From ff45d9c704c3ccb9f0a0cda4959b8f634b251a99 Mon Sep 17 00:00:00 2001
From: Yifei Ge <89774867+goufei123@users.noreply.github.com>
Date: Tue, 22 Oct 2024 16:39:04 +0800
Subject: [PATCH] Update README.md
---
README.md | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/README.md b/README.md
index 64d3a34..b093627 100644
--- a/README.md
+++ b/README.md
@@ -64,7 +64,7 @@ The survey analyzes security threats to CodeLMs, categorizing existing attack ty
[//]: # (The threats discussed in the survey are divided into three main categories:)
According to the document, security threats in CodeLMs are mainly classified into two categories: backdoor attacks and adversarial attacks. Backdoor attacks occur during the training phase, where attackers implant hidden backdoors in the model, allowing it to function normally on benign inputs but behave maliciously when triggered by specific patterns. In contrast, adversarial attacks happen during the testing phase, where carefully crafted perturbations are added to the input, causing the model to make incorrect predictions with high confidence while remaining undetectable to humans.
-
+
### Backdoor Attacks
Backdoor attacks inject malicious behavior into the model during training, allowing the attacker to trigger it at inference time using specific triggers: