BDEKD: mitigating backdoor attacks in NLP models via ensemble knowledge distillation
Abstract Backdoor attacks present significant risks to the security of deep neural networks (DNNs) in NLP domain, as the attackers can covertly manipulate the model’s output behavior either by poisoning the training data or tampering model’s training process. This paper introduces a novel backdoor d...
Saved in:
| Main Authors: | Zijie Zhang, Xinyuan Miao, Chenyu Zhou, Chenming Shang, Xi Chen, Xianglong Kong, Wei Huang, Yi Cao |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
Springer
2025-07-01
|
| Series: | Complex & Intelligent Systems |
| Subjects: | |
| Online Access: | https://doi.org/10.1007/s40747-025-02006-4 |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
Backdoor Attack Against Dataset Distillation in Natural Language Processing
by: Yuhao Chen, et al.
Published: (2024-12-01) -
A survey of backdoor attacks and defences: From deep neural networks to large language models
by: Ling-Xin Jin, et al.
Published: (2025-09-01) -
A4FL: Federated Adversarial Defense via Adversarial Training and Pruning Against Backdoor Attack
by: Saeed-Uz-Zaman, et al.
Published: (2025-01-01) -
Towards Robust Speech Models: Mitigating Backdoor Attacks via Audio Signal Enhancement and Fine-Pruning Techniques
by: Heyan Sun, et al.
Published: (2025-03-01) -
Backdoor defense method in federated learning based on contrastive training
by: Jiale ZHANG, et al.
Published: (2024-03-01)