Relieve Adversarial Attacks Based on Multimodal Training
This paper explores the role of multimodal training in mitigating the problems caused by adversarial attacks, building on the foundations of deep learning. Deep learning models have reached great success in many areas such as image recognition and natural language processing. But their robustness ha...
Saved in:
Main Author: | |
---|---|
Format: | Article |
Language: | English |
Published: |
EDP Sciences
2025-01-01
|
Series: | ITM Web of Conferences |
Online Access: | https://www.itm-conferences.org/articles/itmconf/pdf/2025/01/itmconf_dai2024_02004.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | This paper explores the role of multimodal training in mitigating the problems caused by adversarial attacks, building on the foundations of deep learning. Deep learning models have reached great success in many areas such as image recognition and natural language processing. But their robustness has always been a concern. However, the emergence of adversarial attacks has exposed shortages of neural networks, forcing people to confront their limitations and further increasing concerns about the security of deep learning models. Adversarial training is an effective defense mechanism that incorporates adversarial samples into the training data, enabling models to better detect and resist attacks. This paper first introduces the principles and types of adversarial attacks, as well as basic concepts and related methods, including Fast Gradient Sign Method (FGSM), Projected Gradient Descent (PGD), DeepFool, and Jacobian Saliency Map Attack (JSMA). The paper then focuses on analyzing the robustness of the multimodal model CLIP based on contrastive learning. Finally, the paper proposes whether audio data can be added to the training samples of the CLIP model to further improve its robustness, and raises related issues and bottlenecks. |
---|---|
ISSN: | 2271-2097 |