SURVEY AND PROPOSED METHOD TO DETECT ADVERSARIAL EXAMPLES USING AN ADVERSARIAL RETRAINING MODEL

Artificial intelligence (AI) has found applications across various sectors and industries, offering numerous advantages to human beings. One prominent area where AI has made significant contributions is in machine learning models. These models have revolutionized various fields, benefiting society i...

Full description

Saved in:
Bibliographic Details
Main Authors: Thanh Son Phan, Quang Hua Ta, Duy Trung Pham, Phi Ho Truong
Format: Article
Language:English
Published: Dalat University 2024-08-01
Series:Tạp chí Khoa học Đại học Đà Lạt
Subjects:
Online Access:https://tckh.dlu.edu.vn/index.php/tckhdhdl/article/view/1150
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Artificial intelligence (AI) has found applications across various sectors and industries, offering numerous advantages to human beings. One prominent area where AI has made significant contributions is in machine learning models. These models have revolutionized various fields, benefiting society in numerous ways, from self-driving cars and intelligent chatbots to automated facial authentication systems. However, in recent years, machine learning models have been the target of various attack methods. One common and dangerous attack method is adversarial attack, where modified input images can cause misclassification or erroneous predictions by the models. To confront that challenge, we present a novel approach called adversarial retraining that uses adversarial examples to train machine learning and deep learning models. This technique aims to enhance the robustness and performance of these models by subjecting them to adversarial scenarios during the training process. In this paper, we survey detection methods and propose a method to detect adversarial examples using YOLOv7, a commonly used intensive research model. By training adversarial retraining and conducting experiments, we show that the proposed method is an effective solution for helping deep learning models detect certain cases of adversarial examples.
ISSN:0866-787X