Breaking Machine Learning Models with Adversarial Attacks and its Variants

Machine learning models can be by adversarial attacks, subtle, imperceptible perturbations to inputs that cause the model to produce erroneous outputs. This tutorial introduces adversarial examples and its variants, explaining why even stateof-the-art models are vulnerable and how this impacts secu...

Full description

Saved in:
Bibliographic Details
Main Author: Pavan Reddy
Format: Article
Language:English
Published: LibraryPress@UF 2025-05-01
Series:Proceedings of the International Florida Artificial Intelligence Research Society Conference
Online Access:https://journals.flvc.org/FLAIRS/article/view/139042
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1850277119377014784
author Pavan Reddy
author_facet Pavan Reddy
author_sort Pavan Reddy
collection DOAJ
description Machine learning models can be by adversarial attacks, subtle, imperceptible perturbations to inputs that cause the model to produce erroneous outputs. This tutorial introduces adversarial examples and its variants, explaining why even stateof-the-art models are vulnerable and how this impacts security in AI. It provides an overview of key concepts (such as black-box vs. white-box attack scenarios) and survey common attack techniques and defensive strategies. A hands-on component using Google Colab and the open-source Adversarial Lab toolkit allows attendees to craft adversarial examples and test model robustness in real time. Throughout, we emphasize both the practical skills and the ethical considerations needed to apply adversarial machine learning in a responsible manner. Attendees will gain a comprehensive foundation in adversarial attacks and insights into building more robust, secure machine learning models.
format Article
id doaj-art-e7ec532f92b9418892853dd4fb45e13f
institution OA Journals
issn 2334-0754
2334-0762
language English
publishDate 2025-05-01
publisher LibraryPress@UF
record_format Article
series Proceedings of the International Florida Artificial Intelligence Research Society Conference
spelling doaj-art-e7ec532f92b9418892853dd4fb45e13f2025-08-20T01:49:59ZengLibraryPress@UFProceedings of the International Florida Artificial Intelligence Research Society Conference2334-07542334-07622025-05-0138110.32473/flairs.38.1.139042Breaking Machine Learning Models with Adversarial Attacks and its VariantsPavan Reddy0The George Washington University, Washington DC Machine learning models can be by adversarial attacks, subtle, imperceptible perturbations to inputs that cause the model to produce erroneous outputs. This tutorial introduces adversarial examples and its variants, explaining why even stateof-the-art models are vulnerable and how this impacts security in AI. It provides an overview of key concepts (such as black-box vs. white-box attack scenarios) and survey common attack techniques and defensive strategies. A hands-on component using Google Colab and the open-source Adversarial Lab toolkit allows attendees to craft adversarial examples and test model robustness in real time. Throughout, we emphasize both the practical skills and the ethical considerations needed to apply adversarial machine learning in a responsible manner. Attendees will gain a comprehensive foundation in adversarial attacks and insights into building more robust, secure machine learning models. https://journals.flvc.org/FLAIRS/article/view/139042
spellingShingle Pavan Reddy
Breaking Machine Learning Models with Adversarial Attacks and its Variants
Proceedings of the International Florida Artificial Intelligence Research Society Conference
title Breaking Machine Learning Models with Adversarial Attacks and its Variants
title_full Breaking Machine Learning Models with Adversarial Attacks and its Variants
title_fullStr Breaking Machine Learning Models with Adversarial Attacks and its Variants
title_full_unstemmed Breaking Machine Learning Models with Adversarial Attacks and its Variants
title_short Breaking Machine Learning Models with Adversarial Attacks and its Variants
title_sort breaking machine learning models with adversarial attacks and its variants
url https://journals.flvc.org/FLAIRS/article/view/139042
work_keys_str_mv AT pavanreddy breakingmachinelearningmodelswithadversarialattacksanditsvariants