Localizing Adversarial Attacks To Produces More Imperceptible Noise

Adversarial attacks in machine learning traditionally focus on global perturbations to input data, yet the potential of localized adversarial noise remains underexplored. This study systematically evaluates localized adversarial attacks across widely-used methods, including FGSM, PGD, and C&W,...

Full description

Saved in:
Bibliographic Details
Main Authors: Pavan Reddy, Aditya Sanjay Gujral
Format: Article
Language:English
Published: LibraryPress@UF 2025-05-01
Series:Proceedings of the International Florida Artificial Intelligence Research Society Conference
Online Access:https://journals.flvc.org/FLAIRS/article/view/139004
Tags: Add Tag
No Tags, Be the first to tag this record!