Assessing the adversarial robustness of multimodal medical AI systems: insights into vulnerabilities and modality interactions

The emergence of both task-specific single-modality models and general-purpose multimodal large models presents new opportunities, but also introduces challenges, particularly regarding adversarial attacks. In high-stakes domains like healthcare, these attacks can severely undermine model reliabilit...

Full description

Saved in:
Bibliographic Details
Main Authors: Ekaterina Mozhegova, Asad Masood Khattak, Adil Khan, Roman Garaev, Bader Rasheed, Muhammad Shahid Anwar
Format: Article
Language:English
Published: Frontiers Media S.A. 2025-07-01
Series:Frontiers in Medicine
Subjects:
Online Access:https://www.frontiersin.org/articles/10.3389/fmed.2025.1606238/full
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:The emergence of both task-specific single-modality models and general-purpose multimodal large models presents new opportunities, but also introduces challenges, particularly regarding adversarial attacks. In high-stakes domains like healthcare, these attacks can severely undermine model reliability and their applicability in real-world scenarios, highlighting the critical need for research focused on adversarial robustness. This study investigates the behavior of multimodal models under various adversarial attack scenarios. We conducted experiments involving two modalities: images and texts. Our findings indicate that multimodal models exhibit enhanced resilience against adversarial attacks compared to their single-modality counterparts. This supports our hypothesis that the integration of multiple modalities contributes positively to the robustness of deep learning systems. The results of this research advance understanding in the fields of multimodality and adversarial robustness and suggest new avenues for future studies focused on optimizing data flow within multimodal systems.
ISSN:2296-858X