Multi-model assurance analysis showing large language models are highly vulnerable to adversarial hallucination attacks during clinical decision support

Abstract Background Large language models (LLMs) show promise in clinical contexts but can generate false facts (often referred to as “hallucinations”). One subset of these errors arises from adversarial attacks, in which fabricated details embedded in prompts lead the model to produce or elaborate...

Full description

Saved in:
Bibliographic Details
Main Authors: Mahmud Omar, Vera Sorin, Jeremy D. Collins, David Reich, Robert Freeman, Nicholas Gavin, Alexander Charney, Lisa Stump, Nicola Luigi Bragazzi, Girish N. Nadkarni, Eyal Klang
Format: Article
Language:English
Published: Nature Portfolio 2025-08-01
Series:Communications Medicine
Online Access:https://doi.org/10.1038/s43856-025-01021-3
Tags: Add Tag
No Tags, Be the first to tag this record!