A framework to assess clinical safety and hallucination rates of LLMs for medical text summarisation

Abstract Integrating large language models (LLMs) into healthcare can enhance workflow efficiency and patient care by automating tasks such as summarising consultations. However, the fidelity between LLM outputs and ground truth information is vital to prevent miscommunication that could lead to com...

Full description

Saved in:
Bibliographic Details
Main Authors: Elham Asgari, Nina Montaña-Brown, Magda Dubois, Saleh Khalil, Jasmine Balloch, Joshua Au Yeung, Dominic Pimenta
Format: Article
Language:English
Published: Nature Portfolio 2025-05-01
Series:npj Digital Medicine
Online Access:https://doi.org/10.1038/s41746-025-01670-7
Tags: Add Tag
No Tags, Be the first to tag this record!

Similar Items