Enhancing Anesthetic Patient Education Through the Utilization of Large Language Models for Improved Communication and Understanding

Background/Objectives: The rapid development of Large Language Models (LLMs) presents promising applications in healthcare, including patient education. In anesthesia, where patient anxiety is common due to misunderstandings and fears, LLMs could alleviate perioperative anxiety by providing accessib...

Full description

Saved in:
Bibliographic Details
Main Authors: Jeevan Avinassh Ratnagandhi, Praghya Godavarthy, Mahindra Gnaneswaran, Bryan Lim, Rupeshraj Vittalraj
Format: Article
Language:English
Published: MDPI AG 2025-01-01
Series:Anesthesia Research
Subjects:
Online Access:https://www.mdpi.com/2813-5806/2/1/4
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Background/Objectives: The rapid development of Large Language Models (LLMs) presents promising applications in healthcare, including patient education. In anesthesia, where patient anxiety is common due to misunderstandings and fears, LLMs could alleviate perioperative anxiety by providing accessible and accurate information. This study explores the potential of LLMs to enhance patient education on anesthetic and perioperative care, addressing time constraints faced by anesthetists. Methods: Three language models—ChatGPT-4, Claude 3, and Gemini—were evaluated using three common patient prompts. To minimize bias, incognito mode was used. Readability was assessed with the Flesch–Kincaid, Flesch Reading Ease, and Coleman–Liau indices. Response quality was rated for clarity, comprehension, and informativeness using the DISCERN score and Likert Scale. Results: Claude 3 required the highest reading level, delivering detailed responses but lacking citations. ChatGPT-4o offered accessible and concise answers but missed key details. Gemini provided reliable and comprehensive information and emphasized professional guidance but lacked citations. According to DISCERN and Likert scores, Gemini had the highest rank for reliability and patient friendliness. Conclusions: This study found that Gemini provided the most reliable information, followed by Claude 3, although no significant differences were observed. All models showed limitations in bias and lacked sufficient citations. While ChatGPT-4o was the most comprehensible, it lacked clinical depth. Further research is needed to balance simplicity with clinical accuracy, explore Artificial Intelligence (AI)–physician collaboration, and assess AI’s impact on patient safety and medical education.
ISSN:2813-5806