An Empirical Evaluation of Large Language Models on Consumer Health Questions
<b>Background:</b> Large Language Models (LLMs) have demonstrated strong performances in clinical question-answering (QA) benchmarks, yet their effectiveness in addressing real-world consumer medical queries remains underexplored. This study evaluates the capabilities and limitations of...
Saved in:
| Main Authors: | Moaiz Abrar, Yusuf Sermet, Ibrahim Demir |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
MDPI AG
2025-02-01
|
| Series: | BioMedInformatics |
| Subjects: | |
| Online Access: | https://www.mdpi.com/2673-7426/5/1/12 |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
Cross-Encoder-Based Semantic Evaluation of Extractive and Generative Question Answering in Low-Resourced African Languages
by: Funebi Francis Ijebu, et al.
Published: (2025-03-01) -
Evaluating large language models as graders of medical short answer questions: a comparative analysis with expert human graders
by: Olena Bolgova, et al.
Published: (2025-12-01) -
Intelligent accounting question-answering robot based on a large language model and knowledge graph
by: Shi Shengyun, et al.
Published: (2025-04-01) -
Hierarchical Modeling for Medical Visual Question Answering with Cross-Attention Fusion
by: Junkai Zhang, et al.
Published: (2025-04-01) -
Assessing the quality of automatic-generated short answers using GPT-4
by: Luiz Rodrigues, et al.
Published: (2024-12-01)