Large Language Model Architectures in Health Care: Scoping Review of Research Perspectives

BackgroundLarge language models (LLMs) can support health care professionals in their daily work, for example, when writing and filing reports or communicating diagnoses. With the rise of LLMs, current research investigates how LLMs could be applied in medical practice and th...

Full description

Saved in:
Bibliographic Details
Main Authors: Florian Leiser, Richard Guse, Ali Sunyaev
Format: Article
Language:English
Published: JMIR Publications 2025-06-01
Series:Journal of Medical Internet Research
Online Access:https://www.jmir.org/2025/1/e70315
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1850217906983403520
author Florian Leiser
Richard Guse
Ali Sunyaev
author_facet Florian Leiser
Richard Guse
Ali Sunyaev
author_sort Florian Leiser
collection DOAJ
description BackgroundLarge language models (LLMs) can support health care professionals in their daily work, for example, when writing and filing reports or communicating diagnoses. With the rise of LLMs, current research investigates how LLMs could be applied in medical practice and their benefits for physicians in clinical workflows. However, most studies neglect the importance of selecting suitable LLM architectures. ObjectiveIn this literature review, we aim to provide insights on the different LLM model architecture families (ie, Bidirectional Encoder Representations from Transformers [BERT]–based or generative pretrained transformer [GPT]–based models) used in previous research. We report on the suitability and benefits of different LLM model architecture families for various research foci. MethodsTo this end, we conduct a scoping review to identify which LLMs are used in health care. Our search included manuscripts from PubMed, arXiv, and medRxiv. We used open and selective coding to assess the 114 identified manuscripts regarding 11 dimensions related to usage and technical facets and the research focus of the manuscripts. ResultsWe identified 4 research foci that emerged previously in manuscripts, with LLM performance being the main focus. We found that GPT-based models are used for communicative purposes such as examination preparation or patient interaction. In contrast, BERT-based models are used for medical tasks such as knowledge discovery and model improvements. ConclusionsOur study suggests that GPT-based models are better suited for communicative purposes such as report generation or patient interaction. BERT-based models seem to be better suited for innovative applications such as classification or knowledge discovery. This could be due to the architectural differences where GPT processes language unidirectionally and BERT bidirectionally, allowing more in-depth understanding of the text. In addition, BERT-based models seem to allow more straightforward extensions of their models for domain-specific tasks that generally lead to better results. In summary, health care professionals should consider the benefits and differences of the LLM architecture families when selecting a suitable model for their intended purpose.
format Article
id doaj-art-67b7a932fc8c42fc8202cb3bbcca0c4d
institution OA Journals
issn 1438-8871
language English
publishDate 2025-06-01
publisher JMIR Publications
record_format Article
series Journal of Medical Internet Research
spelling doaj-art-67b7a932fc8c42fc8202cb3bbcca0c4d2025-08-20T02:07:57ZengJMIR PublicationsJournal of Medical Internet Research1438-88712025-06-0127e7031510.2196/70315Large Language Model Architectures in Health Care: Scoping Review of Research PerspectivesFlorian Leiserhttps://orcid.org/0000-0001-5347-0493Richard Gusehttps://orcid.org/0000-0002-0033-8277Ali Sunyaevhttps://orcid.org/0000-0002-4353-8519 BackgroundLarge language models (LLMs) can support health care professionals in their daily work, for example, when writing and filing reports or communicating diagnoses. With the rise of LLMs, current research investigates how LLMs could be applied in medical practice and their benefits for physicians in clinical workflows. However, most studies neglect the importance of selecting suitable LLM architectures. ObjectiveIn this literature review, we aim to provide insights on the different LLM model architecture families (ie, Bidirectional Encoder Representations from Transformers [BERT]–based or generative pretrained transformer [GPT]–based models) used in previous research. We report on the suitability and benefits of different LLM model architecture families for various research foci. MethodsTo this end, we conduct a scoping review to identify which LLMs are used in health care. Our search included manuscripts from PubMed, arXiv, and medRxiv. We used open and selective coding to assess the 114 identified manuscripts regarding 11 dimensions related to usage and technical facets and the research focus of the manuscripts. ResultsWe identified 4 research foci that emerged previously in manuscripts, with LLM performance being the main focus. We found that GPT-based models are used for communicative purposes such as examination preparation or patient interaction. In contrast, BERT-based models are used for medical tasks such as knowledge discovery and model improvements. ConclusionsOur study suggests that GPT-based models are better suited for communicative purposes such as report generation or patient interaction. BERT-based models seem to be better suited for innovative applications such as classification or knowledge discovery. This could be due to the architectural differences where GPT processes language unidirectionally and BERT bidirectionally, allowing more in-depth understanding of the text. In addition, BERT-based models seem to allow more straightforward extensions of their models for domain-specific tasks that generally lead to better results. In summary, health care professionals should consider the benefits and differences of the LLM architecture families when selecting a suitable model for their intended purpose.https://www.jmir.org/2025/1/e70315
spellingShingle Florian Leiser
Richard Guse
Ali Sunyaev
Large Language Model Architectures in Health Care: Scoping Review of Research Perspectives
Journal of Medical Internet Research
title Large Language Model Architectures in Health Care: Scoping Review of Research Perspectives
title_full Large Language Model Architectures in Health Care: Scoping Review of Research Perspectives
title_fullStr Large Language Model Architectures in Health Care: Scoping Review of Research Perspectives
title_full_unstemmed Large Language Model Architectures in Health Care: Scoping Review of Research Perspectives
title_short Large Language Model Architectures in Health Care: Scoping Review of Research Perspectives
title_sort large language model architectures in health care scoping review of research perspectives
url https://www.jmir.org/2025/1/e70315
work_keys_str_mv AT florianleiser largelanguagemodelarchitecturesinhealthcarescopingreviewofresearchperspectives
AT richardguse largelanguagemodelarchitecturesinhealthcarescopingreviewofresearchperspectives
AT alisunyaev largelanguagemodelarchitecturesinhealthcarescopingreviewofresearchperspectives