Review
Abstract
Background: With the rapid development of artificial intelligence, large language models (LLMs) have shown strong capabilities in natural language understanding, reasoning, and generation, attracting much research interest in applying LLMs to health and medicine. Critical care medicine (CCM) provides diagnosis and treatment for patients with critical illness who often require intensive monitoring and interventions in intensive care units (ICUs). Whether LLMs can be applied to CCM, and whether they can operate as ICU experts in assisting clinical decision-making rather than “stochastic parrots,” remains uncertain.
Objective: This scoping review aims to provide a panoramic portrait of the application of LLMs in CCM, identifying the advantages, challenges, and future potential of LLMs in this field.
Methods: This study was conducted in accordance with the PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews) guidelines. Literature was searched across 7 databases, including PubMed, Embase, Scopus, Web of Science, CINAHL, IEEE Xplore, and ACM Digital Library, from the first available paper to August 22, 2025.
Results: From an initial 2342 retrieved papers, 41 were selected for final review. LLMs played an important role in CCM through the following 3 main channels: clinical decision support, medical documentation and reporting, and medical education and doctor-patient communication. Compared to traditional artificial intelligence models, LLMs have advantages in handling unstructured data and do not require manual feature engineering. Meanwhile, applying LLMs to CCM has faced challenges, including hallucinations and poor interpretability, sensitivity to prompts, bias and alignment challenges, and privacy and ethical issues.
Conclusions: Although LLMs are not yet ICU experts, they have the potential to become valuable tools in CCM, helping to improve patient outcomes and optimize health care delivery. Future research should enhance model reliability and interpretability, improve model training and deployment scalability, integrate up-to-date medical knowledge, and strengthen privacy and ethical guidelines, paving the way for LLMs to fully realize their impact in critical care.
Trial Registration: OSF Registries yn328; https://osf.io/yn328/
doi:10.2196/76326
Keywords
Introduction
Critical care medicine (CCM), or intensive care medicine, is an essential field dedicated to managing severely ill patients, emphasizing rapid and life-critical decision-making and interventions. CCM deals with patients who have severe conditions and injuries such as acute kidney injury (AKI), sepsis, and acute respiratory distress syndrome (ARDS), potentially leading to a deteriorative state in the intensive care units (ICUs) [-]. The incidence of AKI in the ICU could reach over 50% worldwide []. Among those who received renal replacement therapy, most of whom were with critical illness in the ICU, the mortality rate was approximately 50% [,]. The prevalence of sepsis is around 30% during ICU stay []. Sepsis accounted for approximately 11 million deaths, making up about 20% of all global deaths []. Recent multicenter epidemiological work shows that the incidence of ARDS in the ICU was between 7.1% and 19% with hospital mortality of 32%-55% []. While in resource-limited settings, the ICU mortality of ARDS could be as high as 50% due to the disparities in health care services []. Therefore, the special environment of the ICU has imposed higher professional requirements on medical staff. Physicians and nurses in ICUs must manage large amounts of patient data while maintaining high efficiency under high pressure [,]. The dynamic and severe nature of critical care demands intelligent decision-support tools to help physicians improve diagnostic accuracy, optimize therapeutic strategies, and facilitate timely clinical decision-making.
Artificial intelligence (AI) technologies, especially generative artificial intelligence (GenAI) models, have developed rapidly in recent years [,]. The advent of large language models (LLMs), such as those based on the Transformer architecture [] and pretrained on extensive text corpora, has marked a substantial advancement in natural language processing (NLP). With billions of parameters, these LLMs have demonstrated remarkable capabilities in understanding and generating human-like text []. LLMs have been implemented in various contexts, including answering questions, summarizing texts, and engaging in open-domain conversations []. Compared to human practitioners, LLMs have been perceived as more understanding and efficient []. Among these LLMs, OpenAI’s ChatGPT [] has become a focal point since its launch in November 2022. OpenAI then introduced upgraded versions of ChatGPT, offering enhanced multimodal capabilities to handle diverse inputs such as text, images, code, and table files. LLMs have revolutionized different fields, including health and medicine [,]. A more detailed description of the evolution and applications of LLMs in health and medicine is provided in Note S1 in [-].
In the field of CCM, the emergence of LLMs demonstrates its unique potential. Similar to the application of LLMs in informing patients with cancer of diagnosis, treatment methods, and side effects, LLMs in CCM can help make life-or-death decisions after fusing large volumes of patient data in a short time []. Physicians in CCM face enormous workloads and pressure, involving LLMs in different clinical decision-making scenarios in CCM will help reduce the workload of physicians and improve health care quality. However, LLMs face challenges when applied in CCM, such as uncertain accuracy and coherence, recency bias, hallucinations, poor interpretability, and ethical issues []. Among them, hallucinations are one of the biggest drawbacks of LLMs, which make them act like stochastic parrots [].
This study aims to review the applications of LLMs in CCM, identifying the advantages, challenges, and future potential of LLMs in this field. Three key research questions were designed to be answered by this review. (1) What is the current status of LLM applications within the critical care setting? (2) What are the recognized advantages and challenges of using LLMs in CCM? (3) What research directions should be taken in the future to promote the application of LLMs in CCM? By addressing the above 3 questions, this review endeavors to provide a clear portrait of and identify the research gaps in the applications of LLMs in CCM, discerning whether they are just stochastic parrots that may mimic human responses based on probability calculation or emerging ICU experts capable of providing timely and highly personalized diagnosis and treatment recommendations. Through this comprehensive review, we aim to outline a roadmap for future research and implementation of LLMs in CCM that could enable them to transform critical care effectively.
Methods
Study Design
This scoping review followed the PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews) guidelines [], and the protocol was registered in the Open Science Framework. We have included a checklist of the PRISMA-ScR guidelines in Table S1 in .
Literature Search Strategy
We conducted a literature search across 7 databases, including PubMed, Embase, Scopus, Web of Science, CINAHL, IEEE Xplore, and ACM Digital Library, from the earliest available paper until August 22, 2025. Keywords related to LLMs included “large language model,” “LLM,” “generative pre-trained transformer,” “GPT,” “generative artificial intelligence,” and “generative AI.” For CCM, the keywords included “critical care,” “intensive care units,” “critical illness,” “intensive care,” and “ICU.” All these terms were combined using the “OR” and “AND” logical operators to ensure the retrieval of literature that addresses both research areas. The detailed search terms for each database are provided in Table S2 in .
Study Selection
The study selection in this scoping review was conducted to ensure comprehensive coverage and relevance of the included literature. In the first phase of the study selection, literature was included based on the following criteria: (1) focusing on LLMs in CCM, including studies that explicitly used or commented on LLMs relevant to the field of CCM, and (2) original research papers from peer-reviewed journals and conferences, perspectives, and letters. Studies were excluded from the review if they met any of the following conditions: (1) irrelevant to LLMs or CCM, including studies that did not focus on applying LLMs within the realm of CCM; (2) conference abstracts, preprint papers, books, patents, editorials, and review papers; and (3) non-English literature. The process for selecting sources of evidence is provided in Note S2 in .
Keyword Co-Occurrence Network Analysis
Keyword co-occurrence network analysis [] is a bibliometric method to explore the relationships between keywords in academic papers. It involves constructing a network where nodes represent keywords and edges represent the co-occurrence of these keywords within the studied documents. It helps to identify the main research themes, trends, and potential research gaps by analyzing the frequency and patterns of keyword co-occurrences. This study used the VOSviewer (version 1.6.20) software to construct a bibliometric network using the visualization of similarity method [,]. The software automatically extracts keywords from a publication’s title, abstract, or author-supplied keyword list. The frequency of co-occurrences of 2 keywords is determined by the number of publications in which both keywords appear together in the title, abstract, or keyword list. The visualization of similarity method starts by calculating the similarity between the keywords of publications based on their co-occurrence. Finally, a matrix is constructed to arrange keywords spatially according to their similarities, and it is the basis for multivariate statistical and network analysis.
Risk of Bias and Applicability Assessment
We critically appraised all included studies using the PROBAST-AI (Prediction Model Risk of Bias Assessment Tool-Artificial Intelligence) [], rating Risk of Bias (RoB) across 4 domains (participants, predictors, outcome, and analysis) and applicability across 3 domains (participants, predictors, and outcome) on a 3-level scale (low, high, and unclear). Full evaluation criteria and rules are provided in Note S3 in .
Results
Literature Search Results
This scoping review covered publications in the 7 databases to August 22, 2025, and retrieved 2342 papers initially. The flowchart of the study selection process is presented in .
The application of LLMs in CCM is a relatively innovative field, but research is still lacking, and the overall number of papers is relatively small. Finally, 41 papers met all the inclusion criteria and were chosen for this review. documents the research contents and publication details of the included studies. The study design and model performance details of the included studies are in Table S3 in , where metrics (such as area under the receiver operating characteristic curve and area under the precision-recall curve, and F1-score) are described, together with the setting, implementation effect on patient outcomes, validation design, and external-validation environment.

| Authors | Published year | Article type | Study design | Journal or conference name | Country | Model | Research contents |
| Savage et al [] | 2023 | Original research | Retrospective | JMIR Medical Informatics | United States | BioMed-RoBERTa | Developed and validated a language model-based screening tool for optimizing Best Practice Alerts. |
| Levin et al [] | 2024 | Original research | Retrospective | International Journal of Nursing Studies | Israel | ChatGPT-4 and Claude-2 (Anthropic PBC) | Compared LLMsa to neonatal nurses in clinical decision support for neonatal care. |
| Pham et al [] | 2024 | Original research | Retrospective | Journal of Medical Internet Research | United States | ChatGPT-3.5 and ChatGPT-4 | Evaluated LLMs in simulations for cardiac arrest and bradycardia based on the American Heart Association’s Advanced Cardiovascular Life Support guidelines. |
| Huespe et al [] | 2023 | Observational study | Retrospective | Critical care explorations | Argentina | ChatGPT-3.5 | Evaluated the capabilities of LLMs in generating the background sections of critical care clinical research questions compared to human researchers. |
| Si et al [] | 2019 | Original research | Retrospective | Journal of the American Medical Informatics Association | United States | ELMo and BERT | Applied contextual embeddings to enhance clinical concept extraction from medical texts. |
| Almazyad et al [] | 2023 | Original research | Retrospective | Cureus | Saudi Arabia | ChatGPT-4 | Used LLMs to enhance expert panel discussions at a medical conference, focusing on pediatric palliative care and ethical decision-making scenarios. |
| Chung et al [] | 2024 | Original research | Retrospective | JAMAb surgery | United States | ChatGPT-4 Turbo | Evaluated the performance of LLMs in perioperative risk stratification and prognostication across various tasks. |
| Abdullahi et al [] | 2024 | Original research | Retrospective | JMIR Medical Education | United States | Bard (Google LLC), ChatGPT-3.5, and ChatGPT-4 | Assessed the effectiveness of LLMs in diagnosing rare and complex medical conditions, focusing on improving medical education and diagnostic accuracy. |
| Benboujja et al [] | 2024 | Original research | Retrospective | Frontiers in Public Health | United States | ChatGPT-4 | Developed and applied a multilingual, AIc-driven educational curriculum in pediatric care to overcome language barriers in global health care education. |
| Lu et al [] | 2023 | Letter | Retrospective | Annals of Biomedical Engineering | China | ChatGPT-3.5 and ChatGPT-4 | Explored potential uses of LLMs in intensive care medicine, focusing on knowledge augmentation, device management, clinical decision support, early warning systems, and ICUd database establishment. |
| Madden et al [] | 2023 | Letter | Retrospective | Intensive Care Medicine | Ireland | ChatGPT-4 | Evaluated the effectiveness of LLMs in querying and summarizing unstructured medical notes in the ICU. |
| Liu et al [] | 2024 | Original research | Prospective | Heliyon | China | ChatGPT-3.5 and ChatGPT-4 | Assessed LLMs in predicting the risk of endotracheal intubation after initiating high-flow oxygen therapy, highlighting potential in decision-making in critical care. |
| Nolan et al [] | 2024 | Original research | Retrospective | Critical Care Explorations | United States | ChatGPT-3.5 and BERT | Investigated the use of LLMs in assisting surrogate and proxy decision-making in critical care, focusing on aligning treatment recommendations with patient values. |
| Shah-Mohammadi and Finkelstein [] | 2024 | Original research | Retrospective | JMIR Medical Informatics | United States | ChatGPT-3.5 | Investigated ChatGPT-3.5’s application in extracting substance use information from ICU discharge summaries, highlighting improvements in accuracy with different learning scenarios. |
| Nawab et al [] | 2024 | Original research | Retrospective | Journal of Medical Artificial Intelligence | United States | ChatGPT-3.5 Turbo | Evaluated ChatGPT-3.5 Turbo’s effectiveness in assigning ICD-10e codes to clinical notes, demonstrating improved accuracy with fine-tuning, particularly in critical care administrative tasks. |
| Soleimani et al [] | 2024 | Original research | Retrospective | Academic Radiology | Iran | ChatGPT-3.5 and Claude.ai | Assessed ChatGPT’s ability to generate radiology reports using MIMIC-CXRf data, with a 3-step prompt guiding report synthesis. Compared outputs to Bart and XLM, showing high similarity to human reports. |
| Oh et al [] | 2024 | Original research | Prospective | Healthcare Informatics Research | Korea | ChatGPT-3.5 and ChatGPT-4 | Analyzed ChatGPT-3.5-turbo and ChatGPT-4 in predicting sepsis mortality using data from the Korean Sepsis Alliance. |
| Urquhart et al [] | 2024 | Original research | Retrospective | Intensive Care Medicine Experimental | Ireland | ChatGPT-3.5, ChatGPT-4, and Llama-2 (Meta) | Evaluated ChatGPT-4, ChatGPT, and Llama 2 for generating ICU discharge summaries, emphasizing event recall and readability. |
| Pabon et al [] | 2024 | Original research | Retrospective | European Journal of Heart Failure | United States | ChatGPT-3.5 | Explored in-hospital outcomes for heart failure patients with improved ejection fraction. ChatGPT-3.5 was used to extract LVEFg data from medical records, but incomplete data detection in some cases required manual review. |
| Akhondi-Asl et al [] | 2024 | Original research | Retrospective | Pediatric Critical Care Medicine | United States | Llama-7B, Llama-65B, and BioGPT-Large | Evaluated domain-specific fine-tuned models against general LLMs for generating differential diagnoses in PICUh patients. Fine-tuned Llama-7B outperformed larger models, demonstrating the importance of domain-specific training. |
| Liu et al [] | 2025 | Original research | Retrospective | Clinical Simulation in Nursing | China | ChatGPT-3.5 | Qualitative exploration of ICU novice simulation instructors’ experience with ChatGPT in case design, focusing on perceived value, potential applications, and limitations. |
| Berger et al [] | 2025 | Original research | Retrospective | Journal of Critical Care | Switzerland | ChatGPT-4o | Qualitative investigation of abbreviation uses in ICU communication, focusing on risks, clinician perceptions, and patient safety implications. |
| Kurz et al [] | 2025 | Original research | Retrospective | NPJ Digital Medicine | Germany | DeepSeek, InternVL, and ChatGPT-4o | Comparative benchmarking of LLMs for diagnostic accuracy using medical images plus clinical context in emergency and critical care. |
| Pham et al [] | 2025 | Original research | Retrospective | Cureus | Vietnam | ChatGPT-4o | Assessment of ChatGPT-4o’s ability to interpret cranial ultrasound images for PV-IVHi diagnosis in very preterm infants, compared to pediatric radiologists. |
| Shi et al [] | 2025 | Original research | Retrospective | Journal of Medical Internet Research | China | SWEDEHEART-AI, Qwen-2, and Llama-3 | Comparison of LLMs for predicting 1-year all-cause mortality post-AMIj, using structured variables versus discharge note analysis. |
| Yitzhaki et al [] | 2025 | Original research | Retrospective | Journal of Paediatrics and Child Health | Israel | ChatGPT-4o | Comparative evaluation of ChatGPT-4 versus PICU specialist in answering open-ended medical education questions sourced from a trainee WhatsApp (WhatsApp LLC) forum. |
| Williams and Erstad [] | 2025 | Original research | Retrospective | American Journal of Health-System Pharmacy | United States | ChatGPT-4, Copilot (Microsoft Corp), Gemini 1.5, and Meta AI | Evaluation of 4 LLMs’ responses to SCCMk guideline–based medication questions. |
| Yang et al [] | 2025 | Original research | Retrospective | JMIR Medical Informatics | China | ICU-GPTl | Development of an automated deployment and extraction platform to allow SQL generation and data retrieval from ICU-related databases without coding. |
| Workum et al [] | 2025 | Original research | Retrospective | Critical Care | Netherlands | ChatGPT-4o, ChatGPT-4o-mini, ChatGPT-3.5-turbo, Mistral Large 2407, and Llama-3.1 70B | Benchmarking 5 LLMs using expert-level ICU MCQsm, compared against human physicians and random guessing. |
| Yang et al [] | 2025 | Original research | Retrospective | Frontiers in Artificial Intelligence | United States | ChatGPT-3.5, ChatGPT-4, Claude 2, Llama2-7B, and Llama2-13B | Comparative evaluation of 5 LLMs on multiple-choice questions in critical care pharmacotherapy education, including prompt-engineering effects and a custom GPT. |
| Ding et al [] | 2024 | Original research | Retrospective | Scientific Reports | United States | BlueBERT | Developed a framework that distills LLM knowledge into structured multimodal EHRn predictive models for ICU health event prediction. |
| Walker et al [] | 2025 | Original research | Retrospective | Journal of the American Medical Informatics Association | United States | ChatGPT-3.5 | Development of CARE-SD: supervised classifiers to detect stigmatizing and doubt-marker language in ICU clinical notes using lexicon- and model-based NLPo. |
| Chen et al [] | 2025 | Original research | Retrospective | BMC Medical Education | United States | DeepL, Gemini (Google LLC), Google Translate, and Microsoft Copilot | Multimodal assessment of freely available MTp tools translating critical care educational content into Chinese, Spanish, and Ukrainian. |
| Ucdal et al [] | 2025 | Original research | Retrospective | Journal of Clinical Medicine | Turkey | Gemini | Evaluation of Gemini’s application of ACGq 2024 guidelines to diagnose severity and guide management in acute pancreatitis. |
| Balta et al [] | 2024 | Original research | Retrospective | Journal of Intensive Care Medicine | Canada | ChatGPT-3.5 and ChatGPT-4 | Comparative evaluation of ChatGPT-3.5 versus 4.0 on appropriateness, consistency, and readability of critical care recommendations. |
| Zhu et al [] | 2025 | Original research | Retrospective | BME Frontiers | China | ChatGPT-4o and ChatGPT-4o mini | Evaluation of contextualized versus static word embeddings in predicting AKIr using ICU clinical notes and structured data, via CNNs models. |
| Pathak et al [] | 2025 | Original research | Retrospective | IEEE Journal of Biomedical and Health Informatics | United States | RespBERT | Development and evaluation of RespBERT that identifies ARDSt from radiology report texts using BERT embeddings and transfer learning. |
| Liu et al [] | 2025 | Original research | Retrospective | IEEE Journal of Biomedical and Health Informatics | Australia | ChatGPT-4o | Development of a note-specific hierarchical network for predicting ICU in-hospital mortality from clinical notes; compares against supervised baselines and LLMs using diverse prompting strategies. |
| Turan et al [] | 2025 | Original research | Prospective | Journal of Clinical Anesthesia | Turkey | ChatGPT-4 | Prospective evaluation of ChatGPT-4 in interpreting ABGu test results compared to expert anesthesiologists. |
| Wang et al [] | 2025 | Original research | Retrospective | Journal of Critical Care | China | ChatGPT-4 | Evaluation of ChatGPT-4’s performance on the Chinese critical care physician qualification examination covering multiple domains. |
| Yang et al [] | 2025 | Original research | Retrospective | Journal of Medical Internet Research | China | ChatGPT-4, Qwen-2, and Llama-3 | LLM-driven extraction of entities and relations to build a sepsis knowledge graph using multicenter clinical data. |
aLLM: large language model.
bJAMA: Journal of the American Medical Association.
cAI: artificial intelligence.
dICU: intensive care unit.
eICD-10: International Statistical Classification of Diseases, Tenth Revision.
fMIMIC-CXR: Medical Information Mart in Intensive Care-Chest X-Ray.
gLVEF: left ventricular ejection fraction.
hPICU: pediatric intensive care unit.
iPV-IVH: periventricular-intraventricular hemorrhage.
jPost-AMI: postacute myocardial infarction.
kSCCM: Society of Critical Care Medicine.
lICU-GPT: Intensive Care Unit-specific Generative Pre-trained Transformer.
mMCQ: multiple choice question.
nEHR: electronic health record.
oNLP: natural language processing.
pMT: machine translation.
qACG: American College of Gastroenterology.
rAKI: acute kidney injury.
sCNN: convolutional neural network.
tARDS: acute respiratory distress syndrome.
uABG: arterial blood gas.
Bibliometric Analysis
This scoping review included a focused selection of 41 papers, providing a global perspective on LLM applications in CCM. This diverse corpus spans several countries, demonstrating widespread research interest in applying LLMs in CCM. The results of keyword co-occurrence network analysis are in Note S4 in . Among the 41 papers, only 2 used prospective data, while the other 39 were retrospective studies. The distribution of the selected publications indicates substantial international collaboration and research efforts. We analyzed the countries where each selected paper’s first and corresponding authors were based. The authors from the United States took the lead in most studies, followed by authors from China, Ireland, Israel, Korea, etc. It revealed that nearly half of the studies were conducted in the United States, with much fewer contributions from other countries and regions. This indicates a concentration of research activities in applying LLMs in CCM within the United States, potentially reflecting the advanced development and adoption of AI technologies in American critical care settings. Among the LLMs used, ChatGPT-4 appears most frequently, demonstrating its relevance and recent prominence in CCM applications. Other models include ChatGPT-3.5 and models such as Llama, Gemini, Claude, and DeepSeek, highlighting the breadth of generative models explored in the included studies. A minority of studies use domain-adapted or clinical NLP backbones, including BioGPT-Large, BioMed-RoBERTa, BlueBERT, and RespBERT.
Applications of LLMs in CCM
Clinical Decision Support
As illustrated in , the primary application of LLMs in CCM is clinical decision support. LLMs can be applied in diagnosis, treatment planning, and prognosis prediction in in-hospital critical care settings.

In diagnosis, LLMs demonstrate the potential to aid physicians in making diagnostic decisions. Gandomi et al [] explored the use of LLMs (Llama 70B and Mistral) in detecting ARDS from radiology reports in the MIMIC-III (Medical Information Mart for Intensive Care-III) database. The study applied LLMs to identify high-probability ARDS cases based on bilateral infiltrates. Akhondi-Asl et al [] investigated the use of domain-specific fine-tuned models (Llama-7B and BioGPT-Large) against larger general-domain models (Llama-65B) for generating differential diagnoses in pediatric intensive care unit (PICU) patients. They found that fine-tuned Llama-7B outperformed both Llama-65B and BioGPT-Large. Kurz et al [] benchmarked the diagnostic performance of various vision-language models on a multimodal dataset comprising medical images and associated clinical context in ICU environments. They compared several open-source vision-language models against ChatGPT-4o, finding that while open models achieved accuracy only up to 40.4%, ChatGPT-4o substantially outperformed them with an approximation of 68.1%. Pham et al [] evaluated the diagnostic utility of ChatGPT-4o in interpreting cranial ultrasound images to detect periventricular-intraventricular hemorrhage among very preterm infants in a neonatal ICU in Vietnam. Comparing ChatGPT-4o’s image-based diagnoses against pediatric radiologists, the model achieved moderate performance (area under the curve [AUC]=0.796), with 75% sensitivity and 84.2% specificity, alongside fair-to-good interrater agreement. The study highlights ChatGPT-4o’s potential as a supplemental tool for early periventricular-intraventricular hemorrhage screening. Pathak et al [] developed RespBERT, leveraging BERT-based embedding and transfer learning to automatically identify ARDS from unstructured radiology notes across multiple hospital datasets. Applying the model to notes from 2 independent institutions, RespBERT achieved F1-scores of 74.5% and 64.2%, demonstrating robust performance across different clinical settings and indicating its potential for ARDS detection in the ICU.
In treatment planning, LLMs show considerable promise in providing personalized treatment recommendations and optimizing clinical pathways for patients with critical illness. Savage et al [] developed and validated the LLM screening tool to selectively identify patients appropriate for Best Practice Alerts of deep vein thrombosis anticoagulation prophylaxis using the MIMIC-III database. They found that the LLM screening tool improved the precision of Best Practice Alerts, reducing the number of unnecessary alerts by 20% and increasing the applicability of alerts by 14.8%. Pham et al [] evaluated ChatGPT’s performance in treating cardiac arrest and bradycardia simulations in accordance with the American Heart Association’s Advanced Cardiovascular Life Support guidelines. Using the 2020 Advanced Cardiovascular Life Support guidelines, ChatGPT’s responses to 2 simulation scenarios were assessed for accuracy. They found that ChatGPT had a median accuracy of 69% for cardiac arrest and 42% for bradycardia, with significant variability in its outputs, often missing critical actions, and having incorrect medication information. Nolan et al [] used LLMs to support critical care decision-making for incapacitated patients. The study simulated scenarios for 50 patients requiring urgent clinical decisions and incorporated patient values captured through various formats, including free-text narratives. The LLMs were tasked with extracting treatments and generating recommendations based on patient profiles. The results showed that LLMs accurately extracted treatments in 88% of cases and received high scores for providing medically plausible and value-aligned recommendations. Williams et al [] evaluated 4 LLMs (ChatGPT-4, Copilot, Gemini version 1.5, and Meta AI) by medication-related questions based on 6 Society of Critical Care Medicine clinical practice guidelines. Copilot yielded the highest proportion of correct answers, followed by Meta AI, ChatGPT-4, and Gemini, which delivered the most incorrect responses. Despite these capabilities, none of the models consistently matched guideline recommendations, indicating that while clinically promising, AI tools require further development for reliable use in the ICU. Ucdal et al [] assessed the performance of Gemini in applying American College of Gastroenterology 2024 guidelines to clinical decisions in acute pancreatitis management. Using MIMIC-III data consisting of 512 patient cases, the study evaluated Gemini’s accuracy in determining disease severity, recommending nutritional strategies, and antibiotic use. Gemini achieved 85% accuracy for mild cases, 82% for severe cases, and 78%-85% compliance with guideline-based nutritional and management recommendations, demonstrating solid agreement with scoring systems. This suggests that AI may support consistent, guideline-concordant decision-making in pancreatitis care. Turan et al [] conducted a prospective observational study comparing ChatGPT-4’s interpretation of arterial blood gas results with that of expert anesthesiologists using 400 ICU patient samples. The model demonstrated excellent accuracy for parameters such as pH, oxygenation, sodium, chloride, and hemoglobin, though it struggled with bilirubin. While generally reliable, it occasionally recommended unnecessary bicarbonate therapy, highlighting the promise of ChatGPT-4 as a rapid interpretive aid, but underscoring the necessity of clinician oversight. Yang et al [] presented a pioneering study that leveraged ChatGPT-4 and a multicenter real-world sepsis dataset (10,544 patients across 3 hospitals in China) to construct a comprehensive sepsis knowledge graph. By combining clinical guidelines, public data, and advanced prompt engineering, they extracted entities and relationships, building a graph with 1894 nodes and 2021 unique connections. ChatGPT-4 achieved a high F1-score of 76.76 on the study’s sepsis-specific dataset and 65.42 under the few-shot condition, surpassing models such as Qwen-2 and Llama-3.
For prognosis prediction, Amacher et al [] used ChatGPT-4 to predict mortality and poor neurological outcomes at hospital discharge for adult patients who had cardiac arrest. The study involved prompting ChatGPT-4 with 16 prognostic parameters from established post–cardiac arrest scores. The findings showed that ChatGPT-4 achieved an AUC of 0.85 for in-hospital mortality and 0.84 for poor neurological outcomes, comparable to traditional scoring systems. Chung et al [] used ChatGPT-4 to perform risk stratification and predict postoperative outcomes based on procedure descriptions and preoperative clinical notes from electronic health records (EHRs). They found that ChatGPT-4 achieved F1-scores of 0.64 for predicting hospital admission, 0.81 for ICU admission, 0.61 for unplanned admission, and 0.86 for predicting hospital mortality. Liu et al [] conducted a prospective multicenter cohort study using ChatGPT-3.5 and ChatGPT-4 to predict the risk of endotracheal intubation within 48 hours following high-flow nasal cannula oxygen therapy in patients with critical illness. They found that ChatGPT-4 achieved an accuracy comparable to that of specialist physicians, with an AUC of 0.82, which was higher than that of non-specialist physicians (AUC=0.66). Oh et al [] conducted a study using ChatGPT-3.5-turbo and ChatGPT-4 to predict in-hospital mortality for sepsis patients. The study used clinical data from the Korean Sepsis Alliance database, focusing on ICU admissions and using metrics such as the SOFA (Sequential Organ Failure Assessment) score and lactic acid levels. The findings demonstrated that ChatGPT-4 performed comparably to a gradient boosting machine in predicting short-term mortality, particularly for 7-day outcomes. Shi et al [] used the MIMIC-IV (Medical Information Mart for Intensive Care-IV) database to compare the performance of 2 LLMs (Qwen-2 and Llama-3) with a specialized artificial neural network (SWEDEHEART-AI) trained on Swedish registry data, in predicting 1-year all-cause mortality among ICU patients with acute myocardial infarction. SWEDEHEART-AI outperformed both LLMs, maintained consistent area under the receiver operating characteristic curve in time-dependent analyses, and demonstrated superior clinical utility and net benefit across risk thresholds, suggesting its stronger reliability for risk stratification. Ding et al [] proposed a novel framework, cross-modality knowledge learning and extraction, that distills knowledge from LLMs into a predictive model trained on multimodal EHR data in the ICU. By refining clinical text using LLM-generated embeddings and using a cross-modality knowledge distillation approach that combines contrastive and patient-similarity learning losses, cross-modality knowledge learning and extraction significantly improved predictive accuracy for hypertension and heart failure events, demonstrating up to a 4.48% boost over state-of-the-art models using data from the MIMIC-III database. Zhu et al [] used clinical data from 2 Chinese hospitals and a public South Korean dataset, comprising a total of 2649 older adult patients who underwent surgery, to use LLMs (GPT-4o, ChatGPT-4o mini, Qwen2-7B-Instruct, and Llama3.1-8B-Instruct), comparing their performance against traditional ML models such as XGBoost (Extreme Gradient Boosting) and Random Forest for the task of predicting postoperative AKI. The study enhanced the LLMs’ capabilities through prompt engineering techniques such as Medical Chain of Thought and instruction fine-tuning. The results demonstrated that the LLM-based frameworks achieved superior generalization on external datasets while also providing human-readable medical rationales for predictions, significantly improving interpretability and clinical utility compared to traditional ML approaches. Liu et al [] investigated risk prediction of in-hospital mortality using routinely collected clinical notes in the ICU. It proposes a note-specific hierarchical network that adapts to different note types and benchmarks it against various supervised baselines and 34 instruction-following LLMs under zero-shot and few-shot settings, as well as chain-of-thought prompting. The hierarchical model outperformed both LLMs and supervised baselines, which consistently underperformed in this critical task, highlighting important constraints of LLMs in risk assessment of critical care patients.
Medical Documentation and Reporting
LLMs are making strides in medical documentation and reporting by automating and streamlining these processes. Shah-Mohammadi et al [] used the ChatGPT-3.5 model to extract substance use information from ICU discharge summaries in the MIMIC-III database, focusing on tobacco, alcohol, and illicit substances. They explored both zero-shot and few-shot prompt learning settings and found that GPT’s performance in identifying tobacco, alcohol, and substance use varied depending on the learning scenario. Zero-shot learning achieved high accuracy in recognizing substance use, while few-shot learning, although lowering accuracy, improved the identification of substance use status, leading to better recall and F1-scores but lower precision. Nawab et al [] conducted a study using the ChatGPT-3.5 Turbo to automate the assignment of ICD-10 codes to clinical notes in the ICU. Their findings demonstrated that fine-tuning the model with a specialized dataset improved its accuracy from 29.7% to 62.6%. Soleimani et al [] conducted a study using ChatGPT-3.5 to evaluate the performance of radiology report generation. Using data from the MIMIC-CXR (Medical Information Mart for Intensive Care-Chest X-Ray) database, the study explored how ChatGPT, guided by a 3-step prompt, synthesized complete radiology reports. They found that ChatGPT effectively generated comprehensive reports by accurately interpreting both patient characteristics and radiological findings. Urquhart et al [] used ChatGPT-4, ChatGPT-3.5, and Llama 2 to extract key information from ICU patient text records in an Irish population. The study evaluated the models’ ability to generate concise and accurate clinical summaries from unstructured ICU admission notes. The results showed that ChatGPT-4 outperformed the other models in readability, organization, and summarization of clinically significant events, but all models struggled with completeness and narrative coherence. Pabon et al [] used ChatGPT-3.5 for extracting left ventricular ejection fraction data from medical records in a study involving patients with heart failure with improved ejection fraction. The model achieved 100% accuracy in identifying reported left ventricular ejection fraction values but struggled with a capture completeness of 75%. Si et al [] explored the impact of ELMo and BERT on clinical concept extraction tasks using data from the MIMIC-III and other clinical corpora. They found that contextual embeddings pretrained on a large clinical corpus outperformed traditional methods. Madden et al [] used ChatGPT-4 to query and summarize unstructured medical notes in the ICU. They found that while the model could produce concise and useful summaries, it also had significant risks of generating hallucinations. Yang et al [] developed a platform to facilitate the deployment and extraction of critical care-related big data using LLMs. The system leverages Docker (Docker Inc)–based automated database deployment and visualization tools, along with an ICU-fine-tuned LLM ICU-GPT to generate SQL queries and extract data from complex ICU datasets without requiring programming knowledge. This platform enables clinicians to manage, visualize, and retrieve structured insights from large critical care databases through a user-friendly web interface, reducing the technical barrier to big data research in clinical settings. Walker et al [] developed CARE-SD, a classifier-based NLP toolkit designed to identify stigmatizing patient labels and doubt markers within ICU clinical notes. By constructing lexicons (127 stigmatizing expressions and 58 doubt markers) using literature-based stems augmented via Word2Vec and ChatGPT-3.5, and training supervised classifiers on annotated samples drawn from 18 million MIMIC-III sentences, the models achieved macro F1-scores of 0.84 (doubt markers) and 0.79 (stigmatizing labels). This approach supports the detection of linguistic biases in critical care EHRs and could inform interventions to reduce stigmatizing language in health care.
Medical Education and Doctor-Patient Communication
LLMs are used more and more frequently in medical education now. One important area closely connected to LLMs is to generate or answer questions in medical examinations. Workum et al [] conducted a benchmark study by evaluating 5 LLMs (GPT-4o, ChatGPT-4o-mini, ChatGPT-3.5-turbo, Mistral Large 2407, and Llama 3.1 70B) on 1181 multiple-choice questions from the European Diploma in Intensive Care examination. All models significantly outperformed human physicians, with ChatGPT-4o achieving the highest accuracy of 93.3%. Despite outstanding consistency and performance, models still produced incorrect answers and raised concerns about energy consumption, especially for ChatGPT-4o, highlighting the need for ongoing evaluation before clinical deployment. Yang et al [] compared the performance and consistency of 5 LLMs (GPT-3.5, ChatGPT-4, Claude 2, Llama2-7B, and Llama2-13B) on a set of 219 multiple-choice questions covering critical care pharmacotherapy for Doctor of Pharmacy students. The study evaluated accuracy, response variance, and the impact of prompt engineering techniques, such as few-shot chain-of-thought prompting, and the use of a custom-trained GPT model. ChatGPT-4 emerged with the highest accuracy (71.6%), chain-of-thought prompting further improved its performance, and the variance in performance differed across models. Notably, customizing models and prompt strategies can enhance LLM reliability in pharmacy education contexts. Chen et al [] developed and applied a multimodal evaluation framework to assess the performance of widely available machine translation (MT) tools (including DeepL, Gemini, Google Translate, and Microsoft Copilot) in translating critical care educational content from English into Mandarin Chinese, Spanish, and Ukrainian. The study used blinded bilingual clinician ratings (for fluency, adequacy, and meaning), BLEU (bilingual evaluation understudy) scores, and usability assessments to compare MT outputs against professional human translations. The results revealed no single MT tool consistently excelled across languages or metrics, human translation scored best for Chinese, Gemini performed strongest for Spanish, and Microsoft Copilot ranked highest for Ukrainian, highlighting the need for ongoing evaluation of MT tools in critical care education as they rapidly evolve. Wang et al [] evaluated ChatGPT-4 against the Chinese Health Professional Technical Qualification Examination for Critical Care Medicine, which comprises 600 questions across fundamental knowledge, specialized knowledge, practical skills, and related medical knowledge. ChatGPT-4 achieved an overall success rate of 73.5%, surpassing the 60% passing threshold, with the highest accuracy in fundamental knowledge (81.94%). Notably, performance was significantly better on single-choice versus multiple-choice questions (76.72% vs 51.32%, P<.001), with no difference between case-based and non–case-based formats. The study underscores its potential as a clinical decision support and educational aid, while cautioning on the need for expert oversight due to potential inaccuracies.
Meanwhile, LLMs use information such as clinical guidelines to answer questions and do clinical reasoning from real-world medical scenarios. Levin et al [] used 2 LLMs, ChatGPT-4 and Claude-2.0, to provide initial assessment and treatment recommendations for patients in neonatal intensive care settings. The results indicated that both models demonstrated clinical reasoning abilities, with Claude-2.0 outperforming ChatGPT-4 in clinical accuracy in providing initial assessments and treatment recommendations, and response speed. Liu et al [] conducted a qualitative study to explore how novice ICU simulation instructors experience the use of GenAI in case design. Using semistructured interviews with 13 instructors and thematic analysis, the study found that GenAI improved efficiency, provided structured and diverse scenario design, and enhanced learning engagement, especially for beginners. The findings suggest that GenAI can serve as a valuable educational tool in ICU simulation, but must be balanced with instructor-led critical thinking and validated clinical accuracy. Yitzhaki et al [] used 100 educational questions from a PICU trainee WhatsApp forum to compare ChatGPT-4’s performance against pediatric intensive care specialists. Evaluated by 10 PICU experts across multiple tertiary centers, ChatGPT-4’s responses were longer and more complete for factual questions, with 60% being preferred for factual knowledge; however, specialists’ responses were favored in clinical reasoning (67%), reflecting higher accuracy. Integrated answers were chosen in 37% of evaluations, emphasizing the need for expert oversight when using ChatGPT-4 in PICU education. Balta et al [] assessed ChatGPT-3.5 versus ChatGPT-4 by having 2 independent intensivists evaluate LLM-generated recommendations to 50 curated core critical care questions from textbooks. ChatGPT-4 delivered significantly higher median appropriateness scores. The study stresses that both models can confidently produce clinically misleading or hallucinated content and thus should be used with caution in the ICU.
LLMs can also be used to overcome language barriers and enhance communication. Benboujja et al [] developed and evaluated a multilingual, AI-driven curriculum to overcome language barriers in pediatric care. Using ChatGPT-4 for translation, the study created 45 educational video modules in English and Spanish, covering surgical procedures, perioperative care, and patient journeys. Almazyad et al [] used ChatGPT-4 to enhance expert panel discussions in pediatric palliative care. They found that ChatGPT-4 effectively facilitated discussions on do-not-resuscitate conflicts by summarizing key themes such as communication, collaboration, patient and family-centered care, trust, and ethical considerations. Berger et al [] conducted a study to investigate the risks associated with the use of abbreviations in critical care communication. By analyzing perspectives from ICU clinicians, the study highlighted how abbreviations, although designed to save time, often introduce ambiguity, misinterpretation, and patient safety risks in high-stakes environments. The findings emphasize that abbreviations can fall short of their intended efficiency, underscoring the importance of clear communication, standardized language, and improved training to minimize preventable errors and improve patient safety in the ICU.
RoB and Applicability Assessment
Across the literature, most studies were judged to have a high RoB in at least 1 RoB domain, most commonly analysis, including reliance on apparent performance without robust internal validation, absent calibration or uncertainty, and risks of temporal or selection leakage, followed by predictors and participants. Outcome definitions were generally aligned with ICU standards, but sometimes lacked blinded ascertainment. Applicability concerns concentrated in predictors (including dependencies on sources not readily available in real-time ICU workflows). Detailed ratings and justifications are presented in Table S4 in .
Discussion
Principal Findings
This scoping review provided a comprehensive portrait of the role of LLMs in CCM, identifying the applications, advantages, challenges, and future research directions of this area. With the recent advent of LLMs, medicine has witnessed groundbreaking developments and advancements []. Many review papers focus on applying LLMs in health and medicine [,]. Particularly, although there are some review papers on AI in CCM [-], few review papers focus on the application of LLMs in CCM. From the 2342 papers initially retrieved, 41 of them were selected for final review. An extensive examination of the selected literature revealed that LLMs have shown promise in some main aspects of CCM: clinical decision support, medical documentation and reporting, medical education, and doctor-patient communication. Compared with traditional AI models, LLMs have advantages in processing unstructured data and do not require manual feature engineering. At the same time, applying LLMs to CCM faces numerous challenges, including hallucinations and poor interpretability, sensitivity to prompts, bias and alignment challenges, and privacy and ethical issues. The current applications, together with the challenges and future directions of LLMs in CCM identified by this review, are shown in . Our findings highlight the potential of LLMs in critical care practices while also underscoring the need for further research to address corresponding challenges and improve the reliability and applicability of LLMs in the critical care domain.

The application of LLMs in CCM has demonstrated numerous advantages. Compared to traditional machine learning techniques, using LLM technology in CCM can effectively cause understanding and generate natural language, aiding clinicians in writing patient medical records and diagnostic notes [-]. The capabilities of LLMs extend beyond text interpretation and generation. They surpass traditional machine learning methods in handling unstructured data. LLMs can learn directly from extensive patient data without manual feature engineering. Moreover, multimodal LLMs can learn and understand medical images, such as x-rays and CT (computed tomography) scans []. In clinical practice, LLMs can extract critical information from a patient’s historical medical records and combine it with the latest medical research, aiding physicians in identifying rare diseases or those with early symptoms that are not clearly defined []. For medical research, LLMs can assist researchers in summarizing data and information in literature research and providing suggestions for manuscript structure and titles, enhancing the readability and completeness of texts []. LLMs have a wide range of knowledge that can provide physicians with a comprehensive analysis of decision-making across different specialties [].
Challenges
Hallucinations and Poor Interpretability
One of the most critical challenges in applying LLMs in CCM is the occurrence of hallucinations, where the model generates plausible-sounding but factually incorrect information [,]. These hallucinations pose risks, where incorrect recommendations can lead to inappropriate diagnoses or treatment plans, potentially endangering ICU patients’ lives. Studies have documented examples where LLMs hallucinate during critical care decision-making, raising concerns about their reliability [,,]. Additionally, LLM outputs often lack transparency, making it difficult for clinicians to understand how specific decisions are reached. This opacity complicates the tracking of decision-making processes, leading to reduced trust in these systems [,]. The opaque nature of LLMs further complicates their application, as it becomes challenging to verify the factual basis of their outputs, particularly in the ICU, where patient safety is important. The lack of clarity regarding the sources of information makes LLMs unreliable for autonomous decision-making in CCM [,]. Therefore, improving the interpretability and transparency of LLMs remains a critical challenge in their integration into clinical workflows [,].
Sensitivity to Prompts and Inconsistent Performance
LLM-generated outputs are highly sensitive to input prompts, and different prompt strategies may affect the model’s capabilities and performance []. In CCM, where the accuracy of information is essential, this sensitivity can lead to inconsistent results across different clinical scenarios [,]. The same LLM might provide different answers to slightly rephrased prompts, requiring clinicians to analyze multiple iterations to ensure accuracy. This lack of consistency raises concerns about the reliability of LLM applications in critical care, especially when time-sensitive decisions must be made. There is no universal prompt strategy that guarantees high performance across all LLMs, which means that clinicians must carefully craft prompts to suit the context of the task []. The absence of a one-size-fits-all approach for prompting remains a significant limitation of LLMs in critical care settings [].
Model Training and Deployment Challenges
Training and deploying LLMs for critical care applications is a resource-intensive process, requiring vast computational power and large, diverse datasets. Public critical care databases, such as MIMIC (Medical Information Mart for Intensive Care) [,] and eICU [], are commonly used for model training. Moreover, hospital regulations and privacy laws often restrict data sharing, complicating the training of models across multiple centers []. The deployment of LLMs in real-time ICU settings also requires substantial computational resources, which may not be feasible in all clinical environments, particularly those with constrained infrastructure [,,]. Additionally, compliance with local data privacy regulations in some hospitals can limit the use of LLMs in health care settings []. The inability of LLMs to adapt to local computing environments due to privacy concerns further hampers their wide-scale application in critical care [].
Timeliness of Knowledge and Model Updates
CCM is a rapidly evolving field, with frequent updates to clinical guidelines, treatments, and best practices. LLMs, however, are typically trained on static datasets, limiting their ability to remain up-to-date with the latest medical advancements []. This delay in knowledge poses a significant challenge, as outdated information could negatively impact patient outcomes [,]. To ensure their continued relevance in clinical practice, LLMs must be regularly updated with the latest medical knowledge. However, this process is computationally expensive and logistically challenging, particularly for large-scale LLMs [,]. Therefore, ensuring the timeliness of LLMs’ knowledge base is crucial to their successful application in CCM.
Bias and Alignment Challenges
LLMs may unintentionally learn biases from the training data and reproduce them in their outputs, potentially leading to skewed or inappropriate recommendations in critical care. These biases can disproportionately affect specific patient populations, potentially leading to disparities in treatment and care [,,]. In the ICU, biased outputs could result in suboptimal or even harmful decisions. Moreover, aligning LLM behavior with clinical guidelines and ethical standards is also a challenge, as models may not always adhere to best practices when generating recommendations [,,]. Addressing bias and ensuring alignment with clinical guidelines and ethics are essential steps for LLMs to function as reliable tools in critical care [,]. Additionally, excessive reliance on AI-generated alerts can lead to “alarm fatigue,” where clinicians become desensitized to frequent, nonurgent predictions, potentially missing critical care events [].
Lack of Evaluation Benchmarks
Currently, there are no universally accepted standards for evaluating the performance of LLMs in critical care settings. Traditional model evaluation primarily focuses on the accuracy of medical question answering, which may not fully reflect the capabilities of LLMs in critical care clinical practice []. Across the included literature, model discrimination should not be conflated with clinical benefit. Consistent with current professional guidance and editorials, routine adoption should follow prospective evaluations that consider patient-centered outcomes rather than relying solely on technical metrics. Current benchmarks rarely include patient-relevant end points [,]. Without appropriate benchmarks, it is challenging to evaluate the effectiveness of LLMs in critical care or compare different models on an equal standard.
Information Overload and Alarm Fatigue in ICU Workflows
Contemporary ICUs are characterized by a proliferation of bedside devices and dense, multistream monitoring data. Qualitative and narrative evidence suggest that this technological abundance, when poorly integrated, can amplify cognitive load, desensitize clinicians to frequent alarms, and undermine situation awareness and team communication []. LLM-enabled systems may help by acting as context-aware filters and data-to-text summarizers that deduplicate near-identical events across monitors, ventilators, or pumps; prioritize alerts using clinical context; and attach traceable rationales for rapid verification [,]. However, naïve deployments could also increase burden (eg, secondary notifications and unverifiable rationales), so any use of LLMs should be aligned with human-centered monitoring principles and embedded in sociotechnical workflows rather than added as another layer.
Clinician Workload, Burnout, and Documentation Quality
While LLMs show promise for text generation and knowledge integration, clinical burden and professional burnout are often overlooked dimensions of the ICU setting [,]. Hallucinations and lack of interpretability shift the burden from “writing” to “reviewing and evidence verification.” Inconsistent sensitivity to prompts and output leads to trial and error for frontline staff []. At the deployment level, the ICU’s high noise level, multirole collaboration, and rigorous legal review may create new bottlenecks for editing and rework. Furthermore, language biases in generated text and paperwork bloat may compromise team communication and information retrieval efficiency. These factors collectively point to the need for systematic evaluation of clinician burden and burnout beyond model effectiveness.
Privacy and Ethical Concerns
Handling patient data responsibly is a significant concern in critical care, where vast amounts of sensitive information must be processed. Ensuring patient privacy while using LLMs presents both technical and legal challenges. Strict compliance with data protection regulations, such as the General Data Protection Regulation and HIPAA (Health Insurance Portability and Accountability Act), is necessary but can hinder the deployment of LLMs in clinical settings [,]. Moreover, the ethical implications of relying on LLMs for life-or-death decisions raise concerns about accountability and the potential over-reliance on AI in medical decision-making []. To navigate these issues, health care systems must establish clear guidelines for the responsible use of LLMs, ensuring that patient privacy is upheld and ethical standards are maintained. Addressing these privacy and ethical challenges will be essential for gaining clinician and patient trust in AI systems used in CCM [].
Future Directions
Enhancing Model Reliability and Interpretability
Improving the reliability and interpretability of LLMs in CCM is critical for their safe integration into real-world clinical workflows. To enhance model reliability, future research should prioritize improving the quality of training data, particularly by incorporating domain-specific knowledge from critical care environments [,]. The accuracy and reliability of LLMs can be enhanced by improving training data quality, using ensemble learning, evidential reasoning, implementing adversarial training, and multiagent systems [-]. Additionally, the use of methods, such as chain-of-thought reasoning, tree-of-thoughts, and retrieval-augmented generation (RAG), can offer greater interpretability, allowing clinicians to understand how LLMs arrive at specific recommendations []. These interpretability techniques would provide clinicians with a clearer rationale for decision-making, thereby building trust in LLM outputs [,]. LLM outputs should include provenance-linked evidence and enforce concise, structured formats to curb alert or note bloat and reduce verification burden in human-centered ICU workflows []. Further, integrating external knowledge databases such as PubMed through plugins can improve the accuracy of LLM outputs and reduce the risk of hallucinations, particularly in critical care [].
Optimizing Prompt Engineering Techniques
LLMs are highly sensitive to prompts, and developing robust prompt engineering techniques is essential for improving consistency and reliability in CCM. Recently, advancements such as Medprompt, which combines dynamic few-shot, self-generated chain-of-thought, and choice shuffle ensemble, have demonstrated improved performance in general LLMs, particularly in medical contexts []. MedGraphRAG is also a novel graph-based RAG framework designed specifically for the medical domain, enhancing the capabilities of LLMs by generating evidence-based, contextually accurate responses through a hierarchical graph structure, thereby improving transparency and reliability in handling private medical data []. These advancements will be particularly useful for critical care environments, where fast and reliable decision-making is essential. Future research should explore the development of prompt engineering strategies to handle complex clinical tasks [,].
Improving Model Training and Deployment Scalability
To address LLM training and deployment challenges, scalable model architectures, transfer learning, model pruning, and federated learning approaches can be explored to reduce computational demands and facilitate practical deployment []. The emergence of low-powered open-source LLMs running locally could circumvent issues related to data privacy and computational resource constraints []. It is crucial to convert medical datasets into easily accessible structured databases and train health care professionals in the ICUs to use LLMs in clinical practice to aid decision-making []. Collaboration with hospitals to develop structured medical databases will also aid in better training of LLMs for real-time decision-making in critical care environments [,].
Integrating Up-to-Date Medical Knowledge
Using web-based learning systems allows models to update and assimilate the latest medical research and changes in clinical practices on time. Additionally, modular update systems can swiftly integrate new medical discoveries, while expert collaboration ensures the scientific validity and timeliness of model outputs. Moreover, using RAG techniques to connect LLMs with databases in CCM can also address the knowledge timeliness issue to some extent [,].
Mitigating Bias and Ensuring Fairness
Bias mitigation should be approached through preprocessing, in-training, intraprocessing, and postprocessing stages []. Preprocessing techniques involve modifying model inputs to ensure balanced representations. In-training methods focus on adjusting model parameters to mitigate biases through gradient-based updates. Intraprocessing methods modify inference behavior without further training, while postprocessing techniques correct model outputs to ensure fair treatment across demographic groups. Developing bias detection and dataset augmentation algorithms to review and adjust model outputs regularly can help reduce model bias and ensure fairness in CCM [].
Developing Comprehensive Evaluation Benchmarks
Recent studies demonstrated that performance varies across different medical tasks, highlighting the need for task-specific evaluations [,]. Future efforts should focus on developing more sophisticated evaluation frameworks that go beyond traditional metrics and consider the specific challenges of critical care []. For instance, the MIMIC-IV-CDM (Medical Information Mart for Intensive Care-IV-Clinical Decision Making) dataset and evaluation framework, focusing on 2400 real patient cases with acute abdominal pain, offers a new benchmark for evaluating LLMs in clinical decision-making, highlighting the need for more rigorous testing to ensure LLMs meet clinical standards, particularly guidelines []. We must standardize clinician-centered metrics, such as time-in-note, click or keystroke counts, handoff completion time, and note quality (accuracy, completeness, consistency, and readability), and report them alongside patient outcomes in prospective, preregistered ICU studies []. To establish clinical benefit and safety beyond model accuracy, future studies should use preregistered, prospective designs that prespecify outcomes such as mortality, ventilator-free days, time-to-critical interventions, and ICU length of stay, alongside calibration and uncertainty reporting [,]. Future work should explore comparing general LLMs against domain-adapted or fine-tuned LLMs (eg, BioGPT-Large and Llama-Med) on tasks and datasets in CCM. Additionally, collaboration between medical professionals and AI researchers will be necessary to design evaluation metrics that are meaningful, clinically applicable, and capable of guiding LLM improvements [].
Strengthening Privacy and Ethical Guidelines
Data privacy and ethical guidelines are crucial for ensuring that LLMs are safely integrated into CCM [,]. As LLMs handle vast amounts of sensitive patient data, their deployment must comply with strict data protection regulations []. Future research should explore synthetic data generation techniques to augment training datasets while protecting patient privacy, allowing for comprehensive model training without compromising confidentiality [,]. Moreover, collaboration with policymakers, ethicists, and legal experts is necessary to ensure LLM applications comply with ethical and legal requirements, thus protecting patient privacy and data security [].
This scoping review may be limited by selection bias due to the literature databases and inclusion criteria, potentially excluding relevant studies in non-English or outside the selected databases. Additionally, the rapid development of LLMs could render the findings quickly outdated, and the broad scope may have limited the depth of analysis for specific LLM applications in CCM.
Conclusions
In conclusion, although LLMs in CCM are not yet ICU experts, they act as more than stochastic parrots. Applying LLMs in CCM presents a transformative potential for enhancing critical care. LLMs are capable of reasoning beyond random generation, and they have demonstrated capabilities to improve diagnostic accuracy, plan optimal treatments, and provide valuable support in prognosis prediction. However, applying LLMs in CCM is still in its early stages, with very few large models specifically designed and fine-tuned for this domain. Future research should focus on enhancing model reliability and interpretability, optimizing prompt engineering techniques, improving training and deployment scalability, integrating up-to-date medical knowledge, mitigating bias and ensuring fairness, developing comprehensive evaluation benchmarks, and strengthening privacy and ethical guidelines. Close collaboration across multiple disciplines, such as medicine, computer science, and data science, may help catalyze the applications of LLMs in CCM. There is some way to go before making LLMs that become true ICU experts. Nevertheless, we are optimistic that LLMs in CCM will become experts in the near future, helping to improve the quality of critical care and the outcomes of patients with critical illness.
Acknowledgments
This work was supported by Grants from the National Natural Science Foundation of China (No. 82372095), the Humanities and Social Science Project of the Chinese Ministry of Education (No. 22YJA630036), the Zhejiang Provincial Natural Science Foundation of China (No. LZ22F020014), and the Beijing Natural Science Foundation of China (No. 7212201, QY23066).
Authors' Contributions
Conceptualization: TS and GK.
Data curation: TS, ZY, RY, and GK.
Formal analysis: TS, ZY, and GK.
Funding acquisition: GK.
Investigation: TS and GK.
Methodology: TS and GK.
Project administration: TS and GK.
Resources: TS, JM, HZ, and GK.
Supervision: GK.
Validation: TS, RY, and GK.
Visualization: TS.
Writing – original draft: TS, JM, ZY, HX, M Xiong, M Xiao, YL, and GK.
Writing – review & editing: TS, RY, and GK.
Conflicts of Interest
None declared.
Additional materials including Note S1 (Overview of LLMs in health and medicine), Note S2 (The process for selecting sources of evidence), Note S3 (Operational rules for PROBAST-AI ratings), Note S4 (Keyword co-occurrence network), Table S1 (PRISMA-ScR checklist), Table S2 (Search terms used across multiple databases for the scoping review), Table S3 (Design and performance summary for included studies), and Table S4 (Risk of bias and applicability assessed using the PROBAST-AI). LLM: large language model; PROBAST-AI: Prediction Model Risk of Bias Assessment Tool-Artificial Intelligence.
PDF File (Adobe PDF File), 550 KBReferences
- Lu Y, Wu H, Qi S, Cheng K. Artificial intelligence in intensive care medicine: toward a ChatGPT/GPT-4 way? Ann Biomed Eng. 2023;51(9):1898-1903. [FREE Full text] [CrossRef] [Medline]
- Shi T, Xu H, Ma J, Kong G, editors. ICU-TGNN: a Hybrid multitask transformer and graph neural network model for predicting clinical outcomes of patients in the ICU. 2024. Presented at: IEEE International Conference on Systems, Man, and Cybernetics (SMC); October 06-10, 2024; Kuching, Malaysia. [CrossRef]
- Xu H, Liu W, Shi T, Kong G. Neural granger causal discovery for derangements in icu-acquired acute kidney injury patients. AMIA Annu Symp Proc. 2024;2024:1265-1274. [Medline]
- Liu W, Shi T, Xu H, Zhao H, Hao J, Kong G. Identifying acute kidney injury subtypes based on serum electrolyte data in ICU via K-medoids clustering. AMIA Annu Symp Proc. 2024;2024:733-737. [Medline]
- Ostermann M, Lumlertgul N, Jeong R, See E, Joannidis M, James M. Acute kidney injury. Lancet. 2025;405(10474):241-256. [CrossRef] [Medline]
- Hoste EAJ, Kellum JA, Selby NM, Zarbock A, Palevsky PM, Bagshaw SM, et al. Global epidemiology and outcomes of acute kidney injury. Nat Rev Nephrol. 2018;14(10):607-625. [CrossRef] [Medline]
- Lin Y, Shi T, Kong G. Acute kidney injury prognosis prediction using machine learning methods: a systematic review. Kidney Med. 2025;7(1):100936. [FREE Full text] [CrossRef] [Medline]
- Li A, Ling L, Qin H, Arabi YM, Myatra SN, Egi M, et al. Epidemiology, management, and outcomes of sepsis in ICUs among countries of differing national wealth across Asia. Am J Respir Crit Care Med. 2022;206(9):1107-1116. [CrossRef] [Medline]
- Rudd KE, Johnson SC, Agesa KM, Shackelford KA, Tsoi D, Kievlan DR, et al. Global, regional, and national sepsis incidence and mortality, 1990-2017: analysis for the global burden of disease study. Lancet. 2020;395(10219):200-211. [FREE Full text] [CrossRef] [Medline]
- Bardají-Carrillo M, López-Herrero R, Aguilar G, Arroyo-Hernantes I, Gómez-Sánchez E, Camporota L, et al. Epidemiological trends of mechanically ventilated acute respiratory distress syndrome in the twenty-first century: a nationwide, population-based retrospective study. J Intensive Care. 2025;13(1):9. [FREE Full text] [CrossRef] [Medline]
- Crawford AM, Shiferaw AA, Ntambwe P, Milan AO, Khalid K, Rubio R, et al. Global critical care: a call to action. Crit Care. 2023;27(1):28. [FREE Full text] [CrossRef] [Medline]
- Shi T, Zhang Z, Liu W, Fang J, Hao J, Jin S. Identifying subgroups of ICU patients using end-to-end multivariate time-series clustering algorithm based on real-world vital signs data. arXiv. Preprint posted online on July 11, 2023. 2023. [CrossRef]
- Cecconi M, Greco M, Shickel B, Vincent J, Bihorac A. Artificial intelligence in acute medicine: a call to action. Crit Care. 2024;28(1):258. [FREE Full text] [CrossRef] [Medline]
- Shi T, Lin Y, Zhao H, Kong G. Artificial intelligence models for predicting acute kidney injury in the intensive care unit: a systematic review of modeling methods, data utilization, and clinical applicability. JAMIA Open. 2025;8(4):ooaf065. [CrossRef] [Medline]
- Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez A. Attention is all you need. 2017. Presented at: NIPS'17: Proceedings of the 31st International Conference on Neural Information Processing Systems; December 4-9, 2017:6000-6010; Long Beach California USA.
- Zhao WX, Zhou K, Li J, Tang T, Wang X, Hou Y. A survey of large language models. arXiv. Preprint posted online November 24, 2023. 2023. [CrossRef]
- Levin C, Kagan T, Rosen S, Saban M. An evaluation of the capabilities of language models and nurses in providing neonatal clinical decision support. Int J Nurs Stud. 2024;155:104771. [CrossRef] [Medline]
- Yuan M, Bao P, Yuan J, Shen Y, Chen Z, Xie Y, et al. Large language models illuminate a progressive pathway to artificial intelligent healthcare assistant. Med Plus. 2024;1(2):100030. [CrossRef]
- Omiye JA, Gui H, Rezaei SJ, Zou J, Daneshjou R. Large language models in medicine: the potentials and pitfalls : a narrative review. Ann Intern Med. 2024;177(2):210-220. [CrossRef] [Medline]
- Brown T, Mann B, Ryder N, Subbiah M, Kaplan JD, Dhariwal P, et al. Language models are few-shot learners. In: Adv Neural Inf Process Syst. 2020. Presented at: NIPS'20: Proceedings of the 34th International Conference on Neural Information Processing Systems; December 6 - 12, 2020:1877-1901; Vancouver BC Canada. [CrossRef]
- Chowdhery A, Narang S, Devlin J, Bosma M, Mishra G, Roberts A, et al. Palm: scaling language modeling with pathways. J Mach Learn Res. 2023;24(240):1-113. [CrossRef]
- Han X, Zhang Z, Ding N, Gu Y, Liu X, Huo Y, et al. Pre-trained models: past, present and future. AI Open. 2021;2:225-250. [CrossRef]
- Harshvardhan G, Gourisaria MK, Pandey M, Rautaray SS. A comprehensive survey and analysis of generative models in machine learning. Comput Sci Rev. 2020;38:100285. [CrossRef]
- Hu Z, Yang Z, Liang X, Salakhutdinov R, Xing EP. Toward controlled generation of text. 2017. Presented at: Proceedings of the 34 th International Conference on Machine Learning; March 02, 2017:1587-1596; Sydney, Australia. [CrossRef]
- Sarkar K, Liu L, Golyanik V, Theobalt C. HumanGAN: a generative model of human images. 2021. Presented at: International Conference on 3D Vision (3DV); December 01-03, 2021:258-267; London, United Kingdom. [CrossRef]
- Kim S, Lee S-G, Song J, Kim J, Yoon S. FloWaveNet : a generative flow for raw audio. arXiv. Preprint posted online on May 20, 2019. 2018. [CrossRef]
- Wu J, Gan W, Chen Z, Wan S, Lin H. AI-generated content (AIGC): a survey. arXiv. Preprint posted online on Mar 26, 2023. 2023. [CrossRef]
- Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, et al. Generative adversarial networks. Commun ACM. 2020;63(11):139-144. [CrossRef]
- Doersch C. Tutorial on variational autoencoders. arXiv. Preprint posted online on Jan 3, 2021. 2016. [CrossRef]
- Lipton ZC, Berkowitz J, Elkan C. A critical review of recurrent neural networks for sequence learning. arXiv. Preprint posted online on Oct 17, 2015. 2015. [CrossRef]
- Mikolov T, Karafiát M, Burget L, Cernocký J, Khudanpur S. Recurrent neural network based language model. 2010. Presented at: INTERSPEECH 2010, 11th Annual Conference of the International Speech Communication Association; September 26-30, 2010:1045-1048; Makuhari, Chiba, Japan. [CrossRef]
- Hochreiter S, Schmidhuber J. Long short-term memory. Neural Comput. 1997;9(8):1735-1780. [CrossRef] [Medline]
- Peters ME, Neumann M, Iyyer M, Gardner M, Clark C, Lee K, et al. Deep contextualized word representations. 2108. Presented at: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers); June 01, 2018; New Orleans, Louisiana. [CrossRef]
- Devlin J, Chang M-W, Lee K, Toutanova K. BERT: pre-training of deep bidirectional transformers for language understanding. 2019. Presented at: Proceedings of NAACL-HLT 2019; June 02-07, 2019:4171-4186; Minneapolis, Minnesota. [CrossRef]
- Radford A, Wu J, Child R, Luan D, Amodei D, Sutskever I. Language models are unsupervised multitask learners. OpenAI Blog. 2019;1(8):9.
- Lewis M, Liu Y, Goyal N, Ghazvininejad M, Mohamed A, Levy O, et al. BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv:1910.13461. 2019. [CrossRef]
- Kaplan J, McCandlish S, Henighan T, Brown TB, Chess B, Child R, et al. Scaling laws for neural language models. arXiv. Posted online on January 23, 2020. 2020. [CrossRef]
- Wei J, Bosma M, Zhao VY, Guu K, Yu AW, Lester B, et al. Finetuned language models are zero-shot learners. arXiv. Posted online on February 8, 2022. 2021. [CrossRef]
- Wei J, Wang X, Schuurmans D, Bosma M, Ichter B, Xia F, et al. Chain-of-thought prompting elicits reasoning in large language models. 2022. Presented at: NIPS'22: Proceedings of the 36th International Conference on Neural Information Processing Systems; November 28, 2022:24824-24837; Red Hook, NY, United States. [CrossRef]
- Puladi B, Gsaxner C, Kleesiek J, Hölzle F, Röhrig R, Egger J. The impact and opportunities of large language models like ChatGPT in oral and maxillofacial surgery: a narrative review. Int J Oral Maxillofac Surg. 2024;53(1):78-88. [FREE Full text] [CrossRef] [Medline]
- Kothari AN. ChatGPT, large language models, and generative AI as future augments of surgical cancer care. Ann Surg Oncol. 2023;30(6):3174-3176. [CrossRef] [Medline]
- Akinci D'Antonoli T, Stanzione A, Bluethgen C, Vernuccio F, Ugga L, Klontzas ME, et al. Large language models in radiology: fundamentals, applications, ethical considerations, risks, and future directions. Diagn Interv Radiol. 2024;30(2):80-90. [FREE Full text] [CrossRef] [Medline]
- Shen Y, Heacock L, Elias J, Hentel KD, Reig B, Shih G, et al. ChatGPT and other large language models are double-edged swords. Radiology. 2023;307(2):e230163. [CrossRef] [Medline]
- Hu X, Ran AR, Nguyen TX, Szeto S, Yam JC, Chan CKM, et al. What can GPT-4 do for diagnosing rare eye diseases? A pilot study. Ophthalmol Ther. 2023;12(6):3395-3402. [FREE Full text] [CrossRef] [Medline]
- Mihalache A, Popovic MM, Muni RH. Performance of an artificial intelligence chatbot in ophthalmic knowledge assessment. JAMA Ophthalmol. 2023;141(6):589-597. [FREE Full text] [CrossRef] [Medline]
- Arora A, Arora A. The promise of large language models in health care. Lancet. 2023;401(10377):641. [CrossRef] [Medline]
- Wu C, Zhang X, Zhang Y, Wang Y, Xie W. Towards generalist foundation model for radiology. arXiv. Posted online on November 16, 2023. 2023. [CrossRef]
- Hyland SL, Bannur S, Bouzid K, Castro DC, Ranjit M, Schwaighofer A, et al. MAIRA-1: a specialised large multimodal model for radiology report generation. arXiv. Posted online on April 26, 2024. 2023. [CrossRef]
- Yang Z, Batra SS, Stremmel J, Halperin E. Surpassing GPT-4 medical coding with a two-stage approach. arXiv. Posted online on November 22, 2023. 2023. [CrossRef]
- Liu J, Yang S, Peng T, Hu X, Zhu Q. ChatICD: prompt learning for few-shot ICD coding through ChatGPT. 2023. Presented at: IEEE International Conference on Bioinformatics and Biomedicine (BIBM); December 05-08, 2023:4360-4367; Istanbul, Turkiye. [CrossRef]
- Zhu L, Mou W, Chen R. Can the ChatGPT and other large language models with internet-connected database solve the questions and concerns of patient with prostate cancer and help democratize medical knowledge? J Transl Med. 2023;21(1):269. [FREE Full text] [CrossRef] [Medline]
- Yoshiyasu Y, Wu F, Dhanda AK, Gorelik D, Takashima M, Ahmed OG. GPT-4 accuracy and completeness against International Consensus Statement on Allergy and Rhinology: rhinosinusitis. Int Forum Allergy Rhinol. 2023;13(12):2231-2234. [CrossRef] [Medline]
- Yeo YH, Samaan JS, Ng WH, Ting P-S, Trivedi H, Vipani A, et al. Assessing the performance of ChatGPT in answering questions regarding cirrhosis and hepatocellular carcinoma. Clin Mol Hepatol. 2023;29(3):721-732. [FREE Full text] [CrossRef] [Medline]
- Lokker C, Bagheri E, Abdelkader W, Parrish R, Afzal M, Navarro T, et al. Deep learning to refine the identification of high-quality clinical research articles from the biomedical literature: Performance evaluation. J Biomed Inform. 2023;142:104384. [FREE Full text] [CrossRef] [Medline]
- Ayers JW, Poliak A, Dredze M, Leas EC, Zhu Z, Kelley JB, et al. Comparing physician and artificial intelligence chatbot responses to patient questions posted to a public social media forum. JAMA Intern Med. 2023;183(6):589-596. [FREE Full text] [CrossRef] [Medline]
- Qureshi R, Shaughnessy D, Gill KAR, Robinson KA, Li T, Agai E. Are ChatGPT and large language models "the answer" to bringing us closer to systematic review automation? Syst Rev. 2023;12(1):72. [FREE Full text] [CrossRef] [Medline]
- Atas Guvenilir H, Doğan T. How to approach machine learning-based prediction of drug/compound-target interactions. J Cheminform. 2023;15(1):16. [FREE Full text] [CrossRef] [Medline]
- Li Z, Li Y, Li Q, Wang P, Guo D, Lu L, et al. LViT: Language meets vision transformer in medical image segmentation. IEEE Trans Med Imaging. 2024;43(1):96-107. [CrossRef] [Medline]
- Guo Y, Qiu W, Leroy G, Wang S, Cohen T. Retrieval augmentation of large language models for lay language generation. J Biomed Inform. 2024;149:104580. [FREE Full text] [CrossRef] [Medline]
- Nadkarni GN, Sakhuja A. Clinical informatics in critical care medicine. Yale J Biol Med. 2023;96(3):397-405. [FREE Full text] [CrossRef] [Medline]
- Thirunavukarasu AJ, Ting DSJ, Elangovan K, Gutierrez L, Tan TF, Ting DSW. Large language models in medicine. Nat Med. 2023;29(8):1930-1940. [CrossRef] [Medline]
- Bender EM, Gebru T, McMillan-Major A, Shmitchell S. On the dangers of stochastic parrots: can language models be too big? 2021. Presented at: FAccT '21: Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency; March 3-10, 2021:610-623; Virtual Event Canada. [CrossRef]
- Tricco AC, Lillie E, Zarin W, O'Brien KK, Colquhoun H, Levac D, et al. PRISMA Extension for Scoping Reviews (PRISMA-ScR): checklist and explanation. Ann Intern Med. 2018;169(7):467-473. [FREE Full text] [CrossRef] [Medline]
- Chen G, Hong S, Du C, Wang P, Yang Z, Xiao L. Comparing semantic representation methods for keyword analysis in bibliometric research. J Informetrics. 2024;18(3):101529. [CrossRef]
- Van EN, Waltman L. Visualizing bibliometric networks. In: Measuring Scholarly Impact: Methods and Practice. Cham. Springer; 2014:285-320.
- Van EN, Waltman L. VOS: a new method for visualizing similarities between objects. 2007. Presented at: Proceedings of the 30th Annual Conference of the Gesellschaft für Klassifikation e.V; March 8-10, 2006:299-306; Freie Universität Berlin. [CrossRef]
- Moons KGM, Wolff RF, Riley RD, Whiting PF, Westwood M, Collins GS, et al. PROBAST: a tool to assess risk of bias and applicability of prediction model studies: explanation and elaboration. Ann Intern Med. 2019;170(1):W1-W33. [FREE Full text] [CrossRef] [Medline]
- Savage T, Wang J, Shieh L. A large language model screening tool to target patients for Best Practice Alerts: development and validation. JMIR Med Inform. 2023;11:e49886. [FREE Full text] [CrossRef] [Medline]
- Pham C, Govender R, Tehami S, Chavez S, Adepoju OE, Liaw W. ChatGPT's performance in cardiac arrest and bradycardia simulations using the American heart association's advanced cardiovascular life support guidelines: exploratory study. J Med Internet Res. 2024;26:e55037. [FREE Full text] [CrossRef] [Medline]
- Huespe IA, Echeverri J, Khalid A, Carboni Bisso I, Musso CG, Surani S, et al. Clinical research with large language models generated writing-clinical research with ai-assisted writing (CRAW) study. Crit Care Explor. 2023;5(10):e0975. [FREE Full text] [CrossRef] [Medline]
- Si Y, Wang J, Xu H, Roberts K. Enhancing clinical concept extraction with contextual embeddings. J Am Med Inform Assoc. 2019;26(11):1297-1304. [FREE Full text] [CrossRef] [Medline]
- Almazyad M, Aljofan F, Abouammoh NA, Muaygil R, Malki KH, Aljamaan F, et al. Enhancing expert panel discussions in pediatric palliative care: innovative scenario development and summarization with ChatGPT-4. Cureus. 2023;15(4):e38249. [FREE Full text] [CrossRef] [Medline]
- Chung P, Fong CT, Walters AM, Aghaeepour N, Yetisgen M, O'Reilly-Shah VN. Large language model capabilities in perioperative risk prediction and prognostication. JAMA Surg. 2024;159(8):928-937. [CrossRef] [Medline]
- Abdullahi T, Singh R, Eickhoff C. Learning to make rare and complex diagnoses with generative ai assistance: qualitative study of popular large language models. JMIR Med Educ. 2024;10:e51391. [FREE Full text] [CrossRef] [Medline]
- Benboujja F, Hartnick E, Zablah E, Hersh C, Callans K, Villamor P, et al. Overcoming language barriers in pediatric care: a multilingual, AI-driven curriculum for global healthcare education. Front Public Health. 2024;12:1337395. [FREE Full text] [CrossRef] [Medline]
- Madden MG, McNicholas BA, Laffey JG. Assessing the usefulness of a large language model to query and summarize unstructured medical notes in intensive care. Intensive Care Med. 2023;49(8):1018-1020. [CrossRef] [Medline]
- Liu T, Duan Y, Li Y, Hu Y, Su L, Zhang A. ChatGPT achieves comparable accuracy to specialist physicians in predicting the efficacy of high-flow oxygen therapy. Heliyon. 2024;10(11):e31750. [FREE Full text] [CrossRef] [Medline]
- Nolan VJ, Balch JA, Baskaran NP, Shickel B, Efron PA, Upchurch GR, et al. Incorporating patient values in large language model recommendations for surrogate and proxy decisions. Crit Care Explor. 2024;6(8):e1131. [FREE Full text] [CrossRef] [Medline]
- Shah-Mohammadi F, Finkelstein J. Extraction of substance use information from clinical notes: generative pretrained transformer-based investigation. JMIR Med Inform. 2024;12:e56243. [FREE Full text] [CrossRef] [Medline]
- Nawab K, Fernbach M, Atreya S, Asfandiyar S, Khan G, Arora R, et al. Fine-tuning for accuracy: evaluation of generative pretrained transformer (GPT) for automatic assignment of international classification of disease (ICD) codes to clinical documentation. J Med Artif Intell. 2024;7:8-8. [CrossRef]
- Soleimani M, Seyyedi N, Ayyoubzadeh SM, Kalhori SRN, Keshavarz H. Practical evaluation of ChatGPT performance for radiology report generation. Acad Radiol. 2024;31(12):4823-4832. [CrossRef] [Medline]
- Oh N, Cha WC, Seo JH, Choi S, Kim JM, Chung CR, et al. ChatGPT predicts in-hospital all-cause mortality for sepsis: in-context learning with the Korean Sepsis Alliance database. Healthc Inform Res. 2024;30(3):266-276. [FREE Full text] [CrossRef] [Medline]
- Urquhart E, Ryan J, Hartigan S, Nita C, Hanley C, Moran P, et al. A pilot feasibility study comparing large language models in extracting key information from ICU patient text records from an Irish population. Intensive Care Med Exp. 2024;12(1):71. [CrossRef] [Medline]
- Pabon MA, Vaduganathan M, Claggett BL, Chatur S, Siqueira S, Marti-Castellote P, et al. In-hospital course of patients with heart failure with improved ejection fraction in the DELIVER trial. Eur J Heart Fail. 2024;26(12):2532-2540. [CrossRef] [Medline]
- Akhondi-Asl A, Yang Y, Luchette M, Burns JP, Mehta NM, Geva A. Comparing the quality of domain-specific versus general language models for artificial intelligence-generated differential diagnoses in PICU patients. Pediatr Crit Care Med. 2024;25(6):e273-e282. [CrossRef] [Medline]
- Liu J, Wang L, He X, Xia Y, Gong X, Wu R, et al. Leveraging generative artificial intelligence to enhance ICU novice simulation instructors’ case design: a qualitative study. Clin Simul Nurs. 2025;105:101770. [CrossRef]
- Berger S, Grzonka P, Hunziker S, Frei AI, Sutter R. When shortcuts fall short: the hidden danger of abbreviations in critical care. J Crit Care. 2025;91:155236. [FREE Full text] [CrossRef] [Medline]
- Kurz CF, Merzhevich T, Eskofier BM, Kather JN, Gmeiner B. Benchmarking vision-language models for diagnostics in emergency and critical care settings. npj Digit Med. 2025;8(1):423. [FREE Full text] [CrossRef] [Medline]
- Pham HQT, Vo TTL, Nguyen TT, Nguyen NTK, Nguyen PMT, Tran N, et al. Validity of ChatGPT in assisting diagnosis of periventricular-intraventricular hemorrhage via cranial ultrasound imaging in very preterm infants. Cureus. 2025;17(4):e82300. [CrossRef] [Medline]
- Shi B, Chen L, Pang S, Wang Y, Wang S, Li F, et al. Large language models and artificial neural networks for assessing 1-year mortality in patients with myocardial infarction: analysis from the Medical Information Mart for Intensive Care IV (MIMIC-IV) database. J Med Internet Res. 2025;27:e67253. [FREE Full text] [CrossRef] [Medline]
- Yitzhaki S, Peled N, Kaplan E, Kadmon G, Nahum E, Gendler Y, et al. Comparing ChatGPT-4 and a paediatric intensive care specialist in responding to medical education questions: a multicenter evaluation. J Paediatr Child Health. 2025;61(7):1084-1089. [CrossRef] [Medline]
- Williams B, Erstad BL. Analysis of responses from artificial intelligence programs to medication-related questions derived from critical care guidelines. Am J Health Syst Pharm. 2025;82(19):e842-e847. [CrossRef] [Medline]
- Yang Z, Xu S, Liu X, Xu N, Chen Y, Wang S, et al. Large language model-based critical care big data deployment and extraction: descriptive analysis. JMIR Med Inform. 2025;13:e63216. [FREE Full text] [CrossRef] [Medline]
- Workum JD, Volkers BWS, van de Sande D, Arora S, Goeijenbier M, Gommers D, et al. Comparative evaluation and performance of large language models on expert level critical care questions: a benchmark study. Crit Care. 2025;29(1):72. [FREE Full text] [CrossRef] [Medline]
- Yang H, Hu M, Most A, Hawkins WA, Murray B, Smith SE, et al. Evaluating accuracy and reproducibility of large language model performance on critical care assessments in pharmacy education. Front Artif Intell. 2025;7:1514896. [FREE Full text] [CrossRef] [Medline]
- Ding S, Ye J, Hu X, Zou N. Distilling the knowledge from large-language model for health event prediction. Sci Rep. 2024;14(1):30675. [FREE Full text] [CrossRef] [Medline]
- Walker A, Thorne A, Das S, Love J, Cooper HLF, Livingston M, et al. CARE-SD: classifier-based analysis for recognizing provider stigmatizing and doubt marker labels in electronic health records: model development and validation. J Am Med Inform Assoc. 2025;32(2):365-374. [CrossRef] [Medline]
- Chen CL, Dong Y, Castillo-Zambrano C, Bencheqroun H, Barwise A, Hoffman A, et al. A systematic multimodal assessment of AI machine translation tools for enhancing access to critical care education internationally. BMC Med Educ. 2025;25(1):1022. [FREE Full text] [CrossRef] [Medline]
- Ucdal M, Bakhshandehpour A, Durak MB, Balaban Y, Kekilli M, Simsek C. Evaluating the role of artificial intelligence in making clinical decisions for treating acute pancreatitis. J Clin Med. 2025;14(12):4347. [FREE Full text] [CrossRef] [Medline]
- Balta KY, Javidan AP, Walser E, Arntfield R, Prager R. Evaluating the appropriateness, consistency, and readability of ChatGPT in critical care recommendations. J Intensive Care Med. 2024;40(2):184-190. [FREE Full text] [CrossRef] [Medline]
- Zhu H, Wang R, Qian J, Wu Y, Jin Z, Shan X, et al. Leveraging large language models for predicting postoperative acute kidney injury in elderly patients. BME Front. 2025;6:0111. [FREE Full text] [CrossRef] [Medline]
- Pathak A, Marshall C, Davis C, Yang P, Kamaleswaran R. RespBERT: a multi-site validation of a natural language processing algorithm, of radiology notes to identify acute respiratory distress syndrome (ARDS). IEEE J Biomed Health Inform. 2025;29(2):1455-1463. [CrossRef] [Medline]
- Liu J, Nguyen A, Capurro D, Verspoor K. Comparing text-based clinical risk prediction in critical care: a note-specific hierarchical network and large language models. IEEE J Biomed Health Inform. 2025;29(10):7657-7667. [CrossRef] [Medline]
- Turan E, Baydemir AE, Balıtatlı AB, Şahin AS. Assessing the accuracy of ChatGPT in interpreting blood gas analysis results ChatGPT-4 in blood gas analysis. J Clin Anesth. 2025;102:111787. [CrossRef] [Medline]
- Wang X, Tang J, Feng Y, Tang C, Wang X. Can ChatGPT-4 perform as a competent physician based on the Chinese critical care examination? J Crit Care. 2025;86:155010. [CrossRef] [Medline]
- Yang H, Li J, Zhang C, Sierra AP, Shen B. Large language model-driven knowledge graph construction in sepsis care using multicenter clinical databases: development and usability study. J Med Internet Res. 2025;27:e65537. [FREE Full text] [CrossRef] [Medline]
- Gandomi A, Wu P, Clement DR, Xing J, Aviv R, Federbush M, et al. ARDSFlag: an NLP/machine learning algorithm to visualize and detect high-probability ARDS admissions independent of provider recognition and billing codes. BMC Med Inform Decis Mak. 2024;24(1):195. [FREE Full text] [CrossRef] [Medline]
- Amacher SA, Arpagaus A, Sahmer C, Becker C, Gross S, Urben T, et al. Prediction of outcomes after cardiac arrest by a generative artificial intelligence model. Resusc Plus. 2024;18:100587. [FREE Full text] [CrossRef] [Medline]
- Hajijama S, Juneja D, Nasa P. Large language model in critical care medicine: opportunities and challenges. Indian J Crit Care Med. 2024;28(6):523-525. [CrossRef] [Medline]
- Meng X, Yan X, Zhang K, Liu D, Cui X, Yang Y, et al. The application of large language models in medicine: a scoping review. iScience. 2024;27(5):109713. [CrossRef] [Medline]
- Singhal K, Azizi S, Tu T, Mahdavi SS, Wei J, Chung HW, et al. Large language models encode clinical knowledge. Nature. 2023;620(7972):172-180. [FREE Full text] [CrossRef] [Medline]
- Shillan D, Sterne JAC, Champneys A, Gibbison B. Use of machine learning to analyse routinely collected intensive care unit data: a systematic review. Crit Care. 2019;23(1):284. [FREE Full text] [CrossRef] [Medline]
- Moazemi S, Vahdati S, Li J, Kalkhoff S, Castano LJV, Dewitz B, et al. Artificial intelligence for clinical decision support for monitoring patients in cardiovascular ICUs: a systematic review. Front Med (Lausanne). 2023;10:1109411. [FREE Full text] [CrossRef] [Medline]
- Syed M, Syed S, Sexton K, Syeda HB, Garza M, Zozus M, et al. Application of machine learning in intensive care unit (ICU) settings using MIMIC dataset: systematic review. Informatics (MDPI). 2021;8(1):16. [FREE Full text] [CrossRef] [Medline]
- Zheng Y, Gan W, Chen Z, Qi Z, Liang Q, Yu PS. Large language models for medicine: a survey. Int J Mach Learn Cyber. 2025;16(2):1015-1040. [CrossRef]
- Doshi R, Amin KS, Khosla P, Bajaj S, Chheang S, Forman HP. Quantitative evaluation of large language models to streamline radiology report impressions: a multimodal retrospective analysis. Radiology. 2024;310(3):e231593. [CrossRef] [Medline]
- Islam K, Nipu A, Madiraju P, Deshpandeditors. Autocompletion of chief complaints in the electronic health records using large language models. IEEE; 2023. Presented at: IEEE International Conference on Big Data (BigData); December 15-18, 2023; Sorrento, Italy. [CrossRef]
- Salvagno M, Taccone FS, Gerli AG. Can artificial intelligence help for scientific writing? Crit Care. 2023;27(1):75. [FREE Full text] [CrossRef] [Medline]
- Nazi Z, Peng W. Large language models in healthcare and medical domain: a review. Informatics. 2024;11(3):57. [CrossRef]
- Huang Q, Dong X, Zhang P, Wang B, He C, Wang J. Opera: Alleviating hallucination in multi-modal large language models via over-trust penalty and retrospection-allocation. 2024. Presented at: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; June 17-21, 2024; Seattle, Washington. [CrossRef]
- Azamfirei R, Kudchadkar SR, Fackler J. Large language models and the perils of their hallucinations. Crit Care. 2023;27(1):120. [FREE Full text] [CrossRef] [Medline]
- Bushuven S, Bentele M, Bentele S, Gerber B, Bansbach J, Ganter J, et al. "ChatGPT, Can You Help Me Save My Child's Life?" - diagnostic accuracy and supportive capabilities to lay rescuers by chatgpt in prehospital basic life support and paediatric advanced life support cases - an in-silico analysis. J Med Syst. 2023;47(1):123. [FREE Full text] [CrossRef] [Medline]
- Harrer S. Attention is not all you need: the complicated case of ethically using large language models in healthcare and medicine. EBioMedicine. 2023;90:104512. [FREE Full text] [CrossRef] [Medline]
- Goodman RS, Patrinely JR, Stone CA, Zimmerman E, Donald RR, Chang SS, et al. Accuracy and reliability of chatbot responses to physician questions. JAMA Netw Open. 2023;6(10):e2336483. [FREE Full text] [CrossRef] [Medline]
- Johnson AEW, Pollard TJ, Shen L, Lehman LH, Feng M, Ghassemi M, et al. MIMIC-III, a freely accessible critical care database. Sci Data. 2016;3:160035. [FREE Full text] [CrossRef] [Medline]
- Johnson AEW, Bulgarelli L, Shen L, Gayles A, Shammout A, Horng S, et al. MIMIC-IV, a freely accessible electronic health record dataset. Sci Data. 2023;10(1):1. [FREE Full text] [CrossRef] [Medline]
- Pollard TJ, Johnson AEW, Raffa JD, Celi LA, Mark RG, Badawi O. The eICU collaborative research database, a freely available multi-center database for critical care research. Sci Data. 2018;5:180178. [FREE Full text] [CrossRef] [Medline]
- Das MK. Multicenter studies: relevance, design and implementation. Indian Pediatr. 2022;59(7):571-579. [FREE Full text] [Medline]
- Wynants L, Kent DM, Timmerman D, Lundquist CM, Van Calster B. Untapped potential of multicenter studies: a review of cardiovascular risk prediction models revealed inappropriate analyses and wide variation in reporting. Diagn Progn Res. 2019;3(1):6. [FREE Full text] [CrossRef] [Medline]
- Khalpey Z, Kumar U, King N, Abraham A, Khalpey AH. Large language models take on cardiothoracic surgery: a comparative analysis of the performance of four models on American board of thoracic surgery exam questions in 2023. Cureus. 2024;16(7):e65083. [CrossRef] [Medline]
- Gallegos I, Rossi R, Barrow J, Tanjim M, Kim S, Dernoncourt F. Bias and fairness in large language models: a survey. Computational Linguistics. 2024;50(3):1097-1179. [CrossRef]
- Suwała S, Szulc P, Guzowski C, Kamińska B, Dorobiała J, Wojciechowska K, et al. ChatGPT-3.5 passes Poland's medical final examination-Is it possible for ChatGPT to become a doctor in Poland? SAGE Open Med. 2024;12:20503121241257777. [FREE Full text] [CrossRef] [Medline]
- Park Y, Pillai A, Deng J, Guo E, Gupta M, Paget M, et al. Assessing the research landscape and clinical utility of large language models: a scoping review. BMC Med Inform Decis Mak. 2024;24(1):72. [FREE Full text] [CrossRef] [Medline]
- Biesheuvel LA, Dongelmans DA, Elbers PWG. Artificial intelligence to advance acute and intensive care medicine. Curr Opin Crit Care. 2024;30(3):246-250. [FREE Full text] [CrossRef] [Medline]
- Wang D, Zhang S. Large language models in medical and healthcare fields: applications, advances, and challenges. Artif Intell Rev. 2024;57(11):299. [CrossRef]
- Armoundas AA, Narayan SM, Arnett DK, Spector-Bagdady K, Bennett DA, Celi LA, American Heart Association Institute for Precision Cardiovascular Medicine, Council on Cardiovascular Stroke Nursing, Council on Lifelong Congenital Heart Disease Heart Health in the Young, Council on Cardiovascular Radiology Intervention, Council on Hypertension, Council on the Kidney in Cardiovascular Disease, et al. Stroke Council. Use of artificial intelligence in improving outcomes in heart disease: a scientific statement from the American heart association. Circulation. 2024;149(14):e1028-e1050. [FREE Full text] [CrossRef] [Medline]
- Nagata JM, Otmar CD, Shim J, Balasubramanian P, Cheng CM, Li EJ, et al. Social media use and depressive symptoms during early adolescence. JAMA Netw Open. 2025;8(5):e2511704. [FREE Full text] [CrossRef] [Medline]
- Olsen E, Novikov Z, Sakata T, Lambert MH, Lorenzo J, Bohn R, et al. More isn't always better: technology in the intensive care unit. Health Care Manage Rev. 2024;49(2):127-138. [CrossRef] [Medline]
- Poncette A, Spies C, Mosch L, Schieler M, Weber-Carstens S, Krampe H, et al. Clinical requirements of future patient monitoring in the intensive care unit: qualitative study. JMIR Med Inform. 2019;7(2):e13064. [FREE Full text] [CrossRef] [Medline]
- Gasciauskaite G, Lunkiewicz J, Roche TR, Spahn DR, Nöthiger CB, Tscholl DW. Human-centered visualization technologies for patient monitoring are the future: a narrative review. Crit Care. 2023;27(1):254. [FREE Full text] [CrossRef] [Medline]
- Pelletier JH, Watson K, Michel J, McGregor R, Rush SZ. Effect of a generative artificial intelligence digital scribe on pediatric provider documentation time, cognitive burden, and burnout. JAMIA Open. 2025;8(4):ooaf068. [CrossRef] [Medline]
- Williams CYK, Subramanian CR, Ali SS, Apolinario M, Askin E, Barish P, et al. Physician- and large language model-generated hospital discharge summaries. JAMA Intern Med. 2025;185(7):818-825. [CrossRef] [Medline]
- Small WR, Austrian J, O'Donnell L, Burk-Rafel J, Hochman KA, Goodman A, et al. Evaluating hospital course summarization by an electronic health record-based large language model. JAMA Netw Open. 2025;8(8):e2526339. [FREE Full text] [CrossRef] [Medline]
- The Lancet Digital Health. ChatGPT: friend or foe? Lancet Digit Health. 2023;5(3):e102. [FREE Full text] [CrossRef] [Medline]
- Shi T, Guo L, Shen Z, Kong G. ERTool: a python package for efficient implementation of the evidential reasoning approach for multi-source evidence fusion. Health Data Sci. 2024;4:0128. [FREE Full text] [CrossRef] [Medline]
- Pan S, Luo L, Wang Y, Chen C, Wang J, Wu X, et al. Unifying large language models and knowledge graphs: a roadmap. IEEE Trans Knowl Data Eng. 2024;36(7):3580-3599. [CrossRef]
- Fan Z, Yu Z, Yang K, Chen W, Liu X, Li G, et al. Diverse models, united goal: a comprehensive survey of ensemble learning. CAAI Trans Intel Tech. 2025;10(4):959-982. [CrossRef]
- Bhardwaj M, Xie T, Boots B, Jiang N. Adversarial model for offline reinforcement learning. 2023. Presented at: NIPS '23: Proceedings of the 37th International Conference on Neural Information Processing Systems; December 10-16, 2023:1245-1269; New Orleans, LA.
- Gao S, Fang A, Huang Y, Giunchiglia V, Noori A, Schwarz JR, et al. Empowering biomedical discovery with AI agents. Cell. 2024;187(22):6125-6151. [FREE Full text] [CrossRef] [Medline]
- Singh C, Inala J, Galley M, Caruana R, Gao J. Rethinking interpretability in the era of large language models. arXiv. Preprint posted online on January 30, 2024. 2024. [CrossRef]
- Wu X, Zhao H, Zhu Y, Shi Y, Yang F, Liu T. Usable XAI: 10 strategies towards exploiting explainability in the LLM era. arXiv. Preprint posted online on May 18, 2025. 2024. [CrossRef]
- Zhao P, Zhang H, Yu Q, Wang Z, Geng Y, Fu F. Retrieval-augmented generation for AI-generated content: a survey. arXiv. Preprint posted online on June 21, 2024. 2024. [CrossRef]
- Nori H, Lee Y, Zhang S, Carignan D, Edgar R, Fusi N. Can generalist foundation models outcompete special-purpose tuning? case study in medicine. arXiv. Preprint posted online on November 28, 2023. 2023. [CrossRef]
- Wu J, Zhu J, Qi Y. Medical graph RAG: towards safe medical large language model via graph retrieval-augmented generation. 2024. Presented at: Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers); July 27-August 1, 2025; Vienna, Austria. [CrossRef]
- Li S, Miao D, Wu Q, Hong C, D'Agostino D, Li X, et al. Federated Learning in Healthcare: A Benchmark Comparison of Engineering and Statistical Approaches for Structured Data Analysis. Health Data Sci. 2024;4:0196. [FREE Full text] [CrossRef] [Medline]
- Mondal D, Lipizzi C. Mitigating large language model bias: automated dataset augmentation and prejudice quantification. Computers. 2024;13(6):141. [CrossRef]
- Jahan I, Laskar MTR, Peng C, Huang JX. A comprehensive evaluation of large language models on benchmark biomedical text processing tasks. Comput Biol Med. 2024;171:108189. [FREE Full text] [CrossRef] [Medline]
- Hager P, Jungmann F, Holland R, Bhagat K, Hubrecht I, Knauer M, et al. Evaluation and mitigation of the limitations of large language models in clinical decision-making. Nat Med. 2024;30(9):2613-2622. [CrossRef] [Medline]
- Duggan MJ, Gervase J, Schoenbaum A, Hanson W, Howell JT, Sheinberg M, et al. Clinician experiences with ambient scribe technology to assist with documentation burden and efficiency. JAMA Netw Open. 2025;8(2):e2460637. [FREE Full text] [CrossRef] [Medline]
- Thirunavukarasu AJ, Mahmood S, Malem A, Foster WP, Sanghera R, Hassan R, et al. Large language models approach expert-level clinical knowledge and reasoning in ophthalmology: a head-to-head cross-sectional study. PLOS Digit Health. 2024;3(4):e0000341. [FREE Full text] [CrossRef] [Medline]
Abbreviations
| AI: artificial intelligence |
| AKI: acute kidney injury |
| ARDS: acute respiratory distress syndrome |
| AUC: area under the curve |
| BLEU: bilingual evaluation understudy |
| CCM: critical care medicine |
| CT: computed tomography |
| EHR: electronic health record |
| GenAI: generative artificial intelligence |
| HIPAA: Health Insurance Portability and Accountability Act |
| ICU: intensive care unit |
| LLM: large language model |
| MIMIC: Medical Information Mart for Intensive Care |
| MIMIC-CXR: Medical Information Mart for Intensive Care-Chest X-Ray |
| MIMIC-III: Medical Information Mart for Intensive Care-III |
| MIMIC-IV: Medical Information Mart for Intensive Care-IV |
| MIMIC-IV-CDM: Medical Information Mart for Intensive Care-IV-Clinical Decision Making |
| MT: machine translation |
| NLP: natural language processing |
| PICU: pediatric intensive care unit |
| PRISMA-ScR: Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews |
| PROBAST-AI: Prediction Model Risk of Bias Assessment Tool-Artificial Intelligence |
| RAG: retrieval-augmented generation |
| RoB: Risk of Bias |
| SOFA: Sequential Organ Failure Assessment |
| XGBoost: Extreme Gradient Boosting |
Edited by A Benis; submitted 23.Apr.2025; peer-reviewed by D Yoo, J-J Beunza, VV Khanna; comments to author 09.Jul.2025; revised version received 04.Sep.2025; accepted 13.Oct.2025; published 24.Nov.2025.
Copyright©Tongyue Shi, Jun Ma, Zihan Yu, Haowei Xu, Rongxin Yang, Minqi Xiong, Meirong Xiao, Yilin Li, Huiying Zhao, Guilan Kong. Originally published in JMIR Medical Informatics (https://medinform.jmir.org), 24.Nov.2025.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on https://medinform.jmir.org/, as well as this copyright and license information must be included.

