Published on in Vol 13 (2025)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/68898, first published .
The Use of Machine Learning for Analyzing Real-World Data in Disease Prediction and Management: Systematic Review

The Use of Machine Learning for Analyzing Real-World Data in Disease Prediction and Management: Systematic Review

The Use of Machine Learning for Analyzing Real-World Data in Disease Prediction and Management: Systematic Review

Review

1College of Medicine, Qassim University, Buraidah, Saudi Arabia

2Applied Biotechnology, Faculty of Chemistry, Warsaw University of Technology, Warsaw, Poland

3Malaysian Health Technology Assessment Section, Medical Development Division, Ministry of Health Malaysia, Wilayah Persekutuan Putrajaya, Malaysia

4Health Economics and Health Technology Assessment, School of Health and Wellbeing, University of Glasgow, Glasgow, United Kingdom

5Health Sciences Research Center, Imam Mohammad ibn Saud Islamic University, Riyadh, Saudi Arabia

*these authors contributed equally

Corresponding Author:

Nasser Alotaiq, PhD

Health Sciences Research Center

Imam Mohammad ibn Saud Islamic University

Othman Bin Affan Rd. Al-Nada 13317

Riyadh

Saudi Arabia

Phone: 966 50 411 9153

Email: naalotaiq@imamu.edu.sa


Background: Machine learning (ML) and big data analytics are rapidly transforming health care, particularly disease prediction, management, and personalized care. With the increasing availability of real-world data (RWD) from diverse sources, such as electronic health records (EHRs), patient registries, and wearable devices, ML techniques present substantial potential to enhance clinical outcomes. Despite this promise, challenges such as data quality, model transparency, generalizability, and integration into clinical practice persist.

Objective: This systematic review aims to examine the use of ML for analyzing RWD in disease prediction and management, identifying the most commonly used ML methods, prevalent disease types, study designs, and the sources of real-world evidence (RWE). It also explores the strengths and limitations of current practices, offering insights for future improvements.

Methods: A comprehensive search was conducted following the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines to identify studies using ML techniques for analyzing RWD in disease prediction and management. The search focused on extracting data regarding the ML algorithms applied; disease categories studied; types of study designs (eg, clinical trials and cohort studies); and the sources of RWE, including EHRs, patient registries, and wearable devices. Studies published between 2014 and 2024 were included to ensure the analysis of the most recent advances in the field.

Results: This review identified 57 studies that met the inclusion criteria, with a total sample size of >150,000 patients. The most frequently applied ML methods were random forest (n=24, 42%), logistic regression (n=21, 37%), and support vector machines (n=18, 32%). These methods were predominantly used for predictive modeling across disease areas, including cardiovascular diseases (n=19, 33%), cancer (n=9, 16%), and neurological disorders (n=6, 11%). RWE was primarily sourced from EHRs, patient registries, and wearable devices. A substantial portion of studies (n=38, 67%) focused on improving clinical decision-making, patient stratification, and treatment optimization. Among these studies, 14 (25%) focused on decision-making; 12 (21%) on health care outcomes, such as quality of life, recovery rates, and adverse events; and 11 (19%) on survival prediction, particularly in oncology and chronic diseases. For example, random forest models for cardiovascular disease prediction demonstrated an area under the curve of 0.85 (95% CI 0.81-0.89), while support vector machine models for cancer prognosis achieved an accuracy of 83% (P=.04). Despite the promising outcomes, many (n=34, 60%) studies faced challenges related to data quality, model interpretability, and ensuring generalizability across diverse patient populations.

Conclusions: This systematic review highlights the significant potential of ML and big data analytics in health care, especially for improving disease prediction and management. However, to fully realize the benefits of these technologies, future research must focus on addressing the challenges of data quality, enhancing model transparency, and ensuring the broader applicability of ML models across diverse populations and clinical settings.

JMIR Med Inform 2025;13:e68898

doi:10.2196/68898

Keywords



Background

Advances in big data analytics and the growing availability of real-world data (RWD) are transforming health care by enabling new applications of machine learning (ML) to improve health outcomes [1]. Real-world evidence (RWE) generated from diverse data sources, such as electronic health records (EHRs), patient registries, and wearable devices, has become central to informed decision-making in clinical practice [2,3]. When combined with ML, RWD present a promising avenue to enhance disease prediction, personalize patient management, and optimize therapeutic effectiveness. By providing a comprehensive view of patient histories and real-world health outcomes, ML applications in health care can drive actionable insights across various domains, including disease diagnosis, treatment planning, and chronic disease management [4,5]. RWD capture information about patients in naturalistic settings, revealing how health care is delivered and its outcomes. Unlike clinical trials that operate within controlled conditions, RWD offer a more representative view of patient experiences, treatment responses, and health outcomes [6]. The rise of big data technology and data management systems has facilitated the integration of vast, heterogeneous data types, allowing ML algorithms to identify complex patterns within high-dimensional datasets [7,8]. These capabilities allow health care providers to predict health outcomes, identify at-risk populations, and tailor interventions based on individual patient factors, thus making strides toward precision medicine [9].

Despite their potential, ML applications in RWD and big data contexts face several challenges. Data quality remains a primary concern, as RWD often feature inconsistencies, missing values, and a lack of standardization [10]. Unlike the structured data from controlled clinical trials, RWD demand extensive preprocessing, including advanced natural language processing (NLP) methods and imputation techniques, to address data gaps. Such efforts are critical to enhancing ML model reliability and ensuring accurate, meaningful outcomes [11,12]. Biases present another key issue. ML models trained on RWD may inherit biases from the data, often stemming from demographic imbalances or regional health care differences. If left unaddressed, these biases can lead to health care disparities, as ML-driven decisions might inaccurately represent racial and ethnic minority populations or certain patient groups [13]. Incorporating fairness-aware ML algorithms and cross validating models across multiple datasets can mitigate this challenge, although developing equitable ML models remains a high priority [14]. Another significant hurdle is the interpretability of ML models, especially deep neural network (DNN) models, which are known for their “black box” nature. While complex models deliver high accuracy, their opaque decision-making process limits the ability to verify or explain predictions. Model transparency is crucial given the high stakes in health care, where ML-based recommendations can impact lives. Advances in interpretability tools, such as Shapley Additive Explanations and Local Interpretable Model-Agnostic Explanations, have helped enhance model transparency; however, balancing interpretability with performance remains an area of active investigation [15,16].

The integration of ML with RWD poses ethical and regulatory challenges, especially regarding patient privacy, data security, and informed consent. Regulations such as the Health Insurance Portability and Accountability Act in the United States and the General Data Protection Regulation in the European Union impose strict standards for data protection. However, adapting these laws to the context of ML in health care is complex due to the scale and diversity of the data involved [17,18]. Solutions such as deidentification, secure data-sharing protocols, and clear data management strategies have become crucial to ensuring patient confidentiality while maximizing data utility [19]. Ensuring equitable treatment outcomes is another ethical imperative. ML models trained on data predominantly representing certain demographics may perform poorly on underrepresented groups; therefore, addressing these disparities is critical. By incorporating fairness-aware ML models and building representative datasets, health care practitioners can ensure that ML applications benefit all patient groups, regardless of demographics [20]. Regulatory bodies have started developing specific guidelines for the use of ML and RWD in health care. The Food and Drug Administration (FDA), for example, has issued draft guidance on using RWE for regulatory decisions, and the European Medicines Agency (EMA) has also recognized the importance of RWE in evaluating drug safety and efficacy [21]. As ML applications in health care continue to grow, a solid regulatory framework will be necessary to safeguard patient health while supporting technological innovation.

Objectives

The objective of this systematic review is to explore and critically analyze the applications, challenges, and future directions of ML in processing real-world health data and big data across various disease domains. Specifically, this review aims to identify the disease areas where ML with RWD has shown clinical utility; examine the ML algorithms and methodologies applied to big data in health care; and analyze the challenges related to data quality, bias, and model interpretability. In addition, this review addresses the ethical and regulatory frameworks pertinent to the use of ML in health care, with an emphasis on patient privacy and fairness. Finally, it outlines future research needs and opportunities for innovation in using ML, RWD, and big data for precision medicine and public health.


Eligibility Criteria

For this systematic review, we focused exclusively on clinical trials and cohort studies that used ML techniques to analyze RWD for disease prediction and management. Studies were included if they met the following criteria: (1) they were randomized controlled trials, pragmatic clinical trials, observational clinical trials, or cohort studies; (2) they involved the application of ML methods (eg, supervised learning, unsupervised learning, and deep learning) to RWD for clinical decision-making, disease prediction, or management of common diseases, such as cardiovascular diseases, diabetes, cancer, and chronic conditions; and (3) they used real-world health data sources, such as EHRs, patient registries, or wearable health devices. Exclusion criteria included trials that did not apply ML techniques or used only data from controlled clinical trials rather than real-world settings.

Information Sources

The following information sources were used to capture the most relevant clinical trials and cohort studies: PubMed, Scopus, and the Cochrane Library. PubMed was specifically targeted for clinical trials and biomedical research, particularly studies published in leading clinical journals. Scopus and the Cochrane Library were also searched to gather clinical trial reports within the health care and ML domains. To ensure comprehensive coverage, Google Scholar was included to identify gray literature, such as theses and reports that were not indexed in traditional databases. These sources were selected to provide a broad overview of clinical trial data and their relevance to ML applications in disease management. In addition, regulatory bodies such as the US FDA and the EMA were consulted to gain insights into clinical trial guidelines and regulatory standards regarding the use of RWD in health care.

Search Strategy

A comprehensive search strategy was developed to identify clinical trials and cohort studies focused on ML applications in RWD. The search query incorporated key terms related to ML (eg, “machine learning,” “deep learning,” and “artificial intelligence”) and clinical trials (eg, “clinical trial,” “randomized controlled trial,” and “pragmatic clinical trial”) along with terms related to disease management (eg, “disease prediction,” and “healthcare outcomes”). For example, the search used the following key terms: (“machine learning” OR “deep learning”) AND (“clinical trial” OR “randomized controlled trial” OR “pragmatic trial” OR “cohort study”) AND (“real-world data” OR “electronic health records” OR “patient registries”). Boolean operators (AND and OR), truncation, and Medical Subject Headings terms were used to refine the search and ensure comprehensive coverage. The search was conducted across multiple databases, including the Cochrane Library, PubMed, and Web of Science, covering the period from January 1, 2014, to December 31, 2024, ensuring that the full range of recent literature was captured. In addition, relevant studies were identified through manual searches of reference lists from key articles and by reviewing clinical trial registries, such as ClinicalTrials.gov, to ensure comprehensive coverage of the clinical trials relevant to ML in disease management. Gray literature was identified by conducting targeted searches in Google Scholar and manually retrieving relevant documents suggested by domain experts. Only English-language publications were included. To enhance transparency and reproducibility, the full search strategy, including specific database queries and search filters, has been provided in Multimedia Appendix 1.

Study Selection

The study selection process was conducted in 2 stages: an initial screening of titles and abstracts by 2 independent reviewers, followed by a full-text review by the same 2 reviewers to ensure consistency and minimize bias. In the first stage, the titles and abstracts of all identified articles were assessed for relevance based on predefined inclusion criteria. Discrepancies between reviewers were resolved through discussion, with a third reviewer consulted when necessary. Studies that met the inclusion criteria proceeded to the second stage, where the full texts were retrieved for further evaluation (Textbox 1).

The evaluation of exclusion criteria was conducted independently by both reviewers, with disagreements resolved through discussion. To maintain transparency, the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) flow diagram [22,23] was used to document the number of studies at each review stage, including identification, screening, eligibility, and final inclusion. The PRISMA checklist is provided in Multimedia Appendix 2.

Textbox 1. Inclusion and exclusion criteria for study selection.

Inclusion criteria

  • Study type: clinical trials or cohort studies
  • Data source: studies using real-world data, such as electronic health records, patient registries, claims data, or wearable device data
  • Machine learning (ML) application: application of ML algorithms for disease prediction or management (eg, supervised, unsupervised, and deep learning models)
  • Clinical focus: studies addressing disease prediction, management, monitoring, or outcome prediction in health care
  • Outcome reporting: studies reporting ML model performance metrics (eg, accuracy, area under the curve, sensitivity, and specificity) and real-world data sources
  • Language: studies published in English
  • Publication type: peer-reviewed articles and indexed gray literature

Exclusion criteria

  • Study type: case reports, cross-sectional studies, reviews, editorials, letters, and conference abstracts
  • Data source: studies using simulated data, animal studies, or laboratory-based data
  • ML application: studies using only conventional statistical models (eg, logistic regression and Cox models) or expert systems
  • Clinical focus: studies unrelated to clinical decision-making, disease outcomes, or patient management
  • Outcome reporting: studies lacking sufficient information on ML model performance or data sources
  • Language: non–English-language publications
  • Publication type: non–peer-reviewed sources (blogs and social media posts)

Data Extraction

Data extraction was performed independently by 2 reviewers using a standardized form. Key data points extracted from each clinical trial and cohort study included study characteristics (eg, authors, year of publication, and trial design), the specific ML methods used (eg, supervised learning, reinforcement learning, deep learning), disease areas targeted (eg, cardiovascular diseases, diabetes, and cancer), and the types of RWD sources used (eg, EHRs, patient registries, and wearable devices). In addition, we extracted the performance metrics of the ML models used, such as accuracy, sensitivity, specificity, and area under the receiver operating characteristic curve (AUROC), to evaluate their effectiveness in disease prediction and management. Information on the challenges and limitations of applying ML to RWD in clinical trials, such as data quality issues, biases, or model interpretability, was also collected. Any disagreements in data extraction were resolved through discussion. The extracted data were organized systematically to synthesize findings across studies.


Systematic Literature Search and Study Selection Workflow

The systematic literature search was conducted to identify studies applying ML techniques to RWD in clinical trials and cohort studies, with a focus on disease prediction and management. The search covered multiple databases, including PubMed, Scopus, Web of Science, and the Cochrane Library, to capture a broad range of studies from biomedical, clinical, and health care research fields. This search yielded 11,252 records, as illustrated in the PRISMA flow diagram (Figure 1). To ensure comprehensive coverage, an additional 7 records were identified through external sources such as Google searches, manual hand searching of nonindexed journals, gray literature, and other nontraditional academic sources. After removing duplicates, 7217 (64.13%) unique studies remained for screening. The selection process followed a rigorous, multistage workflow. Title screening was first performed to assess relevance based on predefined inclusion criteria. Studies with titles not indicating the application of ML to RWD in clinical or disease management contexts were excluded, resulting in the removal of 5930 (82.17%) records and leaving 1287 (17.83%) studies for abstract screening. During the abstract screening, each abstract was carefully evaluated for inclusion criteria, including the use of ML techniques, relevant RWD sources (eg, EHRs, patient registries, or wearable device data), and relevance to disease prediction or management. This led to the exclusion of 967 (75.13%) studies that did not meet these criteria, most commonly for the following reasons: (1) a lack of ML implementation, with some studies using only conventional statistical approaches such as logistic regression (LR) or decision trees (DTs) without learning-based model development; (2) irrelevance to clinical trials or cohort study frameworks, instead focusing on simulations, animal studies, or nonhuman data sources; (3) absence of disease prediction or management applications, such as papers limited to health care policy, infrastructure, or economic modeling without patient-centered outcomes; and (4) insufficient use of RWD sources, as studies often used synthetic or trial-generated data rather than EHRs, registries, claims databases, or wearable device data. A total of 320 (24.86%) studies proceeded to the full-text review stage. At this stage, articles were assessed in detail to confirm adherence to all inclusion criteria. A total of 263 studies were excluded for specific reasons: 98 (37.3%) lacked ML algorithms (using conventional statistics instead), 72 (27.4%) were unrelated to clinical trial methodologies, 23 (8.7%) did not involve study cohorts, 51 (19.4%) were unrelated to health care outcomes, and 19 (7.2%) lacked sufficient information on ML model performance or data sources. Following this thorough, systematic, and transparent selection process, 57 studies met all eligibility criteria and were included in the final systematic review. These selected studies represented a diverse range of clinical applications, disease areas, ML methodologies, and RWD sources, offering a comprehensive overview of the current role of ML in clinical trials and cohort studies for disease prediction and management.

Table 1 summarizes these studies, while key findings and methodological details are provided in Multimedia Appendix 3 [24-80].

Figure 1. The PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) flow diagram depicting the study selection process from initial identification to final inclusion, detailing the number of records screened, excluded, and ultimately included in this systematic review.
Table 1. A summary of the studies included in this systematic review, outlining the study characteristics, diseases or medical conditions, type of study, source of real-world evidence (RWE), and machine learning (ML) methods used.
StudyDatabaseDiseases or medical conditions (category)Type of studyType of RWEML methodsModel performance metrics
Wissel et al [24], 2023PubMedEpilepsy (neurological diseases)Evaluation of health care outcomesEHRaNLPb
  • AUCc=0.79 (95% CI 0.62-0.96)
  • Sensitivity=0.80 (95% CI 0.29-0.99)
  • Specificity=0.77 (95% CI 0.64-0.88)
  • Positive predictive value: 0.25 (95% CI 0.07-0.52)
  • Negative predictive value: 0.98 (95% CI 0.87-1.00).
Ayers et al [25], 2021PubMedOrthotopic heart transplantation (CVDd)Survival predictionPatient registriesDNNe, RFf, and AdaBoostg
  • RF AUROCh=0.691 (95% CI 0.671-0.711)
  • DNN AUROC=0.691 (95% CI 0.671-0.712)
  • AdaBoost AUROC=0.653 (95% CI 0.632-0.674)
Nadarajah et al [26], 2023PubMedAtrial fibrillation (CVD)Disease predictionEHRFIND-AFi ML algorithm
  • AUROC=0.824 (95% CI 0.814-0.834)
Yadgir et al [27], 2022PubMedCognitive impairment (neurological diseases)Disease predictionEHRXGBoostj
  • AUROC=0.720
Liu et al [28], 2023PubMedPeripheral artery disease (CVD)Survival predictionEHRLRk, GBMl, RF, DTm, XGBoost, neural network, Cox regression, and RSFn
  • C-index: 0.788 (compared to 0.730 for GermanVasc Score)
Hill et al [29], 2022PubMedAtrial fibrillation (CVD)Disease prediction and cost-effectivenessEHRPULSE-AIo
  • Sensitivity=50%
  • Specificity=90%
Sheth et al [30], 2019PubMedAcute ischemic stroke (CVD)Disease predictionEHRCNNp
  • AUROC=0.88-0.90
Barton et al [31], 2019PubMedSepsis (infectious diseases)Disease predictionEHRXGBoost
  • AUROC of 0.88, 0.84, and 0.83 for sepsis onset and 24 and 48 h before onset, respectively
Kao et al [32], 2023PubMedAtrial fibrillation (CVD)Disease predictionEHRDT, SVMq, LR, and RF
  • AUROC=0.74
  • Specificity=98.7%
Kim et al [33], 2022PubMedAHREsr (CVD)Disease predictionWearable devicesRF, SVM, LR, and XGBoost
  • RF AUROC=0.742
  • SVM AUROC=0.675
  • XGBoost AUROC=0.745
  • LR AUROC=0.669
Park et al [34], 2023PubMedCoronary artery disease (CVD)Disease predictionPatient registriesBQRs
  • AUC of 0.67, 0.65, 0.78, and 0.73 for per‐patient, LADt, LCxu, and RCAv, respectively
Hilbert et al [35], 2019PubMedAcute ischemic stroke (CVD)Health care outcomes and decision-makingWearable devicesResNetw
  • Average AUC for functional outcome was 0.71
  • Average AUC for reperfusion across all folds was 0.65
Chen et al [36], 2021PubMedEwing sarcoma (tumors)Survival predictionPatient registriesBoosted DT, SVM, nonparametric RF, and neural network
  • Sensitivity=77%-83%
  • Specificity=91%-94%
Koutsouleris et al [37], 2016PubMedSchizophrenia (neurological diseases)Health care outcomes and decision-makingPatient registriesNonlinear SVM
  • Test-fold BACx=75%
Strömblad et al [38], 2021PubMedColorectal and gynecologic cancer (cancers)Health care outcomesEHRGBM and LR
  • —y
Wang et al [39], 2019PubMedAtrial fibrillation (CVD)Decision-makingEHRDT

Tan et al [40], 2021PubMedInfluenza (infectious diseases and respiratory diseases)Health care outcomesEHRRF, XGBoost, and LR
  • Accuracy of RF model for hospitalization=0.840, pneumonia=0.765, and sepsis or septic shock=0.857
  • Accuracy of XGBoost for intensive care unit admission=0.902
  • Accuracy of LR for in-hospital mortality=0.889
Goerigk et al [41], 2020PubMedDepression (neurological diseases)Decision-makingPatient registriesLR, SVM, RF, tree-based stochastic gradient boosting, and XGBoost
  • LR: accuracy=0.75, sensitivity=0.76, specificity=0.73, and AUC=0.792
  • SVM: accuracy=0.88, sensitivity=0.85, specificity=0.91, and AUC=0.939
  • RF: accuracy=0.89, sensitivity=0.88, specificity=0.91, and AUC=0.957
  • XGBoost: accuracy=0.88, sensitivity=0.85, specificity=0.91, and AUC=0.954
Kijpaisalratana, et al [42], 2024PubMedSepsis (infectious disease)Decision-makingEHRRF, XGBoost, LR, and SVM
  • AUROC of ML in early sepsis identification was significantly higher than qSOFAz, SIRSaa, and MEWSab
Sharma et al [43], 2019PubMedAcute coronary syndrome (CVD)Survival predictionPatient registriesCox regression

Singhal et al [44], 2021PubMedARDSac (respiratory diseases)Disease predictionEHRML algorithm called “eARDSad” (neural networks, SVM, RF, LR, and XGBoost)
  • AUROC=0.89 (95% CI 0.88-0.90)
  • Sensitivity=0.77 (95% CI 0.75-0.78)
  • Specificity=0.85 (95% CI 085-0.86)
Kanchanatawan et al [45], 2018PubMedSchizophrenia (neurological diseases)Disease predictionPatient registriesSVM and RF
  • SVM AUROC=0.931
  • RF AUROC=0.898
Huang et al [46], 2022PubMedIschemic stroke (CVD)Survival predictionEHRNBae, XGBoost, and LR
  • NB AUROC=0.767
  • XGBoost AUROC=0.989
  • LR AUROC=0.627
She et al [47], 2023PubMedSepsis (infectious disease)Disease predictionPatient registriesSVM and RF
  • AUC=0.98
Sundar et al [48], 2022PubMedGastric cancer (cancers)Survival predictionPatient registriesRF
  • F-measure: 0.71
  • AUC=0.75 (95% CI 0.50-0.99)
Alaa et al [49], 2019PubMedCardiovascular disease risk (CVD)Disease predictionPatient registriesLinear SVM, RF, neural networks, AdaBoost, and XGBoost
  • AutoPrognosis model improved risk prediction (AUROC=0.774, 95% CI 0.768-0.780)
  • Framingham score (AUROC=0.724, 95% CI 0.720-0.728; P<.001)
  • Cox PHaf model with conventional risk factors (AUROC=0.734, 95% CI 0.729-0.739; P<.001)
  • Cox PH model with all UK Biobank variables (AUROC=0.758, 95% CI 0.753-0.763; P<.001)
Azimi et al [50], 2017PubMedLSCSag (spinal diseases)Decision-makingEHRANNah and LR
  • AUC=0.89
Baxter et al [51], 2019PubMedGlaucoma (ocular diseases)Decision-makingEHRMLRai, RF, and ANN
  • AUC=0.67
Anderson et al [52], 2015PubMedType 2 diabetes (metabolic diseases)Disease predictionEHRRF and SVM
  • AUC=0.78
Bannister et al [53], 2018PubMedStroke and myocardial infarction (CVD)Survival predictionPatient registriesCox regression
  • C-index was 0.59, 0.69, and 0.64 and 0.66, 0.70, and 0.70 for the GPaj and Cox regression models, respectively.
Scheer et al [54], 2017Web of ScienceSpinal deformity surgery (spinal diseases)Decision-makingPatient registriesDT and ANN
  • AUROC=0.89
Rau et al [55], 2016Web of ScienceLiver cancer (cancers)Disease predictionEHRANN and LR
  • ANN sensitivity=0.757
  • Specificity=0.755
  • AUROC=0.873
Ramezankhani et al [56], 2016Web of ScienceType 2 diabetes (metabolic diseases)Disease predictionPatient registriesDT
  • AUC=0.78
  • Sensitivity=78%
Pei et al [57], 2019Web of ScienceType 2 diabetes (metabolic diseases)Disease predictionEHRDT
  • Accuracy=94.2%
  • Precision=94.0%
  • Recall=94.2%
  • AUC=94.8%
Oviedo et al [58], 2019Web of SciencePostprandial hypoglycemia (metabolic diseases)Decision-makingWearable devicesSVM
  • Specificity=79%
  • Sensitivity=71%
Mubeen et al [59], 2017Web of ScienceAlzheimer disease (neurological diseases)Disease predictionEHRRF
  • AUC=0.87
  • Accuracy=80.2%
Lopez-de-Andres et al [60], 2016Web of ScienceType 2 diabetes (metabolic diseases)Survival predictionPatient registriesANN
  • AUROC for Elixhauser comorbidity model=91.7% (95% CI 90.3-93.0)
  • AUROC for Charlson comorbidity model=88.9% (95% CI 87.5-90.2)
Kwon et al [61], 2019Web of ScienceCardiac arrest (CVDSurvival predictionPatient registriesDNN, LR, SVM, and RF
  • DNN AUROC=0.953 (95% CI 0.952-0.954)
  • LR AUROC=0.947 (95% CI 0.943-0.948)
  • RF AUROC=0.943 (95% CI 0.942-0.945)
  • SVM AUROC=0.930 (95% CI 0.929-0.932)
Kim et al [62], 2019Web of ScienceBreast cancer (cancers)Decision-makingEHRTwo-class decision jungle and 2-class neural network
  • AUROC was 0.917 in the high RSak group and 0.744 in the low RS group in the test set
Khanji et al [63], 2019Web of ScienceHypertension and dyslipidemia (CVD)Health care outcomesEHRLR
  • cvAUCsal of 0.73 for hypertension, 0.64 for dyslipidemia, and 0.79 for diabetes
Karhade et al [64], 2019Web of ScienceLumbar disk herniation (spinal diseases)Decision-makingEHRLR, RF, XGBoost, ANN, and SVM
  • AUC=0.81
Jovanovic et al [65], 2014Web of ScienceCholedocholithiasis (gastrointestinal diseases)Decision-makingEHRANN
  • AUROC=0.884 (95% CI 0.831-0.938); P<.001
Kang et al [66], 2020Web of SciencePostinduction hypotension (anesthesia-related complications)Health care outcomesEHRNB, LR, RF, and ANN
  • NB AUROC=0.778 (95% CI 0.650-0.898)
  • LR AUROC=0.756 (95% CI 0.630-0.881)
  • RF AUROC=0.842 (95% CI 0.736-0.948)
  • ANN AUROC=0.760 (95% CI 0.640-0.880)
Isma’eel et al [67], 2018Web of ScienceCoronary artery disease (CVD)Health care outcomesEHRANN
  • Sensitivity=91% (CI 81%-97%)
  • Specificity=65% (CI 60%-79%)
Hill et al [68], 2019Web of ScienceAtrial fibrillation (CVD)Disease predictionEHRRF, SVM, and Cox regression
  • RF AUROC=0.827
  • SVM AUROC=0.725
Dong et al [69], 2019Web of ScienceChinese Crohn disease (gastrointestinal diseases)Decision-makingEHRRF, LR, SVM, DT, and ANN
  • RF AUROC=0.9864
  • LR AUROC=0.9538
  • SVM AUROC=0.9497
  • DT AUROC=0.8809
  • ANN AUROC=0.9059
Bowman et al [70], 2018Web of ScienceCarpal tunnel syndrome (musculoskeletal diseases)Health care outcomesEHRLR and ANN
  • AUROC=0.7
Bertsimas et al [71], 2018Web of ScienceBreast, lung, ovarian cancers (cancers)Survival predictionEHRDT
  • AUROC=0.83-0.86
Manz et al [72], 2020The Cochrane LibraryCancer-related serious illness (cancers)Decision-makingEHRRF and SVM

Tian et al [73], 2023The Cochrane LibraryLung transplantation (respiratory diseases)Survival predictionEHRRSF
  • AUROC=0.879 (95% CI 0.832-0.921)
Li et al [74], 2022The Cochrane LibraryLatent profile analysis (cancers)Decision-makingEHRGBM

Tedeschi et al [75], 2021The Cochrane LibraryPseudogout (rheumatic diseases)Disease predictionEHRNLP
  • AUC=0.86
Ambwani et al [76], 2019The Cochrane LibraryCancer biomarkers (cancers)Health care outcomesEHRLR
  • Sensitivity=97.3%
Jorge et al [77], 2019The Cochrane LibraryLupus (autoimmune disease)Disease predictionEHRLR
  • Specificity=97%
  • Sensitivity=64%
Shimabukuro et al [78], 2017The Cochrane LibrarySepsis (infectious diseases)Health care outcomesEHRLR
  • AUROC=0.952 (95% CI 0.946 to –0.958)
  • Specificity=0.900 (95% CI 0.870 to 0.930)
  • Sensitivity=0.900 (95% CI 0.878 to 0.922)
Sarraju et al [79], 2021The Cochrane LibraryAtherosclerosis
(CVD)
Health care outcomesEHRRF, GBM, XGBoost, and LR
  • XGBoost AUC of 0.70 (95% CI 0.68 to –0.71) in the full CVD cohort and AUC of 0.71 (95% CI 0.69 to –0.73) in patients with ASCVDam, with comparable performance by GBM, RF, and Lasso.
Ye et al [80], 2019The Cochrane LibraryMyopia (ocular diseases)Health care outcomesWearable devicesSVM
  • AUC=0.99

aEHR: electronic health record.

bNLP: natural language processing.

cAUC: area under the curve.

dCVD: cardiovascular disease.

eDNN: deep neural network.

fRF: random forest.

gAdaBoost: adaptive boosting.

hAUROC: area under the receiver operating characteristic curve.

iFIND-AF: Future Innovations in Novel Detection for Atrial Fibrillation.

jXGBoost: Extreme Gradient Boosting.

kLR: logistic regression.

lGBM: gradient boosting machine.

mDT: decision tree.

nRSF: random survival forest.

oPULSE-AI: Prediction of Undiagnosed Atrial Fibrillation Using a Machine Learning Algorithm.

pCNN: convolutional neural network.

qSVM: support vector machine.

rAHRE: atrial high-rate episode.

sBQR: Bayesian quantile regression.

tLAD: left anterior descending artery.

uLCx: left circumflex artery.

vRCA: right coronary artery.

wResNet: residual neural network.

xBAC: balanced accuracy.

yNot applicable.

zqSOFA: quick sequential organ failure assessment.

aaSIRS: systemic inflammatory response syndrome.

abMEWS: modified early warning score.

acARDS: acute respiratory distress syndrome.

adeARDS: early onset acute respiratory distress syndrome

aeNB: naive Bayes.

afCox PH: Cox proportional hazards.

agLSCS: lumbar spinal canal stenosis.

ahANN: artificial neural network.

aiMLR: multivariable logistic regression.

ajGP: genetic programming.

akRS: recurrence score.

alcvAUC: cross-validated area under the curve.

amASCVD: atherosclerotic cardiovascular disease.

Implementation of ML in RWD for Disease Prediction and Management

ML methods have become integral tools for analyzing RWD for disease prediction and management. These methods analyze complex medical data, helping clinicians make informed decisions for better patient care. Random forest (RF) is one of the most widely used ML methods, appearing in 42% (24/57) of the studies (Table 2). It is an ensemble learning technique that builds multiple DTs and combines their outputs to improve model stability and generalizability [81]. Several reviewed studies reported that RF performed well in handling large datasets with numerous variables, particularly EHRs, which are common medical data sources. Its robustness against overfitting and ability to handle missing data made it a frequently chosen method in clinical applications, where data quality could vary [82,83]. While RF has been widely applied in predictive modeling for disease outcomes, treatment responses, and patient risk assessments, further comparative studies are necessary to directly evaluate its performance against other ML models in real-world health care settings [84]. It is often applied to predict disease outcomes, assess treatment responses, and identify patient risk factors.

Table 2. The frequency of machine learning (ML) methods used in studies included in this review (N=57).
ML methodStudies, n (%)
Random forest24 (42)
Logistic regression21 (37)
Support vector machine18 (32)
Extreme Gradient Boosting12 (21)
Artificial neural network11 (19)
Decision tree9 (16)
Gradient boosting machine4 (7)
Cox regression3 (5)
Natural language processing2 (4)
Deep neural network2 (4)

LR was a fundamental method used for binary classification tasks and was used in 37% (21/57) of the studies. LR estimates the probability of a particular class, which is essential for predicting binary outcomes, such as the presence or absence of disease. It is a simple yet powerful tool that works well with smaller datasets and provides results that are easy to interpret, making it particularly useful in clinical settings where transparency is crucial [85,86]. LR is commonly used for disease risk prediction, helping clinicians assess the likelihood of a patient developing a condition based on their medical history and other clinical factors. Its interpretability allows for clear communication of results to health care providers, enhancing decision-making [87,88]. Support vector machine (SVM), used in 32% (18/57) of the studies, is known for its ability to handle high-dimensional data, making it suitable for complex medical datasets, including genomic and imaging data. SVM works by finding the optimal hyperplane that separates different classes in the feature space [89-91]. This method is beneficial in clinical settings where the relationship between variables is nonlinear and can be adapted for classification and regression tasks. SVM is applied in disease prediction, particularly when the dataset has many features relative to the number of observations. It is also useful for classifying patients based on genetic or demographic factors, making it a powerful tool for precision medicine [92,93]. Extreme Gradient Boosting (XGBoost) appeared in 21% (12/57) of the studies and is a highly effective method for improving predictive accuracy through boosting. XGBoost builds models sequentially to correct errors made by previous models and uses regularization to prevent overfitting. This method is effective in handling large datasets, common in clinical studies, and where computational efficiency is essential [94,95]. XGBoost is often used for survival analysis and disease outcome prediction, where it can effectively manage the complexity of large datasets and missing data. Its flexibility allows it to be applied across various disease areas, from cancer prognosis to cardiovascular risk assessment [96,97].

Artificial neural network (ANN) models, used in 19% (11/57) of the studies, are powerful tools for modeling complex, nonlinear relationships in data. With multiple layers of interconnected neurons, an ANN can learn intricate patterns from large datasets. It is widely used in applications that involve unstructured data, such as medical imaging and genetic data, where traditional models might struggle [98,99]. ANN is frequently applied to predict disease progression and response to treatments and identify potential biomarkers. In RWD, an ANN helps identify subtle patterns in complex datasets that simpler models might not capture, such as predicting cancer progression from radiological images [100,101]. DTs, featured in 16% (9/57) of the studies, are straightforward and interpretable models that split data into subsets based on feature values. DT models are highly useful in real-world health care settings, where interpretability is essential for clinical decision-making. They are often applied in clinical decision support systems to guide treatment decisions based on patient data [102,103]. In health care, DTs predict disease outcomes, stratify patients by risk, and recommend treatment plans. Their transparency allows clinicians to understand the decision-making process, which is critical for patient trust and informed consent [104]. Gradient boosting machine (GBM) was used in 7% (4/57) of the studies and is a powerful ensemble method that focuses on correcting errors made by previous models. It is effective in producing highly accurate predictions, particularly in the presence of noisy or incomplete data. GBM is more computationally intensive than other methods, but often outperforms simpler models in accuracy. GBM is particularly useful for predicting disease progression and evaluating treatment efficacy in longitudinal studies, where multiple factors influence outcomes over time [105,106].

NLP, used in 4% (2/57) of the studies, is a subfield of artificial intelligence focused on analyzing unstructured textual data. In health care, NLP extracts relevant information from clinical notes, EHRs, and medical literature. It enables clinicians and researchers to analyze vast amounts of text data to identify trends, predict disease outcomes, and assess treatment effectiveness [107,108]. NLP is crucial in extracting insights from EHRs and other textual data sources. It can help in disease prediction by identifying patterns from patient narratives, diagnostic codes, and clinician notes that would otherwise remain hidden in unstructured formats [109]. Cox Regression, used in 5% (3/57) of the studies, is designed explicitly for survival analysis. It is widely applied in clinical research to model the time of an event, such as the onset of a disease or patient survival. This method is precious for understanding how various predictors affect the risk of an event occurring over time. In RWD, Cox regression is often used in cancer studies and other chronic diseases to predict survival times and assess the impact of different treatment regimens, making it indispensable in clinical trials and outcome-based research [110,111]. DNN models, used in 4% (2/57) of the studies, are a more complex version of ANN with multiple hidden layers. DNN models identify intricate patterns and are increasingly used in health care applications involving large and complex data types, such as medical imaging, genomics, and sensor data. DNN is particularly useful for analyzing high-dimensional data, such as medical images (eg, x-rays and magnetic resonance imaging) or genomic data, where the relationships between variables are complex and nonlinear. It helps identify disease markers and predict outcomes based on these complex datasets [112-114]. The diverse range of ML methods used in RWD for disease prediction and management demonstrates the adaptability of these techniques in clinical practice. From interpretable models such as LR and DTs to more complex methods such as DNN and XGBoost, each ML technique uniquely enhances predictive capabilities. These methods enable health care providers to make more accurate, data-driven decisions, ultimately improving patient outcomes and advancing personalized medicine.

Distribution of Diseases, Study Types, and RWE Sources in ML Applications

The distribution of disease types in studies using ML for disease prediction and management revealed a strong emphasis on cardiovascular diseases, with 19 (33%) of the 57 studies focusing on various conditions within this category (Table 3).

Table 3. Disease categories and the studies included in this review (N=57).
DiseasesStudies, n (%)
Cardiovascular diseases19 (33)
Cancers and tumors9 (16)
Neurological diseases6 (11)
Infectious diseases5 (9)
Metabolic diseases5 (9)
Spinal diseases3 (5)
Gastrointestinal diseases2 (4)
Ocular diseases2 (4)
Respiratory diseases2 (4)
Other diseases4 (7)

This high representation could be attributed to the multifactorial and complex nature of cardiovascular diseases, which often involve a combination of genetic, environmental, and lifestyle factors. Conditions such as atrial fibrillation, heart transplantation, and peripheral artery disease were prominent in these studies, where advanced ML models were used to enhance predictive accuracy and improve patient management. For instance, studies on heart transplantation and atrial fibrillation highlighted the potential of ML algorithms in survival prediction and early disease detection. A work demonstrated that ensemble models, combining RF, DNN, and adaptive boosting, significantly outperformed traditional LR for predicting 1-year survival rates after orthotopic heart transplantation, with an area under the receiver operating characteristic curve of 0.764 [25]. Meanwhile, another study explored the use of the Future Innovations in Novel Detection for Atrial Fibrillation ML algorithm to identify undiagnosed atrial fibrillation using data from EHRs, aiming to improve early detection and intervention. In addition, studies on peripheral artery disease and atrial fibrillation in older adults underscored the utility of ML models in survival prediction and risk assessment [26]. A study developed a predictive model for amputation-free survival after the revascularization process, with the random survival forest model achieving the highest accuracy in predicting long-term outcomes [28]. Similarly, another study used various ML methods, including DTs and RFs, to predict new-onset atrial fibrillation in older adults, achieving high specificity and performance, particularly with the RF model [32]. Furthermore, the use of ML in acute ischemic stroke, including studies by Sheth et al [30] and Hilbert et al [35], illustrated the growing role of deep learning techniques, such as convolutional neural networks and residual neural networks, in improving diagnostic accuracy and predicting patient outcomes [35]. These advancements in ML could potentially revolutionize clinical decision-making and treatment selection, especially for conditions such as stroke, where rapid and accurate assessment is critical.

A significant (9/57, 16%) portion of studies also targeted cancers and tumors, which were often characterized by their heterogeneity and the need for personalized treatment plans. ML algorithms, such as RF and SVM, enhanced early cancer detection, predicted disease recurrence, and assessed the effectiveness of different treatment protocols, demonstrating great potential in oncology settings. One key area of focus was the prediction of disease outcomes. For instance, a study developed a series of ML models to predict the 5-year survival rate for patients with Ewing sarcoma, a rare type of cancer. Using data from 2332 patients, including various algorithms such as boosted DTs, SVMs, RFs, and neural networks, the study found that the RF method performed best, with impressive sensitivity and specificity. This model is now accessible via a web-based application, providing clinicians a valuable tool for assessing survival probabilities for patients with Ewing sarcoma [36]. Another study used a predictive ML model to improve surgical scheduling in cancer surgeries, specifically for colorectal and gynecologic cancers. The research used gradient boosting and LR techniques to predict surgical durations, reducing operational inefficiencies such as patient wait times and optimizing the use of surgical resources, thereby demonstrating how ML could streamline health care operations while maintaining treatment quality [38]. Furthermore, in survival prediction, a study used an RF model to develop a gene signature that predicted the response of patients with gastric cancer to paclitaxel treatment. Their model, which identified a 19-gene signature, enabled the classification of patients into those who would benefit from the treatment, providing a novel approach to personalized cancer therapy [48].

The studies focusing on neurological diseases, including conditions such as epilepsy, cognitive impairment, and schizophrenia, highlighted the significant impact of ML in improving diagnosis, treatment prediction, and health care outcomes. These studies underscored the potential of ML to personalize patient care and optimize clinical decision-making. For instance, a study investigated the application of NLP embedded in EHR to automate alerts for pediatric patients with epilepsy. This ML-driven clinical decision support system successfully increased referrals for epilepsy surgery, with a marked improvement in presurgical evaluation rates and even higher rates of actual surgery, illustrating how NLP-based interventions could influence health care outcomes by improving referral efficiency and treatment access [24]. Similarly, another study focused on the use of XGBoost, an ML algorithm, to identify older patients in the emergency department at high risk for cognitive impairment. This predictive model, using EHR data, demonstrated high sensitivity and specificity, with the potential to reduce the need for in-person screenings and prioritize patients at high risk. By streamlining screening processes, this approach could enhance the detection of cognitive impairments in older adults, potentially leading to earlier interventions and better management of conditions such as dementia [27]. In schizophrenia, a study developed a nonlinear SVM model to predict treatment outcomes for patients with first-episode psychosis. The model was trained on pretreatment patient-reported data and successfully predicted poor versus good treatment outcomes, thus supporting clinical decision-making in terms of which treatments might be more effective for certain patients and identifying those at risk for nonadherence or poor prognosis [37]. These studies collectively demonstrated how ML methods, such as NLP, XGBoost, and RF, were revolutionizing the management of neurological diseases. By enabling early detection, better prediction of disease outcomes, and more informed decision-making, these tools offered substantial improvements in both clinical and health care settings. Infectious diseases, metabolic diseases, spinal diseases, gastrointestinal diseases, ocular diseases, and respiratory diseases each had a smaller but notable presence in the studies (n≤5). These applications generally focused on disease prediction, early diagnosis, and treatment optimization. ML models such as XGBoost and DNN were used to predict disease onset, assess risks, and improve patient management in these areas.

The data revealed that EHRs were the most frequently used form of RWE, accounting for 68% (39/57) of the studies. EHRs were a rich source of patient data, providing comprehensive records of patient health status, diagnoses, treatments, and outcomes over time. This made EHRs particularly valuable for studies that required large-scale data to identify patterns, trends, and correlations in real-world clinical settings. The next most commonly used type of RWE was patient registries, which were used in 26% (15/57) of the studies. Patient registries typically collect data on specific patient populations with particular diseases or conditions, allowing for longitudinal tracking of disease progression and treatment outcomes. Wearable devices were the least used form of RWE, accounting for 7% (4/57) of the studies. Wearables were increasingly being used to collect real-time health data, including vital signs and activity levels, which could provide valuable insights into patients’ health status outside of clinical environments. This distribution highlighted the dominance of EHR as the primary data source in these studies, reflecting its accessibility and broad applicability in health care research.

Disease prediction emerged as the most widely studied area, represented by 35% (20/57) of the studies. This suggested a strong emphasis on using ML and data analytics to predict the onset, progression, or outcomes of various diseases. The next most studied area was decision-making, with 25% (14/57) of the studies that underscored the growing interest in leveraging data-driven insights to inform clinical decisions and treatment strategies. Health care outcomes, such as quality of life, recovery rates, and adverse events, were the focus of 23% (13/57) of the studies, reflecting the importance of understanding how diseases and treatments affect patients’ overall well-being. Survival prediction, accounting for 19% (11/57) of the studies, was another critical area of research, particularly in oncology and chronic diseases, where predicting patient survival and the effectiveness of interventions could guide clinical decision-making. This distribution indicated that disease prediction and decision-making were central to applying RWE in health care, with a significant focus on improving patient outcomes and guiding treatment strategies.


Principal Findings

The findings of this study underscore the growing application of ML techniques in RWD for disease prediction and management. The results reveal that ML methods, particularly ensemble models such as RF, play a crucial role in enhancing prediction accuracy and addressing the complexities of large and high-dimensional datasets common in health care. Among the top ML methods used, RF was the most widely used, featured in 42% (24/57) of the studies, showcasing its adaptability to a variety of clinical datasets such as EHRs and patient registries. RF’s ability to handle missing data, its resistance to overfitting, and its effectiveness in managing imbalanced datasets made it a powerful tool in predicting disease outcomes [115,116], such as survival rates and complications in cardiovascular diseases and cancer. Regarding disease types, cardiovascular diseases dominated the studies, with 33% (19/57) of the studies dedicated to predicting outcomes related to heart transplantation, atrial fibrillation, and peripheral artery disease. This concentration is likely attributed to the critical need for predictive tools in the early diagnosis and management of these conditions, which account for a significant burden on health care systems globally [117]. ML applications, such as DNN and random survival forests, have been shown to improve the accuracy of survival predictions, assess treatment responses, and enhance patient stratification. In addition, the study highlights the increasing application of ML in predicting conditions such as cancers, neurological disorders, and infectious diseases. These findings align with the broader trend of using RWD to bridge the gap between clinical trials and actual patient care by making predictions based on RWD sources, such as EHRs and wearable devices. As evidenced in the studies reviewed, ML techniques can process vast amounts of medical data from various sources, facilitating early detection, timely intervention, and improved management of chronic conditions. Furthermore, these advancements in ML applications are subject to increasing regulatory oversight. Agencies such as the US FDA and the EMA are actively exploring frameworks for the approval and regulation of ML-driven tools in health care. These regulations aim to ensure ML models’ safety, efficacy, and transparency, especially in real-world applications where data variability and model interpretability remain key concerns. As regulatory bodies continue to define standards for using RWD and ML in clinical settings, ensuring compliance with FDA and EMA guidelines will be essential for the broader adoption and integration of these technologies into clinical practice.

Comparison With Prior Work

This systematic review aligns with and extends several recent literature reviews that have explored the application of ML to RWD in health care. Previous studies have highlighted the potential of ML models to transform health care by improving disease prediction and patient management [118,119]. However, our review emphasizes a broader scope by including a wide variety of disease types, from cardiovascular diseases and cancer to neurological and infectious diseases, reflecting the growing versatility of ML tools in clinical settings. A notable comparison can be made with a study that focused on EHRs as the primary data source for ML models, conducted by Miotto et al [120]. While their review identified the challenges associated with EHR-based studies, such as data sparsity and heterogeneity, our study similarly acknowledges these limitations but also expands the discussion to include wearable devices and patient registries as additional data sources. These emerging data sources provide a more complete picture of the patient’s health status, significantly improving model performance and enabling better patient monitoring in real-world settings. Another key comparison is with a review that focused on ML’s role in health care decision-making and its integration into clinical workflows [121]. While the study by Beam and Kohane [121] explored various ML algorithms in health care, our review places a stronger emphasis on the role of ensemble models, such as RF, and their applicability across diverse health care datasets. One of the key contributions of our review, which sets it apart from previous works, is the focus on regulatory challenges associated with the deployment of ML models in clinical practice. While other reviews have discussed technical aspects of ML, we specifically address the urgent need for clearer regulatory frameworks from authorities such as the FDA and EMA to ensure the safe and effective approval of ML models in health care. This is an area that has received limited attention in previous reviews but is critical as health care systems begin to rely more heavily on automated systems for clinical decision-making. Overall, this review builds upon the foundations established by previous literature, offering an updated and comprehensive analysis that incorporates new data sources, encompasses a broader range of diseases, and addresses the challenges of regulatory approval and model interpretability in the context of ML in health care.

Limitations

This review has several limitations. First, while a comprehensive search strategy was used, it is possible that some relevant studies were missed, particularly those that did not explicitly use the selected keywords or were indexed in databases not included in this search. In addition, this review was limited to English-language publications, which may have excluded relevant studies published in other languages. Another potential limitation is the application of strict inclusion and exclusion criteria, which, while ensuring the relevance and quality of the included studies, may have led to the omission of some studies that could have provided valuable insights. For example, studies with limited methodological details or those focusing primarily on deep learning applications were often excluded due to insufficient validation or performance comparisons. Future research could consider broader inclusion criteria to capture a wider range of studies. Furthermore, the variability in study designs and data sources posed challenges in synthesizing findings across different studies. The included studies used different types of RWD, ranging from EHRs and patient registries to imaging and wearable sensor data. This heterogeneity complicates direct comparisons and generalizability. Finally, while efforts were made to minimize bias through independent study screening by 2 reviewers, inherent biases in study selection and data extraction may still exist. The reliance on published literature also introduces publication bias, as studies with negative or inconclusive results may be underrepresented. Future work could integrate unpublished data sources or conduct meta-analyses to provide a more comprehensive assessment of ML applications in RWD.

Future Work

While this systematic review aimed for comprehensive coverage by using a broad search strategy across multiple databases and gray literature sources, the large disparity between the number of initially retrieved studies and those meeting the final inclusion criteria highlights an important area for improvement in future research. A more focused, topic-specific keyword strategy, combined with the application of advanced database filters, could increase the precision of future searches by limiting the retrieval of irrelevant studies. In addition, integrating artificial intelligence–assisted search tools and NLP algorithms might further enhance the efficiency of systematic literature reviews by streamlining the identification of eligible studies based on more nuanced criteria. Future reviews may also benefit from targeting more narrowly defined subtopics within the broader field of ML applications in RWD, such as specific disease domains, ML model types, or clinical trial phases. These refinements would likely improve the overall relevance and manageability of retrieved records, ensuring a more efficient screening process and focused synthesis of findings.

Conclusions

In conclusion, this review highlights the transformative potential of integrating ML techniques with RWD in health care, specifically for disease prediction and management. The use of advanced ML models, such as ensemble methods and deep learning, has demonstrated the ability to enhance predictive accuracy, improve patient stratification, and facilitate more personalized and proactive health care. These advancements are poised to significantly impact clinical decision-making, enabling earlier diagnoses, optimized treatment strategies, and efficient resource allocation. However, despite these promising developments, several challenges remain. Issues related to data quality, generalizability across diverse populations, and the interpretability of complex ML models must be addressed to ensure their effective and widespread application. The lack of transparency in some ML algorithms, which often function as “black boxes,” remains a significant barrier to their integration into clinical workflows. Improving the explainability of these models will be crucial in gaining the trust of health care professionals and enhancing the clinical utility of ML predictions. In addition, regulatory frameworks for ML in health care are still evolving, with clear guidelines needed from regulatory bodies such as the FDA and EMA. This will help ensure that ML models meet safety standards and are deployed in clinical settings in a manner that benefits both health care providers and patients. Furthermore, as health care data becomes increasingly heterogeneous, with sources ranging from EHRs to wearable devices and patient registries, strategies for addressing data inconsistencies and ensuring data quality will be essential. Looking ahead, future research should focus on improving the robustness, transparency, and generalizability of ML models, particularly for underrepresented diseases and diverse patient populations. Establishing ethical and regulatory standards for the use of ML in clinical practice will be crucial for fostering public trust and ensuring equitable access to these innovations. Collaboration among clinicians, data scientists, and policy makers will be key to overcoming these challenges, with the ultimate goal of ensuring that ML-driven advancements in health care lead to improved health outcomes, better care delivery, and more equitable health care systems for all.

Acknowledgments

This work was supported and funded by the Deanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University (IMSIU) (grant number IMSIU-DDRSP2501).

Data Availability

All data generated or analyzed during this study are included in this published article and its supplementary information files.

Authors' Contributions

NHA and DD conceptualized the study, developed the methodology, and wrote the original draft of the manuscript. HFK contributed to data collection, analysis, and interpretation. NHA, DD, and NA provided critical revisions and approved the final version of the manuscript. NA supervised the project study. All authors reviewed and approved the manuscript for publication.

Conflicts of Interest

None declared.

Multimedia Appendix 1

Full search strategy.

DOCX File , 20 KB

Multimedia Appendix 2

PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) checklist.

PDF File (Adobe PDF File), 492 KB

Multimedia Appendix 3

A dataset of selected studies, including details such as the year of publication, authors’ names, study title, database, diseases or medical conditions, disease category, study type, real-world evidence type, machine learning methods used, and key findings.

XLSX File (Microsoft Excel File), 41 KB

  1. Liu F, Panagiotakos D. Real-world data: a brief review of the methods, applications, challenges and opportunities. BMC Med Res Methodol. Nov 05, 2022;22(1):287. [FREE Full text] [CrossRef] [Medline]
  2. Dang A. Real-world evidence: a primer. Pharmaceut Med. Jan 2023;37(1):25-36. [FREE Full text] [CrossRef] [Medline]
  3. Klonoff DC. The expanding role of real-world evidence trials in health care decision making. J Diabetes Sci Technol. Jan 2020;14(1):174-179. [FREE Full text] [CrossRef] [Medline]
  4. Junaid SB, Imam AA, Balogun AO, De Silva LC, Surakat YA, Kumar G, et al. Recent advancements in emerging technologies for healthcare management systems: a survey. Healthcare (Basel). Oct 03, 2022;10(10):1940. [FREE Full text] [CrossRef] [Medline]
  5. Williamson SM, Prybutok V. Balancing privacy and progress: a review of privacy challenges, systemic oversight, and patient perceptions in AI-driven healthcare. Appl Sci. Jan 12, 2024;14(2):675. [CrossRef]
  6. Derman BA, Belli AJ, Battiwalla M, Hamadani M, Kansagra A, Lazarus HM, et al. Reality check: real-world evidence to support therapeutic development in hematologic malignancies. Blood Rev. May 2022;53:100913. [FREE Full text] [CrossRef] [Medline]
  7. Rane NL, Paramesha M, Choudhary SP, Rane J. Machine learning and deep learning for big data analytics: a review of methods and applications. Partn Univers Int Res J. 2024;2(3):172-197. [FREE Full text] [CrossRef]
  8. Saeed S, Ahmed S, Joseph S. Machine learning in the big data age: advancements, challenges, and future prospects. ResearchGate. URL: https:/​/www.​researchgate.net/​publication/​377438052_Machine_Learning_in_the_Big_Data_Age_Advancements_Challenges_and_Future_Prospects [accessed 2025-05-29]
  9. Javaid M, Haleem A, Pratap Singh R, Suman R, Rab S. Significance of machine learning in healthcare: features, pillars and applications. Int J Intell Netw. 2022;3:58-73. [CrossRef]
  10. Ogrizović M, Drašković D, Bojić D. Quality assurance strategies for machine learning applications in big data analytics: an overview. J Big Data. Oct 30, 2024;11:156. [CrossRef]
  11. Badalotti D, Agrawal A, Pensato U, Angelotti G, Marcheselli S. Development of a Natural Language Processing (NLP) model to automatically extract clinical data from electronic health records: results from an Italian comprehensive stroke center. Int J Med Inform. Dec 2024;192:105626. [FREE Full text] [CrossRef] [Medline]
  12. Turchin A, Florez Builes LF. Using natural language processing to measure and improve quality of diabetes care: a systematic review. J Diabetes Sci Technol. May 2021;15(3):553-560. [FREE Full text] [CrossRef] [Medline]
  13. Dominguez-Catena I, Paternain D, Jurio A, Galar M. Less can be more: representational vs. stereotypical gender bias in facial expression recognition. Prog Artif Intell. Oct 14, 2024;14:11-31. [CrossRef]
  14. Ferrara C, Sellitto G, Ferrucci F, Palomba F, De Lucia A. Fairness-aware machine learning engineering: how far are we? Empir Softw Eng. 2024;29(1):9. [FREE Full text] [CrossRef] [Medline]
  15. Hassija V, Chamola V, Mahapatra A, Singal A, Goel D, Huang K, et al. Interpreting black-box models: a review on explainable artificial intelligence. Cogn Comput. Aug 24, 2023;16:45-74. [CrossRef]
  16. Ali S, Abuhmed T, El-Sappagh S, Muhammad K, Alonso-Moral JM, Confalonieri R, et al. Explainable Artificial Intelligence (XAI): what we know and what is left to attain Trustworthy Artificial Intelligence. Inf Fusion. Nov 2023;99:101805. [CrossRef]
  17. El Mestari SZ, Lenzini G, Demirci H. Preserving data privacy in machine learning systems. Comput Secur. Feb 2024;137:103605. [CrossRef]
  18. Lu H, Alhaskawi A, Dong Y, Zou X, Zhou H, Ezzi SH, et al. Patient autonomy in medical education: navigating ethical challenges in the age of artificial intelligence. Inquiry. Sep 18, 2024;61:469580241266364. [FREE Full text] [CrossRef] [Medline]
  19. Tucker K, Branson J, Dilleen M, Hollis S, Loughlin P, Nixon MJ, et al. Protecting patient privacy when sharing patient-level data from clinical trials. BMC Med Res Methodol. Jul 08, 2016;16 Suppl 1(Suppl 1):77. [FREE Full text] [CrossRef] [Medline]
  20. Franklin G, Stephens R, Piracha M, Tiosano S, Lehouillier F, Koppel R, et al. The sociodemographic biases in machine learning algorithms: a biomedical informatics perspective. Life (Basel). May 21, 2024;14(6):652. [FREE Full text] [CrossRef] [Medline]
  21. Considerations for the use of real-world data and real-world evidence to support regulatory decision-making for drug and biological products. U.S. Food & Drug Administration. Aug 2023. URL: https:/​/www.​fda.gov/​regulatory-information/​search-fda-guidance-documents/​considerations-use-real-world-data-and-real-world-evidence-support-regulatory-decision-making-drug [accessed 2025-06-05]
  22. Moher D, Liberati A, Tetzlaff J, Altman DG, PRISMA Group. Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement. PLoS Med. Jul 21, 2009;6(7):e1000097. [FREE Full text] [CrossRef] [Medline]
  23. Alotaiq N, Dermawan D. Advancements in virtual bioequivalence: a systematic review of computational methods and regulatory perspectives in the pharmaceutical industry. Pharmaceutics. Nov 03, 2024;16(11):1414. [FREE Full text] [CrossRef] [Medline]
  24. Wissel BD, Greiner HM, Glauser TA, Mangano FT, Holland-Bouley KD, Zhang N, et al. Automated, machine learning-based alerts increase epilepsy surgery referrals: a randomized controlled trial. Epilepsia. Jul 27, 2023;64(7):1791-1799. [FREE Full text] [CrossRef] [Medline]
  25. Ayers B, Sandholm T, Gosev I, Prasad S, Kilic A. Using machine learning to improve survival prediction after heart transplantation. J Card Surg. Nov 19, 2021;36(11):4113-4120. [CrossRef] [Medline]
  26. Nadarajah R, Wahab A, Reynolds C, Raveendra K, Askham D, Dawson R, et al. Future Innovations in Novel Detection for Atrial Fibrillation (FIND-AF): pilot study of an electronic health record machine learning algorithm-guided intervention to identify undiagnosed atrial fibrillation. Open Heart. Sep 30, 2023;10(2):e002447. [FREE Full text] [CrossRef] [Medline]
  27. Yadgir SR, Engstrom C, Jacobsohn GC, Green RK, Jones CM, Cushman JT, et al. Machine learning-assisted screening for cognitive impairment in the emergency department. J Am Geriatr Soc. Mar 2022;70(3):831-837. [FREE Full text] [CrossRef] [Medline]
  28. Liu Y, Xue J, Jiang J. Application of machine learning algorithms in electronic medical records to predict amputation-free survival after first revascularization in patients with peripheral artery disease. Int J Cardiol. Jul 15, 2023;383:175-184. [CrossRef] [Medline]
  29. Hill NR, Groves L, Dickerson C, Boyce R, Lawton S, Hurst M, et al. Identification of undiagnosed atrial fibrillation using a machine learning risk prediction algorithm and diagnostic testing (PULsE-AI) in primary care: cost-effectiveness of a screening strategy evaluated in a randomized controlled trial in England. J Med Econ. Aug 03, 2022;25(1):974-983. [FREE Full text] [CrossRef] [Medline]
  30. Sheth SA, Lopez-Rivera V, Barman A, Grotta JC, Yoo AJ, Lee S, et al. Machine learning-enabled automated determination of acute ischemic core from computed tomography angiography. Stroke. Nov 2019;50(11):3093-3100. [CrossRef] [Medline]
  31. Barton C, Chettipally U, Zhou Y, Jiang Z, Lynn-Palevsky A, Le S, et al. Evaluation of a machine learning algorithm for up to 48-hour advance prediction of sepsis using six vital signs. Comput Biol Med. Jun 2019;109:79-84. [FREE Full text] [CrossRef] [Medline]
  32. Kao YT, Huang CY, Fang YA, Liu JC, Chang TH. Machine learning-based prediction of atrial fibrillation risk using electronic medical records in older aged patients. Am J Cardiol. Jul 01, 2023;198:56-63. [CrossRef] [Medline]
  33. Kim M, Kang Y, You SC, Park HD, Lee SS, Kim TH, et al. Artificial intelligence predicts clinically relevant atrial high-rate episodes in patients with cardiac implantable electronic devices. Sci Rep. Jan 07, 2022;12(1):37. [FREE Full text] [CrossRef] [Medline]
  34. Park HB, Lee J, Hong Y, Byungchang S, Kim W, Lee BK, et al. Risk factors based vessel-specific prediction for stages of coronary artery disease using Bayesian quantile regression machine learning method: results from the PARADIGM registry. Clin Cardiol. Mar 2023;46(3):320-327. [FREE Full text] [CrossRef] [Medline]
  35. Hilbert A, Ramos LA, van Os HJ, Olabarriaga SD, Tolhuisen ML, Wermer MJ, et al. Data-efficient deep learning of radiological image data for outcome prediction after endovascular treatment of patients with acute ischemic stroke. Comput Biol Med. Dec 2019;115:103516. [CrossRef] [Medline]
  36. Chen W, Zhou C, Yan Z, Chen H, Lin K, Zheng Z, et al. Using machine learning techniques predicts prognosis of patients with Ewing sarcoma. J Orthop Res. Nov 24, 2021;39(11):2519-2527. [FREE Full text] [CrossRef] [Medline]
  37. Koutsouleris N, Kahn RS, Chekroud AM, Leucht S, Falkai P, Wobrock T, et al. Multisite prediction of 4-week and 52-week treatment outcomes in patients with first-episode psychosis: a machine learning approach. Lancet Psychiatry. Oct 2016;3(10):935-946. [CrossRef]
  38. Strömblad CT, Baxter-King RG, Meisami A, Yee SJ, Levine MR, Ostrovsky A, et al. Effect of a predictive model on planned surgical duration accuracy, patient wait time, and use of presurgical resources: a randomized clinical trial. JAMA Surg. Apr 01, 2021;156(4):315-321. [FREE Full text] [CrossRef] [Medline]
  39. Wang SV, Rogers JR, Jin Y, DeiCicchi D, Dejene S, Connors JM, et al. Stepped-wedge randomised trial to evaluate population health intervention designed to increase appropriate anticoagulation in patients with atrial fibrillation. BMJ Qual Saf. Oct 26, 2019;28(10):835-842. [FREE Full text] [CrossRef] [Medline]
  40. Tan TH, Hsu CC, Chen CJ, Hsu SL, Liu TL, Lin HJ, et al. Predicting outcomes in older ED patients with influenza in real time using a big data-driven and machine learning approach to the hospital information system. BMC Geriatr. Apr 27, 2021;21(1):280. [FREE Full text] [CrossRef] [Medline]
  41. Goerigk S, Hilbert S, Jobst A, Falkai P, Bühner M, Stachl C, et al. Predicting instructed simulation and dissimulation when screening for depressive symptoms. Eur Arch Psychiatry Clin Neurosci. Mar 12, 2020;270(2):153-168. [CrossRef] [Medline]
  42. Kijpaisalratana N, Saoraya J, Nhuboonkaew P, Vongkulbhisan K, Musikatavorn K. Real-time machine learning-assisted sepsis alert enhances the timeliness of antibiotic administration and diagnostic accuracy in emergency department patients with sepsis: a cluster-randomized trial. Intern Emerg Med. Aug 21, 2024;19(5):1415-1424. [CrossRef] [Medline]
  43. Sharma A, Sun JL, Lokhnygina Y, Roe MT, Ahmad T, Desai NR, et al. Patient phenotypes, cardiovascular risk, and ezetimibe treatment in patients after acute coronary syndromes (from IMPROVE-IT). Am J Cardiol. Apr 15, 2019;123(8):1193-1201. [CrossRef] [Medline]
  44. Singhal L, Garg Y, Yang P, Tabaie A, Wong AI, Mohammed A, et al. eARDS: A multi-center validation of an interpretable machine learning algorithm of early onset Acute Respiratory Distress Syndrome (ARDS) among critically ill adults with COVID-19. PLoS One. Sep 24, 2021;16(9):e0257056. [FREE Full text] [CrossRef] [Medline]
  45. Kanchanatawan B, Tangwongchai S, Supasitthumrong T, Sriswasdi S, Maes M. Episodic memory and delayed recall are significantly more impaired in younger patients with deficit schizophrenia than in elderly patients with amnestic mild cognitive impairment. PLoS One. May 15, 2018;13(5):e0197004. [FREE Full text] [CrossRef] [Medline]
  46. Huang J, Jin W, Duan X, Liu X, Shu T, Fu L, et al. Twenty-eight-day in-hospital mortality prediction for elderly patients with ischemic stroke in the intensive care unit: interpretable machine learning models. Front Public Health. Jan 12, 2022;10:1086339. [FREE Full text] [CrossRef] [Medline]
  47. She H, Du Y, Du Y, Tan L, Yang S, Luo X, et al. Metabolomics and machine learning approaches for diagnostic and prognostic biomarkers screening in sepsis. BMC Anesthesiol. Nov 09, 2023;23(1):367. [FREE Full text] [CrossRef] [Medline]
  48. Sundar R, Barr Kumarakulasinghe N, Huak Chan Y, Yoshida K, Yoshikawa T, Miyagi Y, et al. Machine-learning model derived gene signature predictive of paclitaxel survival benefit in gastric cancer: results from the randomised phase III SAMIT trial. Gut. Apr 12, 2022;71(4):676-685. [FREE Full text] [CrossRef] [Medline]
  49. Alaa AM, Bolton T, Di Angelantonio E, Rudd JH, van der Schaar M. Cardiovascular disease risk prediction using automated machine learning: a prospective study of 423,604 UK Biobank participants. PLoS One. May 15, 2019;14(5):e0213653. [FREE Full text] [CrossRef] [Medline]
  50. Azimi P, Mohammadi HR, Benzel EC, Shahzadi S, Azhari S. Use of artificial neural networks to decision making in patients with lumbar spinal canal stenosis. J Neurosurg Sci. Dec 2017;61(6):603-611. [CrossRef]
  51. Baxter SL, Marks C, Kuo TT, Ohno-Machado L, Weinreb RN. Machine learning-based predictive modeling of surgical intervention in glaucoma using systemic data from electronic health records. Am J Ophthalmol. Dec 2019;208:30-40. [FREE Full text] [CrossRef] [Medline]
  52. Anderson JP, Parikh JR, Shenfeld DK, Ivanov V, Marks C, Church BW, et al. Reverse engineering and evaluation of prediction models for progression to type 2 diabetes: an application of machine learning using electronic health records. J Diabetes Sci Technol. Dec 20, 2015;10(1):6-18. [FREE Full text] [CrossRef] [Medline]
  53. Bannister CA, Halcox JP, Currie CJ, Preece A, Spasić I. A genetic programming approach to development of clinical prediction models: a case study in symptomatic cardiovascular disease. PLoS One. Sep 4, 2018;13(9):e0202685. [FREE Full text] [CrossRef] [Medline]
  54. Scheer JK, Smith JS, Schwab F, Lafage V, Shaffrey CI, Bess S, et al. Development of a preoperative predictive model for major complications following adult spinal deformity surgery. J Neurosurg Spine. Jun 2017;26(6):736-743. [CrossRef] [Medline]
  55. Rau HH, Hsu CY, Lin YA, Atique S, Fuad A, Wei LM, et al. Development of a web-based liver cancer prediction model for type II diabetes patients by using an artificial neural network. Comput Methods Programs Biomed. Mar 2016;125:58-65. [CrossRef] [Medline]
  56. Ramezankhani A, Hadavandi E, Pournik O, Shahrabi J, Azizi F, Hadaegh F. Decision tree-based modelling for identification of potential interactions between type 2 diabetes risk factors: a decade follow-up in a Middle East prospective cohort study. BMJ Open. Dec 01, 2016;6(12):e013336. [FREE Full text] [CrossRef] [Medline]
  57. Pei D, Zhang C, Quan Y, Guo Q. Identification of potential type II diabetes in a Chinese population with a sensitive decision tree approach. J Diabetes Res. Jan 22, 2019;2019:4248218. [FREE Full text] [CrossRef] [Medline]
  58. Oviedo S, Contreras I, Quirós C, Giménez M, Conget I, Vehi J. Risk-based postprandial hypoglycemia forecasting using supervised learning. Int J Med Inform. Jun 2019;126:1-8. [CrossRef] [Medline]
  59. Mubeen AM, Asaei A, Bachman AH, Sidtis JJ, Ardekani BA, Alzheimer's Disease Neuroimaging Initiative. A six-month longitudinal evaluation significantly improves accuracy of predicting incipient Alzheimer's disease in mild cognitive impairment. J Neuroradiol. Oct 2017;44(6):381-387. [CrossRef] [Medline]
  60. Lopez-de-Andres A, Hernandez-Barrera V, Lopez R, Martin-Junco P, Jimenez-Trujillo I, Alvaro-Meca A, et al. Predictors of in-hospital mortality following major lower extremity amputations in type 2 diabetic patients using artificial neural networks. BMC Med Res Methodol. Nov 22, 2016;16(1):160. [FREE Full text] [CrossRef] [Medline]
  61. Kwon JM, Jeon KH, Kim HM, Kim MJ, Lim S, Kim KH, et al. Deep-learning-based out-of-hospital cardiac arrest prognostic system to predict clinical outcomes. Resuscitation. Jun 2019;139:84-91. [CrossRef] [Medline]
  62. Kim I, Choi HJ, Ryu JM, Lee SK, Yu JH, Kim SW, et al. A predictive model for high/low risk group according to oncotype DX recurrence score using machine learning. Eur J Surg Oncol. Feb 2019;45(2):134-140. [CrossRef] [Medline]
  63. Khanji C, Lalonde L, Bareil C, Lussier MT, Perreault S, Schnitzer ME. Lasso regression for the prediction of intermediate outcomes related to cardiovascular disease prevention using the TRANSIT quality indicators. Med Care. Jan 2019;57(1):63-72. [CrossRef] [Medline]
  64. Karhade AV, Ogink PT, Thio QC, Cha TD, Gormley WB, Hershman SH, et al. Development of machine learning algorithms for prediction of prolonged opioid prescription after surgery for lumbar disc herniation. Spine J. Nov 2019;19(11):1764-1771. [CrossRef] [Medline]
  65. Jovanovic P, Salkic NN, Zerem E. Artificial neural network predicts the need for therapeutic ERCP in patients with suspected choledocholithiasis. Gastrointest Endosc. Aug 2014;80(2):260-268. [CrossRef] [Medline]
  66. Kang AR, Lee J, Jung W, Lee M, Park SY, Woo J, et al. Development of a prediction model for hypotension after induction of anesthesia using machine learning. PLoS One. Apr 16, 2020;15(4):e0231172. [FREE Full text] [CrossRef] [Medline]
  67. Isma'eel HA, Sakr GE, Serhan M, Lamaa N, Hakim A, Cremer PC, et al. Artificial neural network-based model enhances risk stratification and reduces non-invasive cardiac stress imaging compared to Diamond-Forrester and Morise risk assessment models: a prospective study. J Nucl Cardiol. Oct 2018;25(5):1601-1609. [CrossRef] [Medline]
  68. Hill NR, Ayoubkhani D, McEwan P, Sugrue DM, Farooqui U, Lister S, et al. Predicting atrial fibrillation in primary care using machine learning. PLoS One. Nov 1, 2019;14(11):e0224582. [FREE Full text] [CrossRef] [Medline]
  69. Dong Y, Xu L, Fan Y, Xiang P, Gao X, Chen Y, et al. A novel surgical predictive model for Chinese Crohn's disease patients. Medicine (Baltimore). Nov 2019;98(46):e17510. [FREE Full text] [CrossRef] [Medline]
  70. Bowman A, Rudolfer S, Weller P, Bland JD. A prognostic model for the patient-reported outcome of surgical treatment of carpal tunnel syndrome. Muscle Nerve. Dec 15, 2018;58(6):784-789. [CrossRef] [Medline]
  71. Bertsimas D, Dunn J, Pawlowski C, Silberholz J, Weinstein A, Zhuo YD, et al. Applied informatics decision support tool for mortality predictions in patients with cancer. JCO Clin Cancer Inform. Dec 2018;2:1-11. [CrossRef]
  72. Manz CR, Parikh RB, Small DS, Evans CN, Chivers C, Regli SH, et al. Effect of integrating machine learning mortality estimates with behavioral nudges to clinicians on serious illness conversations among patients with cancer: a stepped-wedge cluster randomized clinical trial. JAMA Oncol. Dec 01, 2020;6(12):e204759. [FREE Full text] [CrossRef] [Medline]
  73. Tian D, Yan HJ, Huang H, Zuo YJ, Liu MZ, Zhao J, et al. Machine learning-based prognostic model for patients after lung transplantation. JAMA Netw Open. May 01, 2023;6(5):e2312022. [FREE Full text] [CrossRef] [Medline]
  74. Li E, Manz C, Liu M, Chen J, Chivers C, Braun J, et al. Oncologist phenotypes and associations with response to a machine learning-based intervention to increase advance care planning: secondary analysis of a randomized clinical trial. PLoS One. May 27, 2022;17(5):e0267012. [FREE Full text] [CrossRef] [Medline]
  75. Tedeschi SK, Cai T, He Z, Ahuja Y, Hong C, Yates KA, et al. Classifying pseudogout using machine learning approaches with electronic health record data. Arthritis Care Res (Hoboken). Mar 25, 2021;73(3):442-448. [FREE Full text] [CrossRef] [Medline]
  76. Ambwani G, Cohen A, Estévez M, Singh N, Adamson B, Nussbaum N, et al. PPM8 A machine learning model for cancer biomarker identification in electronic health records. Value Health. May 2019;22:S334. [CrossRef]
  77. Jorge A, Castro VM, Barnado A, Gainer V, Hong C, Cai T, et al. Identifying lupus patients in electronic health records: development and validation of machine learning algorithms and application of rule-based algorithms. Semin Arthritis Rheum. Aug 2019;49(1):84-90. [FREE Full text] [CrossRef] [Medline]
  78. Shimabukuro DW, Barton CW, Feldman MD, Mataraso SJ, Das R. Effect of a machine learning-based severe sepsis prediction algorithm on patient survival and hospital length of stay: a randomised clinical trial. BMJ Open Respir Res. Nov 09, 2017;4(1):e000234. [FREE Full text] [CrossRef] [Medline]
  79. Sarraju A, Ward A, Chung S, Li J, Scheinker D, Rodríguez F. Machine learning approaches improve risk stratification for secondary cardiovascular disease prevention in multiethnic patients. Open Heart. Oct 19, 2021;8(2):e001802. [FREE Full text] [CrossRef] [Medline]
  80. Ye B, Liu K, Cao S, Sankaridurg P, Li W, Luan M, et al. Discrimination of indoor versus outdoor environmental state with machine learning algorithms in myopia observational studies. J Transl Med. Sep 18, 2019;17(1):314. [FREE Full text] [CrossRef] [Medline]
  81. Sun Z, Wang G, Li P, Wang H, Zhang M, Liang X. An improved random forest based on the classification accuracy and correlation measurement of decision trees. Expert Syst Appl. Mar 2024;237:121549. [CrossRef]
  82. Bandyopadhyay A, Albashayreh A, Zeinali N, Fan W, Gilbertson-White S. Using real-world electronic health record data to predict the development of 12 cancer-related symptoms in the context of multimorbidity. JAMIA Open. Oct 2024;7(3):ooae082. [CrossRef] [Medline]
  83. Bairagade DM, Sharma K, More PR. Enhancing prediction accuracy in healthcare data using advanced DataMining techniques. J Comput Sci Eng. 2024. [FREE Full text]
  84. Ghosh D, Cabrera J. Enriched random forest for high dimensional genomic data. IEEE/ACM Trans Comput Biol Bioinf. Sep 1, 2022;19(5):2817-2828. [CrossRef]
  85. Maalouf M. Logistic regression in data analysis: an overview. Int J Data Anal Tech Strateg. 2011;3(3):281. [CrossRef]
  86. Boateng EY, Abaye DA. A review of the logistic regression model with emphasis on medical research. J Data Anal Inf Process. Nov 2019;07(04):190-207. [CrossRef]
  87. Shipe ME, Deppen SA, Farjah F, Grogan EL. Developing prediction models for clinical use using logistic regression: an overview. J Thorac Dis. Mar 2019;11(Suppl 4):S574-S584. [FREE Full text] [CrossRef] [Medline]
  88. Zabor EC, Reddy CA, Tendulkar RD, Patil S. Logistic regression in clinical studies. Int J Radiat Oncol Biol Phys. Feb 01, 2022;112(2):271-277. [CrossRef] [Medline]
  89. Guido R, Ferrisi S, Lofaro D, Conforti D. An overview on the advancements of support vector machine models in healthcare applications: a review. Information. Apr 19, 2024;15(4):235. [CrossRef]
  90. Huang S, Cai N, Pacheco PP, Narrandes S, Wang Y, Xu W. Applications of support vector machine (SVM) learning in cancer genomics. Cancer Genomics Proteomics. Jan 2, 2018;15(1):41-51. [FREE Full text] [CrossRef] [Medline]
  91. Pisner DA, Schnyer DM. Chapter 6 - Support vector machine. In: Mechelli A, Vieira S, editors. Machine Learning: Methods and Applications to Brain Disorders. Cambridge, MA. Academic Press; 2020.
  92. Son YJ, Kim HG, Kim EH, Choi S, Lee SK. Application of support vector machine for prediction of medication adherence in heart failure patients. Healthc Inform Res. Dec 2010;16(4):253-259. [FREE Full text] [CrossRef] [Medline]
  93. Manikandan G, Pragadeesh B, Manojkumar V, Karthikeyan AL, Manikandan R, Gandomi AH. Classification models combined with Boruta feature selection for heart disease prediction. Inf Med Unlock. 2024;44:101442. [CrossRef]
  94. Chen T, Guestrin C. XGBoost: a scalable tree boosting system. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 2016. Presented at: KDD '16; August 13-17, 2016; San Francisco, CA. [CrossRef]
  95. Nadkarni SB, Vijay GS, Kamath RC. Comparative study of random forest and gradient boosting algorithms to predict airfoil self-noise. Eng Proc. 2023;59(1):24. [CrossRef]
  96. Ogunleye A, Wang QG. XGBoost model for chronic kidney disease diagnosis. IEEE/ACM Trans Comput Biol Bioinf. Nov 1, 2020;17(6):2131-2140. [CrossRef]
  97. Chakraborty A, Tsokos CP. An AI-driven predictive model for pancreatic cancer patients using extreme gradient boosting. J Stat Theory Appl. Sep 11, 2023;22(4):262-282. [CrossRef]
  98. Zou J, Han Y, So SS. Overview of artificial neural networks. Methods Mol Biol. 2008;458:15-23. [CrossRef] [Medline]
  99. Kufel J, Bargieł-Łączek K, Kocot S, Koźlik M, Bartnikowska W, Janik M, et al. What is machine learning, artificial neural networks and deep learning?-examples of practical applications in medicine. Diagnostics (Basel). Aug 03, 2023;13(15):2582. [FREE Full text] [CrossRef] [Medline]
  100. Ahmed FE. Artificial neural networks for diagnosis and survival prediction in colon cancer. Mol Cancer. Aug 06, 2005;4(1):29. [FREE Full text] [CrossRef] [Medline]
  101. Abbas S, Asif M, Rehman A, Alharbi M, Khan MA, Elmitwally N. Emerging research trends in artificial intelligence for cancer diagnostic systems: a comprehensive review. Heliyon. Sep 15, 2024;10(17):e36743. [FREE Full text] [CrossRef] [Medline]
  102. Sirocchi C, Bogliolo A, Montagna S. Medical-informed machine learning: integrating prior knowledge into medical decision systems. BMC Med Inform Decis Mak. Jun 28, 2024;24(Suppl 4):186. [FREE Full text] [CrossRef] [Medline]
  103. Rostami M, Oussalah M. A novel explainable COVID-19 diagnosis method by integration of feature selection with random forest. Inform Med Unlocked. 2022;30:100941. [CrossRef] [Medline]
  104. Mohsin SN, Gapizov A, Ekhator C, Ain NU, Ahmad S, Khan M, et al. The role of artificial intelligence in prediction, risk stratification, and personalized treatment planning for congenital heart diseases. Cureus. Aug 2023;15(8):e44374. [FREE Full text] [CrossRef] [Medline]
  105. Natekin A, Knoll A. Gradient boosting machines, a tutorial. Front Neurorobot. 2013;7:21. [FREE Full text] [CrossRef] [Medline]
  106. Ji GW, Jiao CY, Xu ZG, Li XC, Wang K, Wang XH. Development and validation of a gradient boosting machine to predict prognosis after liver resection for intrahepatic cholangiocarcinoma. BMC Cancer. Mar 11, 2022;22(1):258. [FREE Full text] [CrossRef] [Medline]
  107. Hossain E, Rana R, Higgins N, Soar J, Datta Barua P, Pisani AR, et al. Natural language processing in electronic health records in relation to healthcare decision-making: a systematic review. Comput Biol Med. Mar 2023;155:106649. [CrossRef] [Medline]
  108. Rojas-Carabali W, Agrawal R, Gutierrez-Sinisterra L, Baxter SL, Cifuentes-González C, Wei YC, et al. Natural language processing in medicine and ophthalmology: a review for the 21st-century clinician. Asia Pac J Ophthalmol (Phila). Jul 2024;13(4):100084. [FREE Full text] [CrossRef] [Medline]
  109. Sheikhalishahi S, Miotto R, Dudley JT, Lavelli A, Rinaldi F, Osmani V. Natural language processing of clinical notes on chronic diseases: systematic review. JMIR Med Inform. Apr 27, 2019;7(2):e12239. [FREE Full text] [CrossRef] [Medline]
  110. George B, Seals S, Aban I. Survival analysis and regression models. J Nucl Cardiol. Aug 2014;21(4):686-694. [FREE Full text] [CrossRef] [Medline]
  111. Beis G, Iliopoulos A, Papasotiriou I. An overview of introductory and advanced survival analysis methods in clinical applications: where have we come so far? Anticancer Res. Feb 02, 2024;44(2):471-487. [CrossRef] [Medline]
  112. Wassan S, Dongyan H, Suhail B, Jhanjhi NZ, Xiao G, Ahmed S, et al. Deep convolutional neural network and IoT technology for healthcare. Digit Health. Jan 17, 2024;10:20552076231220123. [FREE Full text] [CrossRef] [Medline]
  113. Nazir S, Dickson DM, Akram MU. Survey of explainable artificial intelligence techniques for biomedical imaging with deep neural networks. Comput Biol Med. Apr 2023;156:106668. [FREE Full text] [CrossRef] [Medline]
  114. Lundervold AS, Lundervold A. An overview of deep learning in medical imaging focusing on MRI. Z Med Phys. May 2019;29(2):102-127. [FREE Full text] [CrossRef] [Medline]
  115. Genuer R, Poggi JM, Tuleau-Malot C, Villa-Vialaneix N. Random forests for big data. Big Data Res. Sep 2017;9:28-46. [CrossRef]
  116. Khalilia M, Chakraborty S, Popescu M. Predicting disease risks from highly imbalanced data using random forest. BMC Med Inform Decis Mak. Jul 29, 2011;11(1):51. [FREE Full text] [CrossRef] [Medline]
  117. Kumar Y, Koul A, Singla R, Ijaz MF. Artificial intelligence in disease diagnosis: a systematic literature review, synthesizing framework and future research agenda. J Ambient Intell Humaniz Comput. Jan 13, 2023;14(7):8459-8486. [FREE Full text] [CrossRef] [Medline]
  118. Rajkomar A, Dean J, Kohane I. Machine learning in medicine. N Engl J Med. Apr 04, 2019;380(14):1347-1358. [CrossRef]
  119. Obermeyer Z, Powers B, Vogeli C, Mullainathan S. Dissecting racial bias in an algorithm used to manage the health of populations. Science. Oct 25, 2019;366(6464):447-453. [FREE Full text] [CrossRef] [Medline]
  120. Miotto R, Wang F, Wang S, Jiang X, Dudley JT. Deep learning for healthcare: review, opportunities and challenges. Brief Bioinform. Nov 27, 2018;19(6):1236-1246. [FREE Full text] [CrossRef] [Medline]
  121. Beam AL, Kohane IS. Big data and machine learning in health care. JAMA. Apr 03, 2018;319(13):1317-1318. [CrossRef] [Medline]


ANN: artificial neural network
DNN: deep neural network
DT: decision tree
EHR: electronic health record
EMA: European Medicines Agency
FDA: Food and Drug Administration
GBM: gradient boosting machine
LR: logistic regression
ML: machine learning
NLP: natural language processing
PRISMA: Preferred Reporting Items for Systematic Reviews and Meta-Analyses
RF: random forest
RWD: real-world data
RWE: real-world evidence
SVM: support vector machine
XGBoost: Extreme Gradient Boosting


Edited by A Coristine; submitted 17.11.24; peer-reviewed by A Paradise Vit, A Mavragani; comments to author 13.02.25; revised version received 27.02.25; accepted 22.04.25; published 19.06.25.

Copyright

©Norah Hamad Alhumaidi, Doni Dermawan, Hanin Farhana Kamaruzaman, Nasser Alotaiq. Originally published in JMIR Medical Informatics (https://medinform.jmir.org), 19.06.2025.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on https://medinform.jmir.org/, as well as this copyright and license information must be included.