<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Med Inform</journal-id><journal-id journal-id-type="publisher-id">medinform</journal-id><journal-id journal-id-type="index">7</journal-id><journal-title>JMIR Medical Informatics</journal-title><abbrev-journal-title>JMIR Med Inform</abbrev-journal-title><issn pub-type="epub">2291-9694</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v14i1e77965</article-id><article-id pub-id-type="doi">10.2196/77965</article-id><article-categories><subj-group subj-group-type="heading"><subject>Original Paper</subject></subj-group></article-categories><title-group><article-title>Prospective Diagnostic Accuracy and Technical Feasibility of Artificial Intelligence-Assisted Rib Fracture Detection on Chest Radiographs: Observational Study</article-title></title-group><contrib-group><contrib contrib-type="author"><name name-style="western"><surname>Huang</surname><given-names>Shu-Tien</given-names></name><degrees>MSc, MD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff3">3</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Liu</surname><given-names>Liong-Rung</given-names></name><degrees>MSc, MD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff3">3</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Tsai</surname><given-names>Ming-Feng</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff3">3</xref><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Huang</surname><given-names>Ming-Yuan</given-names></name><degrees>MSc, MD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff3">3</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author" corresp="yes" equal-contrib="yes"><name name-style="western"><surname>Chiu</surname><given-names>Hung-Wen</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff5">5</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib></contrib-group><aff id="aff1"><institution>Department of Emergency Medicine, Mackay Memorial Hospital</institution><addr-line>Taipei</addr-line><country>Taiwan</country></aff><aff id="aff2"><institution>Graduate Institute of Biomedical Informatics, College of Medical Science and Technology, Taipei Medical University</institution><addr-line>9F, Education &#x0026; Research Building, Shuang-Ho Campus, No. 301, Yuantong Rd., Zhonghe Dist.</addr-line><addr-line>New Taipei City</addr-line><country>Taiwan</country></aff><aff id="aff3"><institution>College of Medicine, Mackay Medical University</institution><addr-line>New Taipei City</addr-line><country>Taiwan</country></aff><aff id="aff4"><institution>Division of Plastic Surgery, Mackay Memorial Hospital</institution><addr-line>Taipei</addr-line><country>Taiwan</country></aff><aff id="aff5"><institution>Clinical Big Data Research, Taipei Medical University Hospital</institution><addr-line>Taipei City</addr-line><country>Taiwan</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Coristine</surname><given-names>Andrew</given-names></name></contrib><contrib contrib-type="editor"><name name-style="western"><surname>Benis</surname><given-names>Arriel</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Yoo</surname><given-names>Dongjoon</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Bai</surname><given-names>Enze</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Zhang</surname><given-names>Jun</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Hung-Wen Chiu, PhD, Graduate Institute of Biomedical Informatics, College of Medical Science and Technology, Taipei Medical University, 9F, Education &#x0026; Research Building, Shuang-Ho Campus, No. 301, Yuantong Rd., Zhonghe Dist., New Taipei City, 235603, Taiwan, 886 266202589 ext 10929; <email>hwchiu@tmu.edu.tw</email></corresp><fn fn-type="equal" id="equal-contrib1"><label>*</label><p>these authors contributed equally</p></fn></author-notes><pub-date pub-type="collection"><year>2026</year></pub-date><pub-date pub-type="epub"><day>29</day><month>1</month><year>2026</year></pub-date><volume>14</volume><elocation-id>e77965</elocation-id><history><date date-type="received"><day>22</day><month>05</month><year>2025</year></date><date date-type="rev-recd"><day>30</day><month>12</month><year>2025</year></date><date date-type="accepted"><day>01</day><month>01</month><year>2026</year></date></history><copyright-statement>&#x00A9; Shu-Tien Huang, Liong-Rung Liu, Ming-Feng Tsai, Ming-Yuan Huang, Hung-Wen Chiu. Originally published in JMIR Medical Informatics (<ext-link ext-link-type="uri" xlink:href="https://medinform.jmir.org">https://medinform.jmir.org</ext-link>), 29.1.2026. </copyright-statement><copyright-year>2026</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://medinform.jmir.org/">https://medinform.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://medinform.jmir.org/2026/1/e77965"/><abstract><sec><title>Background</title><p>Rib fractures are present in 10%&#x2010;15% of thoracic trauma cases but are often missed on chest radiographs, delaying diagnosis and treatment. Artificial intelligence (AI) may improve detection and triage in emergency settings.</p></sec><sec><title>Objective</title><p>This study aims to evaluate diagnostic accuracy, processing speed, and technical feasibility of an artificial intelligence&#x2013;assisted rib fracture detection system using prospectively collected data within a real-world, high-volume emergency department workflow.</p></sec><sec sec-type="methods"><title>Methods</title><p>We conducted an observational feasibility study with prospective data collection of a faster region-based convolutional neural network&#x2013;based AI model deployed in the emergency department to analyze 23,251 real-world chest radiographs (22,946 anteroposterior; 305 oblique) from April 1 to July 2, 2023. This study was approved by the Institutional Review Board of MacKay Memorial Hospital (IRB No. 20MMHIS483e). AI operated passively, without influencing clinical decision-making. The reference standard was the final report issued by board-certified radiologists. A subset of discordant cases underwent post hoc computed tomography review for exploratory analysis.</p></sec><sec sec-type="results"><title>Results</title><p>AI achieved 74.5% sensitivity (95% CI 0.708-0.780), 93.3% specificity (95% CI 0.930-0.937), 24.2% positive predictive value, and 99.2% negative predictive value. Median inference time was 10.6 seconds versus 3.3 hours for radiologist reports (paired Wilcoxon signed-rank test <italic>W</italic>=112 987.5, <italic>P</italic>&#x003C;.001). The analysis revealed peak imaging demand between 08:00 and 16:00 and Thursday-Saturday evenings. A 14-day graphics processing unit outage underscored the importance of infrastructure resilience.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>The AI system demonstrated strong technical feasibility for real-time rib fracture detection in a high-volume emergency department setting, with rapid inference and stable performance during prospective deployment. Although the system showed high negative predictive value, the observed false-positive and false-negative rates indicate that it should be considered a supportive screening tool rather than a stand-alone diagnostic solution or a replacement for clinical judgment. These findings support further clinician-in-the-loop studies to evaluate clinical feasibility, workflow integration, and impact on diagnostic decision-making. However, interpretation is limited by reliance on radiology reports as the reference standard and the system&#x2019;s passive, non-interventional deployment.</p></sec></abstract><kwd-group><kwd>artificial intelligence</kwd><kwd>chest radiograph</kwd><kwd>clinical decision support</kwd><kwd>digital health</kwd><kwd>emergency radiology</kwd><kwd>feasibility study</kwd><kwd>faster R-CNN</kwd><kwd>real-world data</kwd><kwd>rib fracture</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>Digital health technologies, particularly artificial intelligence (AI), are increasingly used to address diagnostic delays in high-acuity clinical settings. In emergency departments, timely identification of injuries is essential, yet radiographic interpretation remains constrained by heavy workloads and the inherent complexity of imaging&#x2014;especially for subtle findings such as rib fractures.</p><p>Rib fractures are a frequent consequence of thoracic trauma, occurring in 10%&#x2010;15% of trauma patients and often indicating more serious underlying injuries [<xref ref-type="bibr" rid="ref1">1</xref>-<xref ref-type="bibr" rid="ref3">3</xref>]. When missed, they may lead to inadequate pain management, delayed respiratory support, pneumonia, or even preventable intensive care unit admissions. Beyond clinical harm, undetected fractures also carry medicolegal implications and increase health care costs.</p><p>Despite their significance, rib fractures are notoriously difficult to detect on chest radiographs (CXRs)&#x2014;the first-line imaging modality in most emergency departments&#x2014;due to overlapping anatomical structures and subtle fracture lines. Reported sensitivities for radiologist detection can be as low as 15%, with up to half of fractures potentially missed in high-volume settings [<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref5">5</xref>]. Although computed tomography (CT) and ultrasound can improve accuracy, they are resource-intensive and not always feasible for frontline triage [<xref ref-type="bibr" rid="ref6">6</xref>-<xref ref-type="bibr" rid="ref8">8</xref>]. These limitations highlight an urgent need for AI-driven tools that can assist clinicians by rapidly identifying suspected rib fractures in routine CXRs, enabling more effective prioritization and timely intervention.</p><p>Recent advances in AI, particularly deep learning, have demonstrated strong potential in automating image analysis tasks across medical domains, including dermatology, ophthalmology, and pulmonary imaging [<xref ref-type="bibr" rid="ref9">9</xref>-<xref ref-type="bibr" rid="ref11">11</xref>]. Deep learning models, especially convolutional neural networks, can automatically extract complex image features and have shown superior performance compared to traditional machine learning methods in various image classification tasks [<xref ref-type="bibr" rid="ref11">11</xref>-<xref ref-type="bibr" rid="ref13">13</xref>]. Transfer learning further enables the adaptation of pretrained convolutional neural networks&#x2014;originally developed for natural images&#x2014;for medical image classification tasks, including bone fracture detection [<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref15">15</xref>].</p><p>Although prior studies have applied deep learning to rib fracture detection with promising results, most were retrospective, limited in scale, and did not assess feasibility in operational emergency department workflows [<xref ref-type="bibr" rid="ref7">7</xref>,<xref ref-type="bibr" rid="ref16">16</xref>,<xref ref-type="bibr" rid="ref17">17</xref>]. These proof-of-concept efforts did not address the practical barriers to integrating AI into emergency radiology workflows, such as inference latency, system interoperability, or artifact handling.</p><p>To address this gap, we conducted an observational feasibility study with prospective data collection, evaluating an AI model for rib-fracture detection on CXRs. The system was passively deployed in parallel with routine emergency department imaging workflows using real-world data, without influencing clinical decisions. This design allowed the assessment of diagnostic performance, processing speed, and operational characteristics within standard clinical workflows.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Study Design</title><p>The observational feasibility study protocol was reviewed and approved by the Institutional Review Board of MacKay Memorial Hospital (IRB No. 20MMHIS483e) prior to the initiation of data collection. MacKay Memorial Hospital is a tertiary referral and level 1 trauma center in northern Taiwan. The AI system functioned passively in real time without influencing clinical decisions or patient management. As the system functioned in a noninterventional, observational manner, prospective trial registration was not required.</p><p>From April 1 to July 2, 2023, all chest and rib radiographs acquired in the emergency department were automatically processed by the AI system in near-real-time. Both standard CXRs and rib-only views acquired during the study period were automatically analyzed, as both modalities are routinely used for suspected thoracic trauma. During the study period, a temporary 14-day graphics processing unit (GPU) hardware outage occurred, during which radiographs were not processed in real time; these examinations were excluded from turnaround time analysis but retained for diagnostic accuracy, as formal radiology reports were available. No additional exclusion criteria were applied beyond the 14-day system outage; all eligible emergency radiographs during the study period were included in the analysis. The system operated passively alongside routine clinical workflows, without influencing clinical decisions. AI-identified suspected rib fractures were highlighted using bounding boxes on a backend interface, which was accessible only for research evaluation and remained hidden from the clinical care team (<xref ref-type="fig" rid="figure1">Figure 1</xref>).</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Study workflow of a prospective observational feasibility study evaluating an artificial intelligence&#x2013;assisted rib fracture detection system using chest radiographs in patients admitted to the emergency department at a high-volume tertiary medical center (April 1-July 2, 2023). GPU: graphics processing unit.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e77965_fig01.png"/></fig></sec><sec id="s2-2"><title>Ethical Considerations</title><p>This study was approved by the Institutional Review Board of MacKay Memorial Hospital (IRB No. 20MMHIS483e) and conducted in accordance with the Declaration of Helsinki. Informed consent was waived because the study involved secondary analysis of routinely acquired, deidentified clinical imaging data, and the AI system operated passively without influencing patient management.</p><p>All data were deidentified prior to analysis and processed on secure institutional servers with access limited to authorized research personnel. No compensation was provided to participants. All images included in the manuscript were fully anonymized, and no identifiable patient information is disclosed.</p><p>Consistent with the approved observational study design, all AI outputs&#x2014;including discordant cases&#x2014;were withheld from treating clinicians and did not influence patient management.</p></sec><sec id="s2-3"><title>AI Model Development</title><p>CXRs were retrospectively collected from the hospital picture archiving and communication system (PACS) for model training. All images were deidentified and preprocessed using histogram equalization and image inversion to improve fracture conspicuity. Fracture locations were annotated using bounding boxes by a board-certified emergency physician with 18 years of clinical experience via the DeepQ AI platform [<xref ref-type="bibr" rid="ref18">18</xref>].</p><p>A deep learning model was developed using PyTorch (v1.13) with GPU acceleration. The architecture was based on faster region-based convolutional neural network (R-CNN) [<xref ref-type="bibr" rid="ref19">19</xref>], incorporating a ResNet-50 backbone for feature extraction, a region proposal network for candidate region generation, and a classification head for fracture detection.</p><p>The dataset comprised 2079 CXRs (1065 fracture-positive and 1014 normal) collected between 2010 and 2020. Images were randomly divided into training (80%) and validation (20%) sets at the image level, as each radiograph represented an independent study. When multiple images were obtained from the same patient encounter, each radiograph was treated as an independent sample. Data augmentation&#x2014;including random rotation, flipping, brightness, and contrast adjustment&#x2014;was applied to improve generalization. To address the inherent class imbalance given the low fracture prevalence, class-weighted loss and oversampling of fracture-positive images were employed.</p></sec><sec id="s2-4"><title>Model Validation</title><p>Model performance was assessed on a hold-out test set of 262 CXRs containing 724 expert-annotated rib fractures. Evaluation metrics were reported at both the case and object levels.</p><p>At the case level, the unit of analysis was the radiographic study. A study was considered positive if at least 1 rib fracture was detected, regardless of the number of fractures present. The model correctly identified fractures in 230 of 257 fracture-positive studies, achieving a sensitivity of 89.5%. With only 8 false-positive cases, precision reached 96.6%, yielding an overall <italic>F</italic><sub>1</sub>-score of 0.93.</p><p>At the object level, performance reflected per-lesion detection accuracy. The model correctly localized 680 of 724 annotated rib fractures (recall=94.0%) and generated 55 false-positive boxes. The mean average precision at an intersection-over-union threshold of &#x2265;0.5 was 0.65, indicating robust lesion-level localization (<xref ref-type="table" rid="table1">Table 1</xref>).</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Dataset composition and performance of the artificial intelligence (AI) model using retrospective emergency department chest radiographs, including case-level detection and object-level localization (intersection-over-union [IoU]&#x2265;0.50).</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Category and metric</td><td align="left" valign="bottom">Value</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="2">Dataset</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Total images</td><td align="left" valign="top">262</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Ground-truth boxes</td><td align="left" valign="top">724</td></tr><tr><td align="left" valign="top" colspan="2">Case-level detection (%)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Sensitivity (recall)</td><td align="left" valign="top">89.6</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Precision</td><td align="left" valign="top">96.6</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><italic>F</italic><sub>1</sub>-score</td><td align="left" valign="top">0.93</td></tr><tr><td align="left" valign="top" colspan="2">Object-level localization</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Recall</td><td align="left" valign="top">94.0%</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>mAP<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup> (IoU&#x2265;0.5)</td><td align="left" valign="top">0.65</td></tr></tbody></table><table-wrap-foot><fn id="table1fn1"><p><sup>a</sup>mAP: mean average precision.</p></fn></table-wrap-foot></table-wrap><p>Curve-based analyses further characterized the model&#x2019;s detection behavior (<xref ref-type="fig" rid="figure2">Figure 2</xref>). The precision-recall curve (<xref ref-type="fig" rid="figure2">Figure 2A</xref>) maintained precision &#x2265;0.90 until recall fell below 0.55 (precision-recall area under the curve=0.65), demonstrating high reliability across a broad sensitivity range. The free-response receiver operating characteristic (FROC) curve (<xref ref-type="fig" rid="figure2">Figure 2B</xref>) showed true-positive rates of 0.77 at 1 false positive per image and 0.88 at three, representing practical trade-offs between sensitivity and alert frequency in potential clinical deployment.</p><fig position="float" id="figure2"><label>Figure 2.</label><caption><p>The performance of the artificial intelligence (AI)&#x2013;assisted rib fracture detection model was evaluated in the retrospective model development and validation dataset using emergency department chest radiographs. (A) Precision-recall curve demonstrating case-level detection performance at an intersection-over-union (IoU) threshold of &#x2265;0.50 (area under the curve=0.65). (B) Free-response receiver operating characteristic (FROC) curve showing lesion-level sensitivity as a function of false positives per image, with localization performance assessed at the same IoU threshold (mean average precision=0.65; recall=94%).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e77965_fig02.png"/></fig></sec><sec id="s2-5"><title>Prospective Evaluation of AI Model in Emergency Department Workflow</title><p>The trained model was prospectively evaluated in parallel with clinical workflow, performing automated inference on incoming emergency radiographs. During the automated inference process, all incoming radiographs were standardized and resized to a fixed resolution of 512&#x00D7;512 pixels. The prospective evaluation cohort (April-July 2023) was temporally and operationally independent from the retrospective training and validation dataset (2010-2020), and no patient overlap existed between the 2 cohorts. During the prospective evaluation phase, all chest and rib radiographs from emergency department encounters were automatically processed by the AI system without disrupting clinical workflows. The bounding box outputs were logged for research analysis but were not disclosed to radiologists or used for patient management.</p></sec><sec id="s2-6"><title>Performance Assessment Using NLP-Derived Labels</title><p>To evaluate AI performance in the real-world setting, output was compared to formal radiology reports issued by board-certified radiologists. A rule-based natural language processing (NLP) pipeline was developed to extract structured rib fracture labels (positive, negative, or ambiguous) from free-text reports. The algorithm combined keyword detection (eg, &#x201C;rib fracture,&#x201D; &#x201C;fx&#x201D;) and negation handling (eg, &#x201C;no evidence of,&#x201D; &#x201C;no definite fracture&#x201D;).</p><p>To validate NLP accuracy, a random sample of 200 radiology reports was manually reviewed by 2 emergency physicians blinded to both NLP and AI results. The NLP classification achieved 96.5% agreement (193/200) with manual review, with a Cohen &#x03BA; of 0.91, indicating excellent concordance. Most discrepancies were due to ambiguous language or complex negation structures (<xref ref-type="table" rid="table2">Table 2</xref>).</p><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>Confusion matrix comparing natural language processing (NLP)&#x2013;extracted radiology report labels with manual expert review in a randomly selected subset of 200 emergency department chest radiographs, used to assess labeling accuracy in the retrospective dataset.</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Manual review</td><td align="left" valign="bottom">Fracture present</td><td align="left" valign="bottom">Fracture absent</td><td align="left" valign="bottom">Ambiguous</td><td align="left" valign="bottom">Total (NLP prediction)</td></tr></thead><tbody><tr><td align="left" valign="top">NLP: fracture present</td><td align="left" valign="top">95</td><td align="left" valign="top">2</td><td align="left" valign="top">1</td><td align="left" valign="top">98</td></tr><tr><td align="left" valign="top">NLP: fracture absent</td><td align="left" valign="top">3</td><td align="left" valign="top">90</td><td align="left" valign="top">2</td><td align="left" valign="top">95</td></tr><tr><td align="left" valign="top">NLP: ambiguous</td><td align="left" valign="top">1</td><td align="left" valign="top">2</td><td align="left" valign="top">4</td><td align="left" valign="top">7</td></tr><tr><td align="left" valign="top">Total (manual review)</td><td align="left" valign="top">99</td><td align="left" valign="top">94</td><td align="left" valign="top">7</td><td align="left" valign="top">200</td></tr></tbody></table></table-wrap><p>Formal radiology reports served as the primary reference standard for AI evaluation. This approach may have underestimated AI sensitivity because CT confirmation was not performed systematically. In select discordant cases&#x2014;where the AI flagged fractures not documented in the reports&#x2014;subsequent CT scans confirmed some of these findings. These discrepancies were retrospectively reviewed to investigate potential underreporting by radiologists. While informative, these exploratory adjudications were not used as a universal reference standard due to inconsistent CT availability. Nonetheless, radiology reports remained the definitive benchmark for all performance metrics. Additionally, the selected misclassified cases were examined to identify recurring patterns of diagnostic oversight among frontline physicians.</p></sec><sec id="s2-7"><title>Targeted Adjudication of Discordant Cases</title><p>To further explore potential underreporting within the report-based reference standard, we performed a focused review of discordant cases in which the AI system flagged suspected rib fractures not documented in the corresponding formal radiology reports. Because not all discordant cases had confirmatory CT imaging, an illustrative subset of 11 cases was selected based on the availability of same-encounter chest CT and clinical relevance for qualitative adjudication. Each case was reviewed to determine whether the AI-predicted fractures corresponded to true fractures confirmed on CT. These adjudications were exploratory and intended to contextualize the potential clinical value of AI detection beyond the report-based benchmarking.</p></sec><sec id="s2-8"><title>Data Analysis</title><p>Case-level performance was assessed at the radiographic study (accession) level by comparing AI outputs to NLP-derived labels from radiology reports. A study was considered positive if at least 1 image within the same examination was flagged as having a rib fracture by the AI system; otherwise, the study was classified as negative. Key metrics included sensitivity, specificity, accuracy, positive predictive value, negative predictive value (NPV), and <italic>F</italic><sub>1</sub>-score. Ninety-five percent CIs were calculated using nonparametric bootstrap resampling (1000 iterations). The results were summarized in confusion matrices and diagnostic performance plots.</p></sec><sec id="s2-9"><title>Statistical Analysis</title><p>All statistical analyses were performed using Python 3.11 (pandas v2.2, scikit-learn v1.4) and R 4.3.2. Continuous variables were reported as mean (SD) or median with IQR. Categorical variables were summarized as counts and percentages. A 2-tailed <italic>P</italic>&#x003C;.001 was considered statistically significant.</p><p>This study was reported in accordance with the Checklist for Artificial Intelligence in Medical Imaging (CLAIM) guideline, with the completed checklist provided as <xref ref-type="supplementary-material" rid="app1">Checklist 1</xref>.</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Study Cohort</title><p>From April 1 to July 2, 2023, all chest and dedicated rib radiographs acquired in the emergency department were automatically processed by the AI system in a parallel workflow, yielding 23,251 imaging studies from 20,908 unique patient visits. Population demographics are summarized in <xref ref-type="table" rid="table3">Table 3</xref>. The mean age was 55.9 years (SD 22.3; range 0&#x2010;106), with 10,770 (51.5%) male and 10,138 (48.5%) female patients. A radiologist review identified 589 rib-fracture cases (prevalence 2.8%).</p><table-wrap id="t3" position="float"><label>Table 3.</label><caption><p>Demographic and clinical characteristics of patients in the emergency department included in a prospective observational study of artificial intelligence (AI)&#x2013;assisted rib fracture detection (April 1-July 2, 2023).</p></caption><table id="table3" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Characteristic</td><td align="left" valign="bottom">Value</td></tr></thead><tbody><tr><td align="left" valign="top">Total cases</td><td align="left" valign="top">20,908</td></tr><tr><td align="left" valign="top">Age (y), mean (SD; range)</td><td align="left" valign="top">55.9 (22.3; 0&#x2010;106)</td></tr><tr><td align="left" valign="top">Sex, n (%)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Male</td><td align="left" valign="top">10,770 (51.5)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Female</td><td align="left" valign="top">10,138 (48.5)</td></tr><tr><td align="left" valign="top">Radiologist-confirmed rib fractures, n (%)</td><td align="left" valign="top">589 (2.8)</td></tr></tbody></table></table-wrap></sec><sec id="s3-2"><title>AI Model Performance</title><p>AI model outputs were compared on a per-case basis against structured rib-fracture labels derived from board-certified radiology reports. At the selected operating point&#x2014;corresponding to approximately one false-positive per image on the FROC curve&#x2014;the system achieved a sensitivity of 0.745 (95% CI 0.708&#x2010;0.780) and specificity of 0.933 (95% CI 0.930&#x2010;0.937). Positive predictive value was 0.242 (95% CI 0.223&#x2010;0.262), and negative predictive value was 0.992 (95% CI 0.991&#x2010;0.994). The overall <italic>F</italic><sub>1</sub>-score was 0.365 (95% CI 0.340&#x2010;0.390) with an accuracy of 0.928 (<xref ref-type="table" rid="table4">Table 4</xref>).</p><table-wrap id="t4" position="float"><label>Table 4.</label><caption><p>Case-based diagnostic performance of an artificial intelligence&#x2013;assisted rib fracture detection system in a prospective observational emergency department study<sup><xref ref-type="table-fn" rid="table4fn1">a</xref></sup>.</p></caption><table id="table4" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Metric</td><td align="left" valign="bottom">Estimate (95% CI)</td></tr></thead><tbody><tr><td align="left" valign="top">Sensitivity</td><td align="left" valign="top">0.745 (0.708&#x2010;0.780)</td></tr><tr><td align="left" valign="top">Specificity</td><td align="left" valign="top">0.933 (0.930&#x2010;0.937)</td></tr><tr><td align="left" valign="top">PPV<sup><xref ref-type="table-fn" rid="table4fn2">b</xref></sup></td><td align="left" valign="top">0.242 (0.223&#x2010;0.262)</td></tr><tr><td align="left" valign="top">NPV<sup><xref ref-type="table-fn" rid="table4fn3">c</xref></sup></td><td align="left" valign="top">0.992 (0.991&#x2010;0.994)</td></tr><tr><td align="left" valign="top"><italic>F</italic><sub>1</sub>-score</td><td align="left" valign="top">0.365 (0.340&#x2010;0.390)</td></tr><tr><td align="left" valign="top">Accuracy</td><td align="left" valign="top">0.928 (N/A<sup><xref ref-type="table-fn" rid="table4fn4">d</xref></sup>)</td></tr></tbody></table><table-wrap-foot><fn id="table4fn1"><p><sup>a</sup>Performance metrics are reported with 95% CIs using final radiologist reports as the reference standard.</p></fn><fn id="table4fn2"><p><sup>b</sup>PPV: positive predictive value.</p></fn><fn id="table4fn3"><p><sup>c</sup>NPV: negative predictive value.</p></fn><fn id="table4fn4"><p><sup>d</sup>N/A: not available.</p></fn></table-wrap-foot></table-wrap><p>As shown in <xref ref-type="fig" rid="figure3">Figure 3</xref>, the AI system correctly identified 431 (74.5%) fracture-positive cases while producing 1357 (6.1%) false positives and 148 (0.7%) false negatives across 23,251 studies. This distribution demonstrates the model&#x2019;s high true-negative count (n=18,972, 93.3%) and its strong negative predictive value during deployment in the emergency department. No temporal drift or learning curve effects were observed, as the deployed model remained fixed throughout the study period.</p><fig position="float" id="figure3"><label>Figure 3.</label><caption><p>Case-level confusion matrix of the artificial intelligence (AI)&#x2013;assisted rib fracture detection system during prospective emergency department deployment, using final radiologist reports as the reference standard. Darker blue indicates a higher number of cases (count), as shown in the color bar.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e77965_fig03.png"/></fig></sec><sec id="s3-3"><title>Imaging Workload Patterns</title><p>Analysis of imaging demand revealed predictable diurnal and weekly patterns, with peak volumes between 08:00 and 16:00 daily and secondary surges on Thursday to Saturday evenings. Demand was the lowest between 00:00 and 07:00 across all days of the week (<xref ref-type="fig" rid="figure4">Figure 4</xref>).</p><fig position="float" id="figure4"><label>Figure 4.</label><caption><p>Heatmap illustrating the temporal distribution of chest radiograph imaging workload by hour of day and weekday during prospective emergency department deployment. Color intensity represents the number of chest radiographs acquired per hour, highlighting peak imaging periods across weekdays and weekends.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e77965_fig04.png"/></fig></sec><sec id="s3-4"><title>Inference Turnaround Time</title><p>A total of 19,641 paired cases were included to compare AI inference and radiologist report turnaround times. As shown in <xref ref-type="table" rid="table5">Table 5</xref>, the AI system achieved a median processing time of 10.6 seconds per image (IQR 9.0&#x2010;14.0; range 3&#x2010;35 s), compared with a median of 3.3 hours (IQR 1.31&#x2010;4.80; range 0.08&#x2010;72 h) for radiologist reports. This represents a more than 1000-fold reduction in turnaround time. <xref ref-type="fig" rid="figure5">Figure 5</xref> illustrates this disparity using boxplots on a logarithmic scale. A paired Wilcoxon signed-rank test confirmed that AI inference was significantly faster than radiologist reporting (<italic>W</italic>=112,987.5; <italic>P</italic>&#x003C;.001). This median reporting time reflects the full clinical workflow, including overnight and backlog delays typical of high-volume emergency radiology settings.</p><table-wrap id="t5" position="float"><label>Table 5.</label><caption><p>Turnaround times for artificial intelligence (AI) inference versus radiologist reporting during prospective emergency department deployment (n=19,641)<sup><xref ref-type="table-fn" rid="table5fn1">a</xref></sup>.</p></caption><table id="table5" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Metric</td><td align="left" valign="bottom">AI inference time (s)</td><td align="left" valign="bottom">Radiologist report time (s)</td></tr></thead><tbody><tr><td align="left" valign="top">Mean</td><td align="left" valign="top">10.9</td><td align="left" valign="top">10,877</td></tr><tr><td align="left" valign="top">Standard deviation</td><td align="left" valign="top">3.0</td><td align="left" valign="top">4008</td></tr><tr><td align="left" valign="top">Median (50%)</td><td align="left" valign="top">10.6</td><td align="left" valign="top">11,880</td></tr><tr><td align="left" valign="top">IQR (25%&#x2010;75%)</td><td align="left" valign="top">9.0&#x2010;14.0</td><td align="left" valign="top">4728&#x2010;17,280</td></tr><tr><td align="left" valign="top">Min</td><td align="left" valign="top">3.0</td><td align="left" valign="top">300</td></tr><tr><td align="left" valign="top">Max</td><td align="left" valign="top">35.0</td><td align="left" valign="top">259,200</td></tr></tbody></table><table-wrap-foot><fn id="table5fn1"><p><sup>a</sup>Times are summarized using descriptive statistics, including median and interquartile range.</p></fn></table-wrap-foot></table-wrap><fig position="float" id="figure5"><label>Figure 5.</label><caption><p>Comparison of turnaround times between artificial intelligence (AI) inference and radiologist reporting during prospective emergency department deployment (April 1-July 2, 2023). Boxplots on a logarithmic scale illustrate differences in processing time distributions between real-time AI inference and routine clinical reporting.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e77965_fig05.png"/></fig><p>This processing speed highlights the potential of AI-assisted triage systems to complement radiology workflows by rapidly identifying cases for prioritized review, especially in high-volume emergency settings.</p></sec><sec id="s3-5"><title>System Reliability</title><p>During the prospective evaluation of the AI system operating in a parallel clinical workflow, the AI platform experienced 1 service interruption&#x2014;from 27 April to 10 May 2023&#x2014;caused by GPU overload that halted all inference operations. A total of 3610 studies acquired during this 14-day outage were excluded from the turnaround time analysis but retained in the diagnostic accuracy evaluation (their formal radiology reports remained available). Following hardware replacement, full functionality was restored on 11 May, and no further outages occurred over the remainder of the study period.</p></sec><sec id="s3-6"><title>Illustrative Review of Discordant Cases</title><p>To further evaluate the AI system&#x2019;s potential diagnostic value beyond report-based benchmarking, a targeted adjudication was conducted on 11 representative discordant cases in which the AI system flagged suspected rib fractures not described in the corresponding radiology reports (<xref ref-type="table" rid="table6">Table 6</xref>). Among these, 7 cases (cases 1&#x2010;7) were subsequently confirmed as true fractures on CT (&#x201C;AI-CT concordant&#x201D;), indicating that several AI-labeled false positives in the report-based analysis represented true fractures missed in the reference standard. The remaining 4 cases (cases 8&#x2010;11) were confirmed negative on CT, primarily attributable to nonfracture anatomical structures or imaging artifacts. Including these CT-confirmed fractures as true positives would modestly increase the model&#x2019;s effective sensitivity and positive predictive value, highlighting the underestimation inherent in report-based benchmarking.</p><table-wrap id="t6" position="float"><label>Table 6.</label><caption><p>Targeted post hoc computed tomography (CT) adjudication of representative discordant cases in which artificial intelligence (AI) flagged suspected rib fractures not described in the corresponding radiology reports during prospective emergency department deployment.</p></caption><table id="table6" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Case</td><td align="left" valign="bottom">AI output</td><td align="left" valign="bottom">Radiologist report</td><td align="left" valign="bottom">Emergency physician</td><td align="left" valign="bottom">Outcome</td><td align="left" valign="bottom">Note (AI significance)</td></tr></thead><tbody><tr><td align="left" valign="top">1</td><td align="left" valign="top">Flagged right rib fracture</td><td align="left" valign="top">No fracture</td><td align="left" valign="top">Noted with POCUS<sup><xref ref-type="table-fn" rid="table6fn1">a</xref></sup></td><td align="left" valign="top">CT-confirmed</td><td align="left" valign="top">Triage value, prompting clinicians to perform US<sup><xref ref-type="table-fn" rid="table6fn2">b</xref></sup></td></tr><tr><td align="left" valign="top">2</td><td align="left" valign="top">Flagged right rib fracture</td><td align="left" valign="top">No fracture</td><td align="left" valign="top">Missed</td><td align="left" valign="top">CT-confirmed</td><td align="left" valign="top">AI-CT concordance</td></tr><tr><td align="left" valign="top">3</td><td align="left" valign="top">Flagged left fifth rib fracture</td><td align="left" valign="top">No fracture</td><td align="left" valign="top">Missed</td><td align="left" valign="top">CT-confirmed</td><td align="left" valign="top">AI-CT concordance</td></tr><tr><td align="left" valign="top">4</td><td align="left" valign="top">Flagged rib fracture post chest tube</td><td align="left" valign="top">No fracture</td><td align="left" valign="top">Missed</td><td align="left" valign="top">CT-confirmed</td><td align="left" valign="top">Chest tube artifact did not impair detection</td></tr><tr><td align="left" valign="top">5</td><td align="left" valign="top">Flagged lower-rib fracture</td><td align="left" valign="top">No fracture</td><td align="left" valign="top">Missed</td><td align="left" valign="top">CT-confirmed</td><td align="left" valign="top">AI-CT concordance</td></tr><tr><td align="left" valign="top">6</td><td align="left" valign="top">Flagged fracture near hardware</td><td align="left" valign="top">No fracture</td><td align="left" valign="top">Missed</td><td align="left" valign="top">CT-confirmed</td><td align="left" valign="top">Detected fracture adjacent to surgical hardware</td></tr><tr><td align="left" valign="top">7</td><td align="left" valign="top">Flagged fracture under scapular shadow</td><td align="left" valign="top">No fracture</td><td align="left" valign="top">Missed</td><td align="left" valign="top">CT-confirmed</td><td align="left" valign="top">AI-CT concordance</td></tr><tr><td align="left" valign="top">8</td><td align="left" valign="top">Flagged fracture at scapula border</td><td align="left" valign="top">No fracture</td><td align="left" valign="top">No fracture</td><td align="left" valign="top">False positive</td><td align="left" valign="top">Scapular margin misidentified</td></tr><tr><td align="left" valign="top">9</td><td align="left" valign="top">Flagged rib fracture at bra clasp</td><td align="left" valign="top">No fracture</td><td align="left" valign="top">No fracture</td><td align="left" valign="top">False positive</td><td align="left" valign="top">Bra hardware artifact</td></tr><tr><td align="left" valign="top">10</td><td align="left" valign="top">Flagged fracture at chest tube marker</td><td align="left" valign="top">No fracture</td><td align="left" valign="top">No fracture</td><td align="left" valign="top">False positive</td><td align="left" valign="top">Chest tube marker misinterpreted</td></tr><tr><td align="left" valign="top">11</td><td align="left" valign="top">Flagged fracture (image noise)</td><td align="left" valign="top">No fracture</td><td align="left" valign="top">No fracture</td><td align="left" valign="top">False positive</td><td align="left" valign="top">Image noise</td></tr></tbody></table><table-wrap-foot><fn id="table6fn1"><p><sup>a</sup>POCUS: point-of-care ultrasound. </p></fn><fn id="table6fn2"><p><sup>b</sup>US: ultrasound.</p></fn></table-wrap-foot></table-wrap><p>In one representative case (Case 3), the AI system correctly identified a subtle nondisplaced fracture of the left fifth rib that was not documented in the radiology report but later verified on 3D CT reconstruction (<xref ref-type="fig" rid="figure6">Figure 6</xref>). In contrast, <xref ref-type="fig" rid="figure7">Figure 7</xref> illustrates the main sources of false positives, including scapular margin misinterpretation, chest-tube hardware artifacts, and motion-induced noise.</p><fig position="float" id="figure6"><label>Figure 6.</label><caption><p>Representative true-positive rib fracture detected by artificial intelligence (AI) and confirmed by computed tomography (CT) during prospective emergency department deployment. (A) Chest radiograph showing an AI-flagged fracture of the left fifth rib that was not described in the initial radiology report. (B) Corresponding CT image confirming the fracture at the same anatomical location.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e77965_fig06.png"/></fig><fig position="float" id="figure7"><label>Figure 7.</label><caption><p>Representative false-positive detections generated by an artificial intelligence (AI)&#x2013;assisted rib fracture detection system during prospective emergency department deployment. The examples illustrate common sources of false-positive signals on chest radiographs. (A) Scapular margin overlap misinterpreted as a rib fracture. (B) Chest tube marker misidentified as a rib discontinuity. (C) Bra hardware producing a linear opacity mimicking a fracture. (D) Image noise and low-contrast regions leading to spurious detection.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e77965_fig07.png"/></fig><p>This targeted CT adjudication underscores the potential of AI-assisted screening to augment clinical vigilance by identifying subtle or overlooked fractures, while also emphasizing the need to improve artifact robustness and optimize false-positive suppression for practical clinical integration.</p></sec><sec id="s3-7"><title>Common Pitfalls in Frontline Rib Fracture Detection</title><p>Our review of discordant cases identified 3 principal drivers of missed rib fractures by emergency physicians: first, non-thoracic presenting complaints (eg, catheter malfunction or abdominal pain) led interpreters to focus on unrelated findings and overlook subtle rib breaks; second, the absence of classic chest pain&#x2014;patients describing only mild discomfort or a vague &#x201C;pop&#x201D;&#x2014;lowered clinical vigilance for nondisplaced fractures; and third, competing urgent injuries (facial, limb, or soft-tissue trauma) diverted attention from the chest, resulting in underappreciated fractures.</p></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Principal Findings</title><p>In this observational feasibility study using real-world emergency department imaging data, we demonstrated that a faster R-CNN&#x2013;based AI system can operate in parallel with routine clinical workflows to provide near-instantaneous rib fracture triage without influencing patient care. During a 3-month evaluation period, the model automatically processed 23,251 CXRs with a median inference time of 10.6 seconds per image, achieving a turnaround time reduction exceeding 3 orders of magnitude compared with formal radiologist reporting, while maintaining 74.5% sensitivity and 93.3% specificity. These findings position AI as a potential automated screening aid capable of rapidly identifying low-risk examinations and generating signals that could inform future prioritization strategies. The observed discrepancy between the 10.6-second AI inference time and the 3.3-hour radiologist turnaround time reflects a critical clinical bottleneck in busy emergency departments. Although these metrics represent different stages&#x2014;technical processing versus final clinical documentation&#x2014;the delay in official reporting highlights the diagnostic gap AI aims to address. In this context, near-instantaneous AI alerts may support case prioritization before formal reporting, although no clinician-facing alerts were implemented in this study.</p><p>However, the current findings primarily demonstrate technical feasibility rather than full clinical feasibility, as the AI system operated passively without direct clinician interaction. Future clinician-in-the-loop evaluations will be necessary to assess workflow integration, usability, and impact on diagnostic behavior or patient outcomes.</p><p>Although the system&#x2019;s positive predictive value was relatively low (24.2%), this trade-off aligns with its design as a triage support tool rather than a stand-alone diagnostic system. In high-volume emergency departments, the ability to rapidly identify examinations with a low likelihood of fracture is crucial. The model&#x2019;s high NPV of 99.2% allows clinicians to focus on a smaller, higher-risk subset of cases, thereby improving efficiency and reducing cognitive load. Given the observed sensitivity of 74.5%, false-negative cases remain possible, and the system should not be used as a stand-alone rule-out mechanism or as a substitute for clinical judgment.</p></sec><sec id="s4-2"><title>Comparison to Prior Work</title><p>Similar findings have been reported by Yao et al [<xref ref-type="bibr" rid="ref20">20</xref>], who demonstrated that deep learning systems with high NPV on chest CT can reduce radiologists&#x2019; workload by effectively identifying nonfracture cases. A recent systematic review by van den Broek et al [<xref ref-type="bibr" rid="ref21">21</xref>] further emphasized the triage value of AI in fracture detection, underscoring its potential across multiple imaging modalities.</p><p>Our AI system also demonstrated robust performance across circadian and weekly imaging surges, with peak volumes observed between 08:00 and 16:00 and during Thursday to Saturday evenings. However, a 14-day GPU hardware outage during the study period highlighted a real-world challenge of maintaining the AI system&#x2019;s reliability in clinical environments. This incident underscores the need for infrastructure redundancy, real-time monitoring, and failover protocols&#x2014;key considerations for sustainable AI deployment. These practical aspects of AI deployment remain underreported in most published studies [<xref ref-type="bibr" rid="ref22">22</xref>].</p><p>Compared with prior approaches such as the PACS-AI platform [<xref ref-type="bibr" rid="ref23">23</xref>], our system offered full automation, operating continuously in real time and without the need for manual image selection. This better reflects the demands of frontline emergency radiology. Herpe et al [<xref ref-type="bibr" rid="ref24">24</xref>] demonstrated improved diagnostic accuracy with PACS-integrated AI for limb fractures; however, their study did not evaluate scalability or autonomous triage capability under high-throughput conditions. In contrast, our study incorporated prospective data collection within a real-world emergency department workflow, allowing the assessment of AI performance, reliability, and operational feasibility under authentic clinical conditions.</p><p>Focused adjudication of discordant cases revealed that the AI system correctly identified rib fractures that were missed by both radiologists and emergency physicians in 7 cases, all subsequently confirmed on follow-up CT (&#x201C;AI-CT concordant&#x201D;). These findings highlight the potential of AI to strengthen diagnostic vigilance in complex clinical scenarios. While prior studies&#x2014;such as Zhou et al [<xref ref-type="bibr" rid="ref25">25</xref>]&#x2014;have shown that AI can detect rib fractures overlooked in initial CT interpretations, with confirmation on follow-up imaging, these investigations have largely centered on CT-based workflows. Although CT is highly sensitive, its routine use is limited by concerns over radiation, cost, and logistics. In contrast, our CXR-focused approach targets the most widely used imaging modality in acute care, offering a more scalable and practical solution for real-world emergency triage. Notably, Brady et al [<xref ref-type="bibr" rid="ref26">26</xref>] have emphasized that diagnostic errors and discrepancies are not uncommon in radiology, with daily error rates estimated at 3%&#x2010;5%, reinforcing the importance of AI as a complementary tool to enhance diagnostic accuracy.</p><p>Four false-positive cases revealed predictable pitfalls, including the misinterpretation of scapular margins, chest-tube hardware, and motion artifacts. Similar findings were reported by Sun et al [<xref ref-type="bibr" rid="ref27">27</xref>], who noted frequent false positives in an AI model for rib fracture detection on CXR, often due to anatomical overlap and imaging artifacts. These results support the need for artifact-aware retraining and preprocessing optimization to reduce false alerts and improve clinical integration.</p><p>This targeted CT adjudication further highlights the complementary role of AI in identifying subtle or overlooked fractures and underscores the inherent limitation of using report-based labels as the reference standard in real-world studies.</p></sec><sec id="s4-3"><title>Future Directions</title><p>In addition, integrating AI-generated alerts into emergency radiology workflows will require careful calibration of alert thresholds to minimize false positives and prevent alert fatigue among clinicians. Human-centered design, interface refinement, and iterative feedback from end users will be critical to achieving effective and sustainable adoption.</p><p>While most prior prospective studies have emphasized diagnostic performance or radiologist feedback, our findings extend beyond these metrics to include diagnostic efficacy, operational resilience, and system failure contingencies. These real-world insights support the feasibility and clinical value of embedding AI into routine emergency department workflows. Recent work has highlighted the importance of not only measuring accuracy but also assessing robustness across patient and workflow variability [<xref ref-type="bibr" rid="ref28">28</xref>]. Furthermore, the need for deployment frameworks that address hardware resilience, continuous quality monitoring, and interpretability safeguards is increasingly recognized as essential for sustainable AI adoption in high-acuity settings [<xref ref-type="bibr" rid="ref29">29</xref>].</p><p>Recent reports and position statements have highlighted a persistent gap between the promising diagnostic performance of AI systems and their limited demonstrated clinical benefit. Robust, prospective, and randomized clinical studies remain urgently needed to justify large-scale implementation [<xref ref-type="bibr" rid="ref30">30</xref>,<xref ref-type="bibr" rid="ref31">31</xref>]. Even high-performing AI models (area under the curve&#x2248;0.85) have failed to surpass standard clinical practice in improving patient outcomes [<xref ref-type="bibr" rid="ref32">32</xref>,<xref ref-type="bibr" rid="ref33">33</xref>]. These findings reinforce ongoing concerns that most AI or machine learning devices, despite regulatory authorization, are primarily validated using retrospective data and therefore remain susceptible to selection bias, distributional shift, and overestimation of generalizability [<xref ref-type="bibr" rid="ref34">34</xref>,<xref ref-type="bibr" rid="ref35">35</xref>].</p><p>Clinical decision-making in emergency care is inherently multimodal: physicians integrate imaging findings with the mechanism of injury, examination, and vital signs to guide judgment. In contrast, this AI system analyzes images in isolation and is designed not to replace but to support clinicians as a rapid screening aid&#x2014;enhancing vigilance in high-volume, high-pressure environments where missed fractures may occur. Incorporating multimodal clinical data in future models could further improve diagnostic relevance and workflow integration.</p><p>Although this was an observational feasibility study, it represents one of the largest evaluations of an AI-assisted rib fracture detection system in real-world emergency radiology. The findings demonstrate that such a system can provide meaningful diagnostic support, maintain consistent performance at scale, and potentially enhance patient safety. Future implementation should therefore shift from technical to clinical feasibility, focusing on clinician-in-the-loop impact studies, PACS-integrated trials, and workflow efficiency assessments. Although user perception was not formally assessed, informal feedback from emergency physicians indicated strong interest in AI-supported flagging&#x2014;particularly for subtle fractures and during periods of high patient volume.</p><p>Future research should prioritize prospective, multicenter studies to validate generalizability and quantify AI impact on workflow, resource utilization, and patient outcomes. Model improvements&#x2014;including artifact-aware retraining, expanded fracture coverage in challenging scenarios such as subtle or anatomically obscured fractures, and continuous learning&#x2014;will be critical to enhance diagnostic precision. Finally, building infrastructure resilience and integrating effective alert management into radiologist workflows are essential for sustainable clinical adoption.</p></sec><sec id="s4-4"><title>Limitations</title><p>First, this single-center observational study may limit generalizability to other institutions with different imaging protocols, patient populations, or workflow environments. Second, during retrospective model development, training and validation were performed at the image level rather than the patient level. This may have resulted in optimistic internal validation estimates due to potential within-patient similarity, which likely contributes to the observed performance gap between retrospective validation and prospective real-world deployment. Additionally, stratified performance analyses by age group and sex were not performed due to the low prevalence of rib fractures in certain subgroups, particularly pediatric patients. Similarly, the small fraction of oblique views (approximately 1.3%) prevented a dedicated analysis by imaging view, as the limited sample size would yield unstable estimates for these specific cohorts.</p><p>The prospective evaluation period also did not include winter months. Seasonal variation in trauma mechanisms or imaging artifacts may influence fracture detectability in certain settings; therefore, caution is warranted when generalizing these findings across different seasonal contexts.</p><p>Additionally, the use of a 512&#x00D7;512 resolution for model inference represents a technical trade-off; while it facilitates rapid processing, the associated downsampling may limit the detection of very subtle cortical disruptions.</p><p>Third, using radiology reports as the reference standard&#x2014;while pragmatic&#x2014;may underestimate the AI system&#x2019;s true performance, as subtle or occult fractures can be underreported in clinical practice. A focused CT review of representative discordant cases further supported this concern, revealing instances where AI-predicted fractures were subsequently confirmed as true fractures on CT. This approach likely yielded conservative performance estimates, since NLP-derived labels may not capture subtle fractures identified by AI or CT.</p><p>Regarding the study design, although post hoc adjudication is generally more appropriate for hypothesis generation than for definitive performance reassessment, modifying the reference standard after study completion may introduce bias. Accordingly, in this study, performance evaluation was anchored to the contemporaneous clinical reference standard used in routine practice. More comprehensive adjudication strategies&#x2014;such as consensus radiologist review of AI-positive, report-negative cases&#x2014;may provide additional insights when implemented within a separately designed study.</p><p>Finally, because AI predictions were not disclosed to clinicians, we did not assess downstream clinical outcomes, including changes in diagnostic behavior, time to intervention, or patient management. As for the data pipeline, although we used NLP to extract rib fracture labels from radiology reports, which may introduce misclassification in ambiguous cases, the pipeline demonstrated high agreement with manual review (&#x03BA;=0.91). Given the observed 3.5% discrepancy rate, any residual label noise propagating through the large-scale dataset may introduce modest uncertainty into performance estimates. However, the substantial sample size (n=23,251) is expected to attenuate the impact of such noise, supporting the stability of the resulting confidence intervals for large-scale clinical benchmarking.</p></sec><sec id="s4-5"><title>Conclusions</title><p>In this observational feasibility study, we evaluated a faster R-CNN&#x2013;based AI system deployed in parallel with clinical workflows to automatically detect rib fractures on CXRs using real-world emergency department data. Although AI outputs were not visible to clinicians, the system processed over 23,000 studies with high throughput, achieving 74.5% sensitivity and 93.3% specificity and delivering results within seconds&#x2014;over 1000 times faster than formal radiologist reports.</p><p>These findings demonstrate strong technical feasibility of real-time AI-assisted rib fracture detection in emergency radiology. While clinical decisions remained unaffected during this observational phase, future studies should validate clinical feasibility through clinician-in-the-loop evaluation, PACS integration, and workflow optimization to address potential alert fatigue and false-positive management.</p></sec></sec></body><back><ack><p>We thank the emergency department, radiology department, and information technology teams for their support during system deployment and data extraction. Generative artificial intelligence tools were used solely for language editing and grammar checking. All scientific content, analyses, interpretations, and conclusions were developed by the authors. The artificial intelligence system evaluated in this study was deployed on the DeepQ AI platform as a technical deployment environment. The authors have no financial, employment, or equity relationship with DeepQ Technology or HTC, and the platform was used solely for noncommercial research purposes.</p></ack><notes><sec><title>Funding</title><p>This research was financially supported by National Science and Technology Council, Taiwan, under Grant No. NSTC113-2221-E-038&#x2010;006</p></sec><sec><title>Data Availability</title><p>The imaging data analyzed in this study are not publicly available due to institutional regulations and patient privacy protections. Deidentified data may be made available from the corresponding author upon reasonable request and subject to institutional review board approval.</p></sec></notes><fn-group><fn fn-type="con"><p>Conceptualization: H-WC, S-TH</p><p>Methodology: H-WC, S-TH</p><p>Data curation: M-YH, S-TH</p><p>Formal analysis: M-YH, S-TH</p><p>Investigation: M-YH, S-TH</p><p>Writing &#x2013; original draft: M-YH, S-TH</p><p>Writing &#x2013; review &#x0026; editing: H-WC, L-RL, M-FT, M-YH, S-TH</p><p>Supervision: CH-W</p></fn><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AI</term><def><p>artificial intelligence</p></def></def-item><def-item><term id="abb2">AP</term><def><p>anteroposterior</p></def></def-item><def-item><term id="abb3">CLAIM</term><def><p>Checklist for Artificial Intelligence in Medical Imaging</p></def></def-item><def-item><term id="abb4">CT</term><def><p>computed tomography</p></def></def-item><def-item><term id="abb5">CXR</term><def><p>chest radiograph</p></def></def-item><def-item><term id="abb6">FROC</term><def><p>free-response receiver operating characteristic</p></def></def-item><def-item><term id="abb7">GPU</term><def><p>graphics processing unit</p></def></def-item><def-item><term id="abb8">NLP</term><def><p>natural language processing</p></def></def-item><def-item><term id="abb9">NPV</term><def><p>negative predictive value</p></def></def-item><def-item><term id="abb10">PACS</term><def><p>picture archiving and communication system</p></def></def-item><def-item><term id="abb11">R-CNN</term><def><p>region-based convolutional neural network</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Tignanelli</surname><given-names>CJ</given-names> </name><name name-style="western"><surname>Rix</surname><given-names>A</given-names> </name><name name-style="western"><surname>Napolitano</surname><given-names>LM</given-names> </name><name name-style="western"><surname>Hemmila</surname><given-names>MR</given-names> </name><name name-style="western"><surname>Ma</surname><given-names>S</given-names> </name><name name-style="western"><surname>Kummerfeld</surname><given-names>E</given-names> </name></person-group><article-title>Association between adherence to evidence-based practices for treatment of patients with traumatic rib fractures and mortality rates among US trauma centers</article-title><source>JAMA Netw Open</source><year>2020</year><month>03</month><day>2</day><volume>3</volume><issue>3</issue><fpage>e201316</fpage><pub-id pub-id-type="doi">10.1001/jamanetworkopen.2020.1316</pub-id><pub-id pub-id-type="medline">32215632</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Colling</surname><given-names>KP</given-names> </name><name name-style="western"><surname>Goettl</surname><given-names>T</given-names> </name><name name-style="western"><surname>Harry</surname><given-names>ML</given-names> </name></person-group><article-title>Outcomes after rib fractures: more complex than a single number</article-title><source>J Trauma Inj</source><year>2022</year><month>12</month><volume>35</volume><issue>4</issue><fpage>268</fpage><lpage>276</lpage><pub-id pub-id-type="doi">10.20408/jti.2021.0096</pub-id><pub-id pub-id-type="medline">39380934</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sharma</surname><given-names>OP</given-names> </name><name name-style="western"><surname>Oswanski</surname><given-names>MF</given-names> </name><name name-style="western"><surname>Jolly</surname><given-names>S</given-names> </name><name name-style="western"><surname>Lauer</surname><given-names>SK</given-names> </name><name name-style="western"><surname>Dressel</surname><given-names>R</given-names> </name><name name-style="western"><surname>Stombaugh</surname><given-names>HA</given-names> </name></person-group><article-title>Perils of rib fractures</article-title><source>Am Surg</source><year>2008</year><month>04</month><volume>74</volume><issue>4</issue><fpage>310</fpage><lpage>314</lpage><pub-id pub-id-type="doi">10.1177/000313480807400406</pub-id><pub-id pub-id-type="medline">18453294</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Griffith</surname><given-names>JF</given-names> </name><name name-style="western"><surname>Rainer</surname><given-names>TH</given-names> </name><name name-style="western"><surname>Ching</surname><given-names>AS</given-names> </name><name name-style="western"><surname>Law</surname><given-names>KL</given-names> </name><name name-style="western"><surname>Cocks</surname><given-names>RA</given-names> </name><name name-style="western"><surname>Metreweli</surname><given-names>C</given-names> </name></person-group><article-title>Sonography compared with radiography in revealing acute rib fracture</article-title><source>AJR Am J Roentgenol</source><year>1999</year><month>12</month><volume>173</volume><issue>6</issue><fpage>1603</fpage><lpage>1609</lpage><pub-id pub-id-type="doi">10.2214/ajr.173.6.10584808</pub-id><pub-id pub-id-type="medline">10584808</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Tomas</surname><given-names>X</given-names> </name><name name-style="western"><surname>Facenda</surname><given-names>C</given-names> </name><name name-style="western"><surname>Vaz</surname><given-names>N</given-names> </name><etal/></person-group><article-title>Thoracic wall trauma-misdiagnosed lesions on radiographs and usefulness of ultrasound, multidetector computed tomography and magnetic resonance imaging</article-title><source>Quant Imaging Med Surg</source><year>2017</year><month>08</month><volume>7</volume><issue>4</issue><fpage>384</fpage><lpage>397</lpage><pub-id pub-id-type="doi">10.21037/qims.2017.08.02</pub-id><pub-id pub-id-type="medline">28932697</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chapman</surname><given-names>BC</given-names> </name><name name-style="western"><surname>Overbey</surname><given-names>DM</given-names> </name><name name-style="western"><surname>Tesfalidet</surname><given-names>F</given-names> </name><etal/></person-group><article-title>Clinical utility of chest computed tomography in patients with rib fractures CT chest and rib fractures</article-title><source>Arch Trauma Res</source><year>2016</year><month>12</month><volume>5</volume><issue>4</issue><fpage>e37070</fpage><pub-id pub-id-type="doi">10.5812/atr.37070</pub-id><pub-id pub-id-type="medline">28144607</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Huang</surname><given-names>ST</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>LR</given-names> </name><name name-style="western"><surname>Chiu</surname><given-names>HW</given-names> </name><name name-style="western"><surname>Huang</surname><given-names>MY</given-names> </name><name name-style="western"><surname>Tsai</surname><given-names>MF</given-names> </name></person-group><article-title>Deep convolutional neural network for rib fracture recognition on chest radiographs</article-title><source>Front Med</source><year>2023</year><volume>10</volume><pub-id pub-id-type="doi">10.3389/fmed.2023.1178798</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Pishbin</surname><given-names>E</given-names> </name><name name-style="western"><surname>Ahmadi</surname><given-names>K</given-names> </name><name name-style="western"><surname>Foogardi</surname><given-names>M</given-names> </name><name name-style="western"><surname>Salehi</surname><given-names>M</given-names> </name><name name-style="western"><surname>Seilanian Toosi</surname><given-names>F</given-names> </name><name name-style="western"><surname>Rahimi-Movaghar</surname><given-names>V</given-names> </name></person-group><article-title>Comparison of ultrasonography and radiography in diagnosis of rib fractures</article-title><source>Chin J Traumatol</source><year>2017</year><month>08</month><volume>20</volume><issue>4</issue><fpage>226</fpage><lpage>228</lpage><pub-id pub-id-type="doi">10.1016/j.cjtee.2016.04.010</pub-id><pub-id pub-id-type="medline">28687342</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gulshan</surname><given-names>V</given-names> </name><name name-style="western"><surname>Peng</surname><given-names>L</given-names> </name><name name-style="western"><surname>Coram</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs</article-title><source>JAMA</source><year>2016</year><month>12</month><day>13</day><volume>316</volume><issue>22</issue><fpage>2402</fpage><lpage>2410</lpage><pub-id pub-id-type="doi">10.1001/jama.2016.17216</pub-id><pub-id pub-id-type="medline">27898976</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Rajpurkar</surname><given-names>P</given-names> </name><name name-style="western"><surname>Irvin</surname><given-names>J</given-names> </name><name name-style="western"><surname>Zhu</surname><given-names>K</given-names> </name><etal/></person-group><article-title>CheXNet: radiologist-level pneumonia detection on chest x-rays with deep learning</article-title><source>arXiv</source><comment>Preprint posted online on  Nov 14, 2017</comment><pub-id pub-id-type="doi">10.48550/arXiv.1711.05225</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Esteva</surname><given-names>A</given-names> </name><name name-style="western"><surname>Kuprel</surname><given-names>B</given-names> </name><name name-style="western"><surname>Novoa</surname><given-names>RA</given-names> </name><etal/></person-group><article-title>Dermatologist-level classification of skin cancer with deep neural networks</article-title><source>Nature</source><year>2017</year><month>02</month><day>2</day><volume>542</volume><issue>7639</issue><fpage>115</fpage><lpage>118</lpage><pub-id pub-id-type="doi">10.1038/nature21056</pub-id><pub-id pub-id-type="medline">28117445</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Litjens</surname><given-names>G</given-names> </name><name name-style="western"><surname>Kooi</surname><given-names>T</given-names> </name><name name-style="western"><surname>Bejnordi</surname><given-names>BE</given-names> </name><etal/></person-group><article-title>A survey on deep learning in medical image analysis</article-title><source>Med Image Anal</source><year>2017</year><month>12</month><volume>42</volume><fpage>60</fpage><lpage>88</lpage><pub-id pub-id-type="doi">10.1016/j.media.2017.07.005</pub-id><pub-id pub-id-type="medline">28778026</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lundervold</surname><given-names>AS</given-names> </name><name name-style="western"><surname>Lundervold</surname><given-names>A</given-names> </name></person-group><article-title>An overview of deep learning in medical imaging focusing on MRI</article-title><source>Z Med Phys</source><year>2019</year><month>05</month><volume>29</volume><issue>2</issue><fpage>102</fpage><lpage>127</lpage><pub-id pub-id-type="doi">10.1016/j.zemedi.2018.11.002</pub-id><pub-id pub-id-type="medline">30553609</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kora</surname><given-names>P</given-names> </name><name name-style="western"><surname>Ooi</surname><given-names>CP</given-names> </name><name name-style="western"><surname>Faust</surname><given-names>O</given-names> </name><etal/></person-group><article-title>Transfer learning techniques for medical image analysis: a review</article-title><source>Biocybern Biomed Eng</source><year>2022</year><month>01</month><volume>42</volume><issue>1</issue><fpage>79</fpage><lpage>107</lpage><pub-id pub-id-type="doi">10.1016/j.bbe.2021.11.004</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Olczak</surname><given-names>J</given-names> </name><name name-style="western"><surname>Fahlberg</surname><given-names>N</given-names> </name><name name-style="western"><surname>Maki</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Artificial intelligence for analyzing orthopedic trauma radiographs</article-title><source>Acta Orthop</source><year>2017</year><month>12</month><volume>88</volume><issue>6</issue><fpage>581</fpage><lpage>586</lpage><pub-id pub-id-type="doi">10.1080/17453674.2017.1344459</pub-id><pub-id pub-id-type="medline">28681679</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>N</given-names> </name><name name-style="western"><surname>Li</surname><given-names>X</given-names> </name><etal/></person-group><article-title>Convolutional neural network for detecting rib fractures on chest radiographs: a feasibility study</article-title><source>BMC Med Imaging</source><year>2023</year><month>01</month><day>30</day><volume>23</volume><issue>1</issue><fpage>18</fpage><pub-id pub-id-type="doi">10.1186/s12880-023-00975-x</pub-id><pub-id pub-id-type="medline">36717773</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lee</surname><given-names>K</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>S</given-names> </name><name name-style="western"><surname>Kwak</surname><given-names>JS</given-names> </name><name name-style="western"><surname>Park</surname><given-names>H</given-names> </name><name name-style="western"><surname>Oh</surname><given-names>H</given-names> </name><name name-style="western"><surname>Koh</surname><given-names>JC</given-names> </name></person-group><article-title>Development and validation of an artificial intelligence model for detecting rib fractures on chest radiographs</article-title><source>J Clin Med</source><year>2024</year><month>06</month><day>30</day><volume>13</volume><issue>13</issue><fpage>3850</fpage><pub-id pub-id-type="doi">10.3390/jcm13133850</pub-id><pub-id pub-id-type="medline">38999416</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Chang</surname><given-names>EY</given-names> </name></person-group><article-title>DeepQ: advancing healthcare through artificial intelligence and virtual reality</article-title><conf-name>MM &#x2019;17: Proceedings of the 25th ACM international conference on Multimedia New York</conf-name><conf-date>Oct 23-27, 2017</conf-date><pub-id pub-id-type="doi">10.1145/3123266.3130875</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Ren</surname><given-names>S</given-names> </name><name name-style="western"><surname>He</surname><given-names>K</given-names> </name><name name-style="western"><surname>Girshick</surname><given-names>R</given-names> </name><name name-style="western"><surname>Sun</surname><given-names>J</given-names> </name></person-group><article-title>Faster R-CNN: towards real-time object detection with region proposal networks</article-title><source>arXiv</source><comment>Preprint posted online on  Jun 4, 2015</comment><pub-id pub-id-type="doi">10.48550/arXiv.1506.01497</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yao</surname><given-names>L</given-names> </name><name name-style="western"><surname>Guan</surname><given-names>X</given-names> </name><name name-style="western"><surname>Song</surname><given-names>X</given-names> </name><etal/></person-group><article-title>Rib fracture detection system based on deep learning</article-title><source>Sci Rep</source><year>2021</year><month>12</month><day>6</day><volume>11</volume><issue>1</issue><fpage>23513</fpage><pub-id pub-id-type="doi">10.1038/s41598-021-03002-7</pub-id><pub-id pub-id-type="medline">34873241</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>van den Broek</surname><given-names>MCL</given-names> </name><name name-style="western"><surname>Buijs</surname><given-names>JH</given-names> </name><name name-style="western"><surname>Schmitz</surname><given-names>LFM</given-names> </name><name name-style="western"><surname>Wijffels</surname><given-names>MME</given-names> </name></person-group><article-title>Diagnostic performance of artificial intelligence in rib fracture detection: systematic review and meta-analysis</article-title><source>Surgeries</source><year>2024</year><volume>5</volume><issue>1</issue><fpage>24</fpage><lpage>36</lpage><pub-id pub-id-type="doi">10.3390/surgeries5010005</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kelly</surname><given-names>CJ</given-names> </name><name name-style="western"><surname>Karthikesalingam</surname><given-names>A</given-names> </name><name name-style="western"><surname>Suleyman</surname><given-names>M</given-names> </name><name name-style="western"><surname>Corrado</surname><given-names>G</given-names> </name><name name-style="western"><surname>King</surname><given-names>D</given-names> </name></person-group><article-title>Key challenges for delivering clinical impact with artificial intelligence</article-title><source>BMC Med</source><year>2019</year><month>10</month><day>29</day><volume>17</volume><issue>1</issue><fpage>195</fpage><pub-id pub-id-type="doi">10.1186/s12916-019-1426-2</pub-id><pub-id pub-id-type="medline">31665002</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Theriault-Lauzier</surname><given-names>P</given-names> </name><name name-style="western"><surname>Cobin</surname><given-names>D</given-names> </name><name name-style="western"><surname>Tastet</surname><given-names>O</given-names> </name><etal/></person-group><article-title>A responsible framework for applying artificial intelligence on medical images and signals at the point of care: the PACS-AI platform</article-title><source>Can J Cardiol</source><year>2024</year><month>10</month><volume>40</volume><issue>10</issue><fpage>1828</fpage><lpage>1840</lpage><pub-id pub-id-type="doi">10.1016/j.cjca.2024.05.025</pub-id><pub-id pub-id-type="medline">38885787</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Herpe</surname><given-names>G</given-names> </name><name name-style="western"><surname>Nelken</surname><given-names>H</given-names> </name><name name-style="western"><surname>Vendeuvre</surname><given-names>T</given-names> </name><etal/></person-group><article-title>Effectiveness of an artificial intelligence software for limb radiographic fracture recognition in an emergency department</article-title><source>J Clin Med</source><year>2024</year><month>09</month><day>20</day><volume>13</volume><issue>18</issue><fpage>5575</fpage><pub-id pub-id-type="doi">10.3390/jcm13185575</pub-id><pub-id pub-id-type="medline">39337062</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhou</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Qin</surname><given-names>P</given-names> </name><name name-style="western"><surname>Luo</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Evaluating AI rib fracture detections using follow-up CT scans</article-title><source>Am J Emerg Med</source><year>2023</year><month>10</month><volume>72</volume><fpage>34</fpage><lpage>38</lpage><pub-id pub-id-type="doi">10.1016/j.ajem.2023.07.018</pub-id><pub-id pub-id-type="medline">37478635</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Brady</surname><given-names>AP</given-names> </name></person-group><article-title>Error and discrepancy in radiology: inevitable or avoidable?</article-title><source>Insights Imaging</source><year>2017</year><month>02</month><volume>8</volume><issue>1</issue><fpage>171</fpage><lpage>182</lpage><pub-id pub-id-type="doi">10.1007/s13244-016-0534-1</pub-id><pub-id pub-id-type="medline">27928712</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sun</surname><given-names>H</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>X</given-names> </name><name name-style="western"><surname>Li</surname><given-names>Z</given-names> </name><etal/></person-group><article-title>Automated rib fracture detection on chest X-ray using contrastive learning</article-title><source>J Digit Imaging</source><year>2023</year><month>10</month><volume>36</volume><issue>5</issue><fpage>2138</fpage><lpage>2147</lpage><pub-id pub-id-type="doi">10.1007/s10278-023-00868-z</pub-id><pub-id pub-id-type="medline">37407842</pub-id></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hong</surname><given-names>WS</given-names> </name><name name-style="western"><surname>Haimovich</surname><given-names>AD</given-names> </name><name name-style="western"><surname>Taylor</surname><given-names>RA</given-names> </name></person-group><article-title>Predicting hospital admission at emergency department triage using machine learning</article-title><source>PLoS ONE</source><year>2018</year><volume>13</volume><issue>7</issue><fpage>e0201016</fpage><pub-id pub-id-type="doi">10.1371/journal.pone.0201016</pub-id><pub-id pub-id-type="medline">30028888</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Moskalenko</surname><given-names>V</given-names> </name><name name-style="western"><surname>Kharchenko</surname><given-names>V</given-names> </name></person-group><article-title>Resilience-aware MLOps for AI-based medical diagnostic system</article-title><source>Front Public Health</source><year>2024</year><volume>12</volume><fpage>1342937</fpage><pub-id pub-id-type="doi">10.3389/fpubh.2024.1342937</pub-id><pub-id pub-id-type="medline">38601490</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Khera</surname><given-names>R</given-names> </name><name name-style="western"><surname>Butte</surname><given-names>AJ</given-names> </name><name name-style="western"><surname>Berkwits</surname><given-names>M</given-names> </name><etal/></person-group><article-title>AI in medicine-JAMA&#x2019;s focus on clinical outcomes, patient-centered care, quality, and equity</article-title><source>JAMA</source><year>2023</year><month>09</month><day>5</day><volume>330</volume><issue>9</issue><fpage>818</fpage><lpage>820</lpage><pub-id pub-id-type="doi">10.1001/jama.2023.15481</pub-id><pub-id pub-id-type="medline">37566406</pub-id></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Armoundas</surname><given-names>AA</given-names> </name><name name-style="western"><surname>Narayan</surname><given-names>SM</given-names> </name><name name-style="western"><surname>Arnett</surname><given-names>DK</given-names> </name><etal/></person-group><article-title>Use of artificial intelligence in improving outcomes in heart disease: a scientific statement from the American Heart Association</article-title><source>Circulation</source><year>2024</year><month>04</month><day>2</day><volume>149</volume><issue>14</issue><fpage>e1028</fpage><lpage>e1050</lpage><pub-id pub-id-type="doi">10.1161/CIR.0000000000001201</pub-id><pub-id pub-id-type="medline">38415358</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mazor</surname><given-names>T</given-names> </name><name name-style="western"><surname>Farhat</surname><given-names>KS</given-names> </name><name name-style="western"><surname>Trukhanov</surname><given-names>P</given-names> </name><etal/></person-group><article-title>Clinical trial notifications triggered by artificial intelligence-detected cancer progression: a randomized trial</article-title><source>JAMA Netw Open</source><year>2025</year><month>04</month><day>1</day><volume>8</volume><issue>4</issue><fpage>e252013</fpage><pub-id pub-id-type="doi">10.1001/jamanetworkopen.2025.2013</pub-id><pub-id pub-id-type="medline">40257799</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mandair</surname><given-names>D</given-names> </name><name name-style="western"><surname>Elia</surname><given-names>MV</given-names> </name><name name-style="western"><surname>Hong</surname><given-names>JC</given-names> </name></person-group><article-title>Considerations in translating AI to improve care</article-title><source>JAMA Netw Open</source><year>2025</year><month>04</month><day>1</day><volume>8</volume><issue>4</issue><fpage>e252023</fpage><pub-id pub-id-type="doi">10.1001/jamanetworkopen.2025.2023</pub-id><pub-id pub-id-type="medline">40257804</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Habib</surname><given-names>AR</given-names> </name><name name-style="western"><surname>Gross</surname><given-names>CP</given-names> </name></person-group><article-title>FDA regulations of AI-driven clinical decision support devices fall short</article-title><source>JAMA Intern Med</source><year>2023</year><month>12</month><day>1</day><volume>183</volume><issue>12</issue><fpage>1401</fpage><lpage>1402</lpage><pub-id pub-id-type="doi">10.1001/jamainternmed.2023.5006</pub-id><pub-id pub-id-type="medline">37812411</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chouffani El Fassi</surname><given-names>S</given-names> </name><name name-style="western"><surname>Abdullah</surname><given-names>A</given-names> </name><name name-style="western"><surname>Fang</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>Not all AI health tools with regulatory authorization are clinically validated</article-title><source>Nat Med</source><year>2024</year><month>10</month><volume>30</volume><issue>10</issue><fpage>2718</fpage><lpage>2720</lpage><pub-id pub-id-type="doi">10.1038/s41591-024-03203-3</pub-id><pub-id pub-id-type="medline">39187696</pub-id></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Checklist 1</label><p>CLAIM (Checklist for Artificial Intelligence in Medical Imaging).</p><media xlink:href="medinform_v14i1e77965_app1.pdf" xlink:title="PDF File, 3 KB"/></supplementary-material></app-group></back></article>