Original Paper
Abstract
Background: Large language models show promise for improving radiology workflows, but their performance on structured radiological tasks such as Reporting and Data Systems (RADS) categorization remains unexplored.
Objective: This study aims to evaluate 3 large language model chatbots—Claude-2, GPT-3.5, and GPT-4—on assigning RADS categories to radiology reports and assess the impact of different prompting strategies.
Methods: This cross-sectional study compared 3 chatbots using 30 radiology reports (10 per RADS criteria), using a 3-level prompting strategy: zero-shot, few-shot, and guideline PDF-informed prompts. The cases were grounded in Liver Imaging Reporting & Data System (LI-RADS) version 2018, Lung CT (computed tomography) Screening Reporting & Data System (Lung-RADS) version 2022, and Ovarian-Adnexal Reporting & Data System (O-RADS) magnetic resonance imaging, meticulously prepared by board-certified radiologists. Each report underwent 6 assessments. Two blinded reviewers assessed the chatbots’ response at patient-level RADS categorization and overall ratings. The agreement across repetitions was assessed using Fleiss κ.
Results: Claude-2 achieved the highest accuracy in overall ratings with few-shot prompts and guideline PDFs (prompt-2), attaining 57% (17/30) average accuracy over 6 runs and 50% (15/30) accuracy with k-pass voting. Without prompt engineering, all chatbots performed poorly. The introduction of a structured exemplar prompt (prompt-1) increased the accuracy of overall ratings for all chatbots. Providing prompt-2 further improved Claude-2’s performance, an enhancement not replicated by GPT-4. The interrun agreement was substantial for Claude-2 (k=0.66 for overall rating and k=0.69 for RADS categorization), fair for GPT-4 (k=0.39 for both), and fair for GPT-3.5 (k=0.21 for overall rating and k=0.39 for RADS categorization). All chatbots showed significantly higher accuracy with LI-RADS version 2018 than with Lung-RADS version 2022 and O-RADS (P<.05); with prompt-2, Claude-2 achieved the highest overall rating accuracy of 75% (45/60) in LI-RADS version 2018.
Conclusions: When equipped with structured prompts and guideline PDFs, Claude-2 demonstrated potential in assigning RADS categories to radiology cases according to established criteria such as LI-RADS version 2018. However, the current generation of chatbots lags in accurately categorizing cases based on more recent RADS criteria.
doi:10.2196/55799
Keywords
Introduction
Since ChatGPT’s public release in November 2022, large language models (LLMs) have attracted great interest in medical imaging applications [
]. Research indicated that ChatGPT showed promising applications in various aspects of the medical imaging process. Even without radiology-specific pretraining, LLMs can pass board examinations [ ], provide radiology decision support [ ], assist in differential diagnosis [ - ], and generate impressions from findings or structured reports [ - ]. These applications not only accelerate the imaging diagnosis process and alleviate the workload of doctors but also improve the accuracy of diagnosis [ ]. However, limitations exist, with 1 study showing ChatGPT-3 producing erroneous answers for a third of daily clinical questions and about 63% of provided references were not found [ ]. ChatGPT’s dangerous tendency to produce inaccurate responses is less frequent in GPT-4 but still limits usability in medical education and practice at present [ ]. Tailoring LLMs to radiology may enhance reliability, as an appropriateness criteria context aware chatbot outperformed generic chatbots and radiologists [ ].The American College of Radiology Reporting and Data Systems (RADS) standardizes communication of imaging findings. As of August 2023, there have been 9 disease-specific systems endorsed by the American College of Radiology, referring to products from the lexicons to report templates [
]. RADS reduces terminology variability, facilitates communication between radiologists and referring physicians, allows consistent evaluations, and conveys clinical significance to improve care. However, complexity and unfamiliarity limit adoption. Consequently, endeavors should be pursued to broaden the implementation of RADS. Therefore, we conducted this study to evaluate LLM’s capabilities on a focused RADS assignment task for radiology reports.A prompt serves as a directive or instruction given to LLMs to generate a particular response. The technique of “prompt tuning” has emerged as a valuable approach to refine the performance of LLMs, particularly for specific domains or tasks [
]. By providing structured queries or exemplary responses, the output of chatbots can be tailored for accurate and relevant answers. Such prompt-tuning strategies leverage LLMs’ knowledge while guiding appropriate delivery for particular challenges [ ]. Given the complexity and specificity of the RADS categorization, our investigation emphasizes different prompt impacts to assess chatbot capabilities and potential performance enhancement through refined prompting tuning.In this study, our primary objective was to meticulously evaluate the performance of 3 LLMs (GPT-3.5, GPT-4, and Claude-2) for RADS categorization using different prompt-tuning strategies. We aimed to test their accuracy and consistency in RADS categorization and shed light on the potential benefits and limitations of relying on chatbot-derived information for the categorization of specific RADS.
Methods
Ethical Considerations
As the study was based on radiological data that were artificially generated by radiologists and did not involve the participation of human subjects, the study was determined to be exempt from ethical review, in accordance with the regulations established by the institutional review board of Henan Provincial People’s Hospital.
Study Design
The workflow of the study is shown in
. We conducted a cross-sectional analysis in September 2023 to evaluate the competency of 3 chatbots—GPT-3.5, GPT-4 (OpenAI, August 30, 2023 version) [ ], and Claude-2 (Anthropic) [ ]—in the task of assigning 3 RADS categorizations to radiology reports. Given the chatbot’s knowledge cessation was as of September 2021, we opted for Liver Imaging Reporting & Data System (LI-RADS) version 2018 [ ], Lung CT (computed tomography) Screening Reporting & Data System (Lung-RADS) version 2022 [ ], and Ovarian-Adnexal Reporting & Data System (O-RADS) magnetic resonance imaging (MRI) (developed in 2022) [ ] as the yardsticks to compare the responses engendered by GPT-3.5, GPT-4, and Claude-2. A total of thirty radiology reports for either CT or MRI examinations were composed for this analysis, with 10 cases representing each of the 3 RADS reporting systems. The radiology reports used for testing were generated by radiologists with more than 10 years’ experience to correct the wording styles from real-life cases based on respective RADS systems. For each RADS (ie, LI-, Lung-, and O-RADS), we attempted to reflect the complexity and diversity so that the reports cover typical cases in clinical practice. Therefore, reports with 2-3 simple cases and 7-8 challenging cases were generated for 1 RADS. These include scenarios such as prior examination comparison, the presence of multiple nodules, extensive categorization under different RADS systems, and updates from the most recent LI-RADS and Lung-RADS guidelines. The characteristics of radiology reports for each RADS and the distribution of the number of the reports across the 3 RADS are shown in . The objective was to evaluate the performance of chatbots on a highly structured radiology workflow task involving cancer risk categorization based on structured report inputs. The study design focused on a defined use case to illuminate the strengths and limitations of existing natural language-processing technology in this radiology subdomain.Prompts
We collected and analyzed responses from GPT-3.5, GPT-4, and Claude-2 for each case. To mitigate bias, the radiological findings were presented individually via separate interactions, with corresponding responses saved for analysis. Three prompt templates were designed to elicit each RADS categorization along with explanatory rationale: Prompt-0 was a zero-shot prompt, merely introducing the RADS assignment task, such as “Your task is to follow Lung-RADS version 2022 guideline to give Lung-RADS category of the radiological findings delimited by angle brackets.”
Prompt-1 was a few-shot prompt, furnishing an exemplar of RADS categorization including the reasoning, summarized impression, and final category. The following is an example:
Your task is to follow Lung-RADS version 2022 guideline to give Lung-RADS category of the radiological findings delimited by angle brackets. “”“ < …Radiological Findings… > Answer:Rationale: {…} Overall: {…} Summary: {…} Lung-RADS Category: X ”“”
Prompt-2 distinctly instructed chatbots to consult the PDF of corresponding RADS guidelines, compensating for these chatbots’ lack of radiology-specific pretraining. For Claude-2, the PDF could be directly ingested, while GPT-4 required the use of an “Ask for PDF” plug-in to extract pertinent information [
, ].Each case was evaluated 6 times with each chatbot across the 3 prompt levels. The representative radiological reports and prompts are shown in
. The links to all the prompts and guideline PDFs are shown in .Evaluation of Chatbots
Two study authors (QW and HL) independently evaluated the following for each chatbot response in a blinded manner, with any discrepancies resolved by a third senior radiologist (YW). The following were assessed for each response:
- Patient-level RADS categorization: judged as correct, incorrect, or unsure. “Correct” denotes that the chatbot accurately identified the patient-level RADS category, irrespective of the rationale provided. “Unsure” denotes that the chatbot’s response failed to provide a decisive RADS category. For example, a response articulating that “a definitive Lung-RADS category cannot be assigned” would be categorized as “unsure.”
- Overall rating: assessed as either correct or incorrect. A response is judged incorrect if any errors (Es) are identified, including the following:
- E1: a factual extraction error that denotes the chatbots’ inability to paraphrase the radiological findings accurately, consequently misinterpreting the information.
- E2: hallucination, encompassing the fabrication of nonexistent RADS categories (E2a) and RADS criteria (E2b).
- E3: a reasoning error, which includes the incapacity to logically interpret the imaging description (E3a) and the RADS category accurately (E3b). The subtype errors for reasoning imaging description include the inability to reason lesion signal (E3ai), lesion size (E3aii), and enhancement (E3aiii) accurately.
- E4: an explanatory error, encompassing inaccurate elucidation of RADS category meaning (E4a) and erroneous explanation of the recommended management and follow-up corresponding to the RADS category (E4b).
If a chatbot’s feedback manifested any of the aforementioned infractions, it was labeled as incorrect, with the specific type of error documented. To assess the consistency of the evaluations, a k-pass voting method was also applied. Specifically, a case was deemed accurately categorized if it met the criteria in a minimum of 4 out of the 6 runs.
Statistical Analyses
The accuracy of the patient-level RADS categorization and overall rating for each chatbot was compared using the chi-square test. The agreement across the 6 repeated runs was assessed using Fleiss κ. Agreement strength was interpreted as follows: <0 signified poor, 0-0.20 indicated slight, 0.21-0.40 represented fair, 0.41-0.60 was interpreted as moderate, 0.61-0.80 denoted substantial, and 0.81-1 was characterized as almost perfect. Statistical significance was defined as 2-sided P<.05. All analyses were performed using R statistical software (version 4.1.2; The R Foundation).
Results
Performance of Chatbots
The performance of chatbots is shown in
and and , with the links to case-level details provided in . For the overall rating ( , average row and A), Claude-2 with prompt-2 demonstrated significantly higher average accuracy across the 30 cases than Claude-2 with prompt-0 (odds ratio [OR] 8.16; P<.001). GPT-4 with prompt-2 also showed improved average accuracy compared with GPT-4 with prompt-0, but the difference was not statistically significant (OR 3.19; P=.13). When using the k-pass voting method ( , k-pass voting row), Claude-2 with prompt-2 had significantly higher accuracy than Claude-2 with prompt-0 (OR 8.65; P=.002). Similarly, GPT-4 with prompt-2 was significantly more accurate than GPT-4 with prompt-0 (OR 11.98; P=.01). For the exact assignment of the patient-level RADS categorization ( , average row and B), Claude-2 with Prompt-2 showed significantly more average accuracy than Claude-2 with prompt-0 (P=.04).Chatbots and prompts | Prompt-0, n (%; 95% CI) | Prompt-1, n (%; 95% CI) | Prompt-2, n (%; 95% CI) | |||
GPT-3.5 | ||||||
Run 1 | 3 (10; 3-28) | 9 (30; 15-50) | N/Aa | |||
Run 2 | 3 (10; 3-28) | 9 (30; 15-50) | N/A | |||
Run 3 | 4 (13; 4-32) | 7 (23; 11-43) | N/A | |||
Run 4 | 4 (13; 4-32) | 5 (17; 6-35) | N/A | |||
Run 5 | 3 (10; 3-28) | 6 (20; 8-39) | N/A | |||
Run 6 | 3 (10; 3-28) | 4 (13; 4-32) | N/A | |||
Averageb | 3 (10; 3-28) | 7 (23; 11-43) | N/A | |||
K-pass votingc | 1 (3; 0-19) | 2 (7; 1-24) | N/A | |||
GPT-4 | ||||||
Run 1 | 4 (13; 4-32) | 11 (37; 21-56) | 12 (40; 23-59) | |||
Run 2 | 4 (13; 4-32) | 7 (23; 11-43) | 8 (27; 13-46) | |||
Run 3 | 4 (13; 4-32) | 9 (30; 15-50) | 9 (30; 15-50) | |||
Run 4 | 2 (7; 1-24) | 9 (30; 15-50) | 13 (43; 26-62) | |||
Run 5 | 5 (17; 6-35) | 11 (37; 21-56) | 8 (27; 13-46) | |||
Run 6 | 6 (20; 8-39) | 9 (30; 15-50) | 8 (27; 13-46) | |||
Averageb | 4 (13; 4-32) | 9 (30; 15-50) | 10 (33; 18-53) | |||
K-pass votingc | 1 (3; 0-19) | 6 (20; 8-39) | 9 (30; 15-50)d | |||
Claude-2 | ||||||
Run 1 | 4 (13; 4-32) | 10 (33; 18-53) | 19 (63; 44-79) | |||
Run 2 | 5 (17; 6-35) | 8 (27; 13-46) | 16 (53; 35-71) | |||
Run 3 | 5 (17; 6-35) | 7 (23; 11-43) | 15 (50; 33-67) | |||
Run 4 | 5 (17; 6-35) | 6 (20; 8-39) | 17 (57; 38-74) | |||
Run 5 | 3 (10; 3-28) | 7 (23; 11-43) | 18 (60; 41-77) | |||
Run 6 | 3 (10; 3-28) | 7 (23; 11-43) | 14 (47; 29-65) | |||
Averageb | 4 (13; 4-32) | 8 (27; 13-46) | 17 (57; 38-74)d | |||
K-pass votingc | 3 (10; 3-28) | 7 (23; 11-43) | 15 (50; 33-67)d |
aN/A: not applicable.
bAccuracy by the average method.
cAccuracy by k-pass voting (≥4/6 runs correct).
dSignificant between prompt-0 and prompt-2.
Chatbots and prompts | Correct/incorrect/unsure patient-level Reporting and Data Systems categories, n/n/n | ||||||||||||||||
Run 1 | Run 2 | Run 3 | Run 4 | Run 5 | Run 6 | Averagea | K-pass votingb | ||||||||||
GPT-3.5 | |||||||||||||||||
Prompt-0 | 7/23/0 | 7/23/0 | 7/23/0 | 9/21/0 | 8/21/1 | 8/20/2 | 8/22/0 | 7/23/0 | |||||||||
Prompt-1 | 13/15/2 | 11/19/0 | 8/21/1 | 8/21/1 | 11/19/0 | 8/22/0 | 10/20/0 | 7/23/0 | |||||||||
GPT-4 | |||||||||||||||||
Prompt-0 | 10/20/0 | 8/19/3 | 9/20/1 | 8/22/0 | 16/14/0 | 13/15/2 | 11/18/1 | 8/22/0 | |||||||||
Prompt-1 | 15/14/1 | 10/18/2 | 11/18/1 | 14/15/1 | 15/14/1 | 12/18/0 | 13/16/1 | 11/19/0 | |||||||||
Prompt-2 | 13/16/1 | 11/18/1 | 12/18/0 | 14/16/0 | 9/21/0 | 11/16/3 | 12/18/0 | 11/19/0 | |||||||||
Claude-2 | |||||||||||||||||
Prompt-0 | 13/17/0 | 12/18/0 | 12/18/0 | 15/15/0 | 10/20/0 | 9/21/0 | 12/18/0 | 13/17/0 | |||||||||
Prompt-1 | 11/19/0 | 14/16/0 | 11/19/0 | 11/19/0 | 13/17/0 | 12/18/0 | 12/18/1 | 11/19/0 | |||||||||
Prompt-2 | 21/9/0 | 21/9/0 | 20/10/0 | 22/8/0 | 21/9/0 | 2021/8/1 | 21/9/0c | 21/9/0 |
aAccuracy by the average method.
bAccuracy by k-pass voting (≥4/6 runs correct).
cSignificant between prompt-0 and prompt-2.
Consistency of Chatbots
As shown in
, among the 30 cases evaluated in 6 runs, Claude-2 with prompt-2 showed substantial agreement (k=0.65 for overall rating; k=0.74 for RADS categorization). GPT-4, when interfaced with prompt-2, demonstrated moderate agreement (k=0.46 for overall rating; k=0.41 for RADS categorization). When evaluated with prompt-1, GPT-4 presented moderate agreement (k=0.38 for overall rating; k=0.42 for RADS categorization). In contrast, Claude-2 showed substantial agreement (k=0.63 for overall rating; k=0.61 for RADS categorization), while GPT-3.5 exhibited a range from slight to fair agreement. With prompt-0, Claude-2 showed moderate agreement (k=0.49) for overall rating and substantial agreement for RADS categorization (k=0.65). GPT4 manifested slight agreement (k=0.19) for the overall rating and fair agreement for RADS categorization. Meanwhile, GPT-3.5 showed fair agreement (k=0.28) for the overall rating and moderate agreement (k=0.57) for RADS categorization.Prompt-0, Fleiss κ (95% CI) | Prompt-1, Fleiss κ (95% CI) | Prompt-2, Fleiss κ (95% CI) | All, Fleiss κ (95% CI) | |||||
Patient-level RADSacategorization | ||||||||
GPT-3.5 | 0.57 (0.48-0.65) | 0.24 (0.15-0.32) | N/Ab | 0.39 (0.33-0.46) | ||||
GPT-4 | 0.33 (0.25-0.42) | 0.42 (0.34-0.5) | 0.41 (0.33-0.5) | 0.39 (0.34-0.44) | ||||
Claude-2 | 0.65 (0.56-0.74) | 0.61 (0.52-0.7) | 0.74 (0.65-0.83) | 0.69 (0.64-0.74) | ||||
Overall rating | ||||||||
GPT-3.5 | 0.28 (0.19-0.37) | 0.14 (0.05-0.23) | N/A | 0.21 (0.14-0.27) | ||||
GPT-4 | 0.19 (0.1-0.28) | 0.38 (0.29-0.47) | 0.46 (0.37-0.55) | 0.39 (0.34-0.45) | ||||
Claude-2 | 0.49 (0.4-0.58) | 0.63 (0.53-0.72) | 0.65 (0.56-0.75) | 0.66 (0.61-0.72) |
aRADS: Reporting and Data Systems.
bN/A: not applicable.
Subgroup Analysis
Since the knowledge base for ChatGPT was frozen as of September 2021, accounting for the knowledge limitations of LLMs developed before the latest RADS guideline updates, we compared the responses of different RADS criteria. The total accurate responses across 6 runs were computed for all prompts. Both GPT-4 and Claude-2 demonstrated superior performance in the context of LI-RADS CT/MRI version 2018 as opposed to Lung-RADS version 2022 and O-RADS MRI (all P<.05;
). delineates the performance of various chatbots across different prompts and RADS categories. For the overall rating ( A), Claude-2 exhibited a progressive trend of enhancement of overall rating accuracy from prompt-0 to prompt-1 to prompt-2, with 20.0% (12/60), 36.7% (22/60), and 75.0% (45/60) for LIRADS; 11.7% (7/60), 18.3% (11/60), and 48.3% (29/60) for Lung-RADS; and 10.0% (6/60), 20.0% (12/60), and 41.7% (25/60) for O-RADS, respectively. Notably, with prompt-2, Claude-2 achieved the highest overall rating accuracy of 75% in older systems such as LI-RADS version 2018. Conversely, GPT-4 improved with prompt-1/2 over prompt-0, but prompt-2 did not exceed prompt-1. For the RADS categorization ( B), prompt-1 and prompt-2 outperformed prompt-0 for LI-RADS, irrespective of chatbots. However, for Lung-RADS and O-RADS, prompt-0 sometimes superseded prompt-1.Chatbots and RADSb | Year of development | RADS categorization (correct/incorrect/unsure), n/n/n | P value | Overall rating (correct/incorrect), n/n | P value | |||||
GPT-3.5 | ||||||||||
LI-RADSc CTd/MRIe | 2018 | 32/86/2 | Reference | 22/98 | Reference | |||||
Lung-RADSf | 2022 | 38/78/4 | .83 | 14/106 | .15 | |||||
O-RADSg MRI | 2022 | 35/84/1 | .46 | 24/96 | .87 | |||||
GPT-4 | ||||||||||
LI-RADS CT/MRI | 2018 | 104/74/2 | Reference | 78/102 | Reference | |||||
Lung-RADS | 2022 | 40/128/12 | <.001 | 21/159 | <.001 | |||||
O-RADS MRI | 2022 | 67/110/3 | <.001 | 40/140 | <.001 | |||||
Claude-2 | ||||||||||
LI-RADS CT/MRI | 2018 | 93/86/1 | Reference | 79/101 | Reference | |||||
Lung-RADS | 2022 | 63/117/0 | .001 | 47/133 | <.001 | |||||
O-RADS MRI | 2022 | 113/67/0 | .04 | 43/137 | <.001 |
aData are aggregate numbers across 6 runs.
bRADS: Reporting and Data Systems.
cLI-RADS: Liver Imaging Reporting and Data System.
dCT: computed tomography.
eMRI: magnetic resonance imaging.
fLung-RADS: Lung CT Screening Reporting and Data System.
gO-RADS: Ovarian-Adnexal Reporting and Data System.
Analysis of Error Types
A total of 1440 cases were analyzed for error types, with details provided in
. The bar plot illustrating the distribution of errors across the 3 chatbots is shown in . A typical example of factual extraction error (E1) occurred in response to the seventh Lung-RADS question. The statement “The 3mm solid nodule in the lateral basal segmental bronchus is subsegmental” is inaccurate, as the lateral basal segmental bronchus represents one of the 18 defined lung segments and not a subsegment [ ].Hallucination of inappropriate RADS categories (E2a) occurred more frequently with prompt-0 across all 3 chatbots. However, this error rate decreased to zero for Claude-2 when using prompt-2, a trend not seen with GPT-3.5 or GPT-4. A recurrent E2a error in LI-RADS was the obsolete category LR-5V from the 2014 version, now superseded by LR-TIV in subsequent editions [
, ]. Furthermore, hallucination of invalid RADS criteria (E2b) was more prevalent than that of E2a. For instance, the LI-RADS second question response stating “T2 marked hyperintensity is a feature commonly associated with hepatocellular carcinoma (HCC)” is inaccurate, as T2-marked hyperintensity is characteristic of hemangioma and not hepatocellular carcinoma. Despite initial higher E2b rates, Claude-2 demonstrated a substantial reduction with prompt-2 (105 to 38 instances), exceeding the decrement seen with GPT-4 (71 to 57 instances).Regarding reasoning error, incorrect RADS category reasoning (E3b) was the most frequent error but decreased for all chatbots with prompt-1 and prompt-2 versus prompt-0. Claude-2 reduced errors by almost half with prompt-2, while the GPT-4 decrease was less pronounced. Lesion signal interpretation errors (E3ai) included misinterpreting hypointensity on diffusion-weighted imaging as “restricted diffusion,” rather than facilitated diffusion. Lesion size reasoning errors (E3aii) occurred in 34 of 1440 cases, predominantly by Claude-2 (25/34, 73.5%), especially in systems such as Lung-RADS and LI-RADS where size is critical for categorization. Examples were attributing a 12-mm pulmonary nodule to the ≥6-mm but <8-mm range, or assigning a hepatic lesion measuring 2.3 cm × 1.5 cm to the 10- to 19-mm category. Reasoning enhancement errors (E3aiii) were exclusive to Claude-2 in O-RADS, where enhancement significantly impacts categorization. Misclassifying images at 40 seconds postcontrast as early or delayed enhancement exemplifies this error.
Explanatory errors (E4) including incorrect RADS category definitions (E4a) and inappropriate management recommendations (E4b) also substantially declined with prompt-1 and prompt-2. For instance, in the first Lung-RADS question response, the statement “The 4X designation indicates infectious/inflammatory etiology is suspected.” is incorrect. Lung-RADS 4X means category 3 or 4 nodules with additional features or imaging findings that increase suspicion of lung cancer [
].Discussion
Principal Findings
In this study, we evaluated the performance of 3 chatbots—GPT-3.5, GPT-4, and Claude-2—in categorizing radiological findings according to RADS criteria. Using 3 levels of prompts providing increasing structure, examples, and domain knowledge, the chatbots’ accuracies and consistencies were quantified across 30 cases. The best performance was achieved by Claude-2 when provided with few-shot prompting and the RADS criteria PDFs. Interestingly, the chatbots tended to categorize better for the relatively older LI-RADS version 2018 criteria in contrast to the more recent Lung-RADS version 2022 and O-RADS guidelines published after the chatbots’ training cutoff.
The incorporation of RADS, which standardizes reporting in radiology, has been a significant advancement, although the multiplicity and complexity of these systems impose a steep learning curve for radiologists [
]. Even for subspecialized radiologists at tertiary hospitals, mastering the numerous RADS guidelines poses challenges, requiring familiarity with the lexicons, regular application in daily practice, and ongoing learning to remain current with new versions. While previous studies have shown that LLMs could assist radiologists in various tasks [ - , , ], their performance at RADS categorization from imaging findings is untested. We therefore evaluated LLMs for focused RADS categorization of testing cases.Without prompt engineering (prompt-0), all chatbots performed poorly. However, accuracy improved for all chatbots when provided an exemplar prompt demonstrating the desired response structure (prompt-1). This underscores the use of prompt tuning for aligning LLMs to specific domains such as radiology. Further enriching prompt-1 with the RADS guideline PDFs as a relevant knowledge source (prompt-2) considerably enhanced Claude-2’s accuracy, a feat not mirrored by GPT-4. This discrepancy could stem from ChatGPT’s reliance on an external plug-in to access documents, while Claude-2’s architecture accommodates the direct assimilation of expansive texts, benefiting from its larger-context window and superior long document–processing capabilities.
Notably, we discerned performance disparities across RADS criteria. When queried on older established guidelines such as LI-RADS version 2018 [
], the chatbots demonstrated greater accuracy than more recent schemes such as Lung-RADS version 2022 and O-RADS [ , , ]. Specifically, GPT-4 and Claude-2 had significantly higher total correct ratings for LI-RADS than for Lung-RADS and O-RADS (all P<.05). This could be attributed to their extensive exposure to the voluminous data related to the matured LI-RADS during their pretraining phase. With prompt-2, Claude-2 achieved 75% (45/60) accuracy for overall rating LI-RADS categorization. The poorer performance on newer RADS criteria highlights the need for strategies to continually align LLMs with the most up-to-date knowledge.A deep dive into the error-type analysis revealed informative trends. Incorrect RADS category reasoning (E3b) constituted the most frequent error across chatbots, decreasing with prompt tuning. Targeted prompting also reduced critical errors such as hallucinations of RADS criteria (E2b) and categories (E2a) likely by constraining output to valid responses. During pretraining, GPT-liked LLMs predict the next word in the unlabeled data set, risking learning fallacious relationships between RADS features. For instance, Lung-RADS version 2022 lacks categories 5 and 6 [
], though some other RADS such as Breast Imaging Reporting and Data System include them [ ]. Using prompt-0, chatbots erroneously hallucinated Lung-RADS categories 5 and 6. Explanatory errors (E4) including inaccurate definition of the assigned RADS category (E4a) and inappropriate management recommendations (E4b) also substantially declined with prompt tuning. For instance, when queried on the novel O-RADS criteria with prompt-0, chatbots hallucinated follow-up recommendations from other RADS criteria and responded “O-RADS category 3 refers to an indeterminate adnexal mass and warrants short-interval follow-up.” Targeted prompting appears to mitigate these critical errors such as hallucination and incorrect reasoning. Careful prompt engineering is essential to properly shape LLM knowledge for radiology tasks.Limitations
There are also several limitations in this study. First, only the LI-RADS CT/MRI and O-RADS MRI were included, excluding LI-RADS ultrasound (US) and O-RADS US guidelines, which are often practiced in an independent US department [
, ]. Second, the chatbot’s performance was heavily dependent on prompt quality. We test only 3 types of prompts and further prompt strategies studies are warranted to investigate the impact of more exhaustive engineering on chatbots’ accuracy. Third, GPT-4-turbo was released on November 6, 2023, representing the latest GPT-4 model with improvements in instruction following, reproducible outputs, and more [ ]. Furthermore, its training data extend to April 2023 compared with September 2021 for the base GPT-4 model tested here. We are uncertain about this newest GPT-4-turbo model’s performance on the RADS categorization task. Evaluating GPT-4-turbo represents an important direction for future work. Fourth, our study focused on 3 of 9 RADS [ ], with a limited 10 cases for each RADS category. Although our choice ensured a blend of old and new guidelines and tried to cover all the RADS scores as much as possible, extending evaluations to all the RADS guidelines and incorporating more radiology reports from real clinical scenarios could offer deeper insights into potential limitations. Nonetheless, this initial study highlights critical considerations of prompt design and knowledge calibration required for safely applying LLMs in radiology. Fifth, evaluating the performance of the LLM in comparison with radiologists of varying expertise levels proves valuable for discerning its strengths and weaknesses in real-world applications. This comparative analysis will be undertaken in our forthcoming studies.Conclusions
When equipped with structured prompts and guideline PDFs, Claude-2 demonstrates potential in assigning RADS categories to radiology cases according to established criteria such as LI-RADS version 2018. However, the current generation of chatbots lags in accurately categorizing cases based on more recent RADS criteria. Our study highlights the potential of LLMs in streamlining radiological categorizations while also pinpointing the enhancements necessary for their dependable application in clinical practice for RADS categorization tasks.
Acknowledgments
This study has received funding from the National Natural Science Foundation of China (82371934 and 82001783) and Joint Fund of Henan Province Science and Technology R&D Program (225200810062). The authors thank Chuanjian Lv, MD; Zejun Wen, MM; and Jianghua Lou, MM, for their help in drafting the radiology reports with regard to Lung CT Screening Reporting and Data System, Liver Imaging Reporting and Data System, and Ovarian-Adnexal Reporting and Data System, respectively.
Authors' Contributions
QW (Henan Provincial People’s Hospital & People’s Hospital of Zhengzhou University), QW (Beijing United Imaging Research Institute of Intelligent Imaging), HL, Y Wang, YB, Y Wu, XY, and MW contributed to study design. QW (Henan Provincial People’s Hospital & People’s Hospital of Zhengzhou University) and QW (Beijing United Imaging Research Institute of Intelligent Imaging) contributed to the statistical analysis. All authors contributed to the acquisition, analysis, or interpretation of the data; the drafting of the manuscript; and critical revision of the manuscript.
Conflicts of Interest
QW and PD are senior engineers of Beijing United Imaging Research Institute of Intelligent Imaging and United Imaging Intelligence (Beijing) Co, Ltd. JX and DS are senior specialists of Shanghai United Imaging Intelligence Co, Ltd. The companies have no role in designing and performing the surveillance and analyzing and interpreting the data. All other authors report no conflicts of interest relevant to this article.
The characteristics of radiology reports for each of the Reporting and Data Systems (RADS) and the distribution of the number of the reports across the 3 RADS.
DOCX File , 107 KBRepresentative radiology reports and prompts.
DOCX File , 18 KBLinks to prompts and guideline PDFs.
DOCX File , 12 KBLinks to prompt engineering results.
DOCX File , 11 KBReferences
- Li R, Kumar A, Chen JH. How chatbots and large language model artificial intelligence systems will reshape modern medicine: fountain of creativity or Pandora's box? JAMA Intern Med. Jun 01, 2023;183(6):596-597. [CrossRef] [Medline]
- Bhayana R, Krishna S, Bleakney RR. Performance of ChatGPT on a radiology board-style examination: insights into current strengths and limitations. Radiology. Jun 2023;307(5):e230582. [CrossRef] [Medline]
- Rao A, Kim J, Kamineni M, Pang M, Lie W, Dreyer KJ, et al. Evaluating GPT as an adjunct for radiologic decision making: GPT-4 versus GPT-3.5 in a breast imaging pilot. J Am Coll Radiol. Oct 2023;20(10):990-997. [FREE Full text] [CrossRef] [Medline]
- Ueda D, Mitsuyama Y, Takita H, Horiuchi D, Walston SL, Tatekawa H, et al. ChatGPT's diagnostic performance from patient history and imaging findings on the diagnosis please quizzes. Radiology. Jul 2023;308(1):e231040. [CrossRef] [Medline]
- Kottlors J, Bratke G, Rauen P, Kabbasch C, Persigehl T, Schlamann M, et al. Feasibility of differential diagnosis based on imaging patterns using a large language model. Radiology. Jul 2023;308(1):e231167. [CrossRef] [Medline]
- Gertz RJ, Bunck AC, Lennartz S, Dratsch T, Iuga AI, Maintz D, et al. GPT-4 for automated determination of radiological study and protocol based on radiology request forms: a feasibility study. Radiology. 2023;307(5):e230877. [CrossRef] [Medline]
- Adams LC, Truhn D, Busch F, Kader A, Niehues SM, Makowski MR, et al. Leveraging GPT-4 for post hoc transformation of free-text radiology reports into structured reporting: a multilingual feasibility study. Radiology. May 2023;307(4):e230725. [CrossRef] [Medline]
- Wagner MW, Ertl-Wagner BB. Accuracy of information and references using ChatGPT-3 for retrieval of clinical radiological information. Can Assoc Radiol J. Feb 2024;75(1):69-73. [FREE Full text] [CrossRef] [Medline]
- Ziegelmayer S, Marka AW, Lenhart N, Nehls N, Reischl S, Harder F, et al. Evaluation of GPT-4's chest x-ray impression generation: a reader study on performance and perception. J Med Internet Res. Dec 22, 2023;25:e50865. [FREE Full text] [CrossRef] [Medline]
- Bhayana R, Bleakney RR, Krishna S. GPT-4 in radiology: improvements in advanced reasoning. Radiology. Jun 2023;307(5):e230987. [CrossRef] [Medline]
- Liu J, Wang C, Liu S. Utility of ChatGPT in clinical practice. J Med Internet Res. Jun 28, 2023;25:e48568. [FREE Full text] [CrossRef] [Medline]
- Rau A, Rau S, Zoeller D, Fink A, Tran H, Wilpert C, et al. A context-based chatbot surpasses trained radiologists and generic ChatGPT in following the ACR appropriateness guidelines. Radiology. Jul 2023;308(1):e230970. [CrossRef] [Medline]
- American College of Radiology. Reporting and Data Systems (RADS). URL: https://www.acr.org/Clinical-Resources/Reporting-and-Data-Systems [accessed 2023-08-26]
- Meskó B. Prompt engineering as an important emerging skill for medical professionals: tutorial. J Med Internet Res. Oct 04, 2023;25:e50638. [FREE Full text] [CrossRef] [Medline]
- OpenAI. URL: https://openai.com [accessed 2023-11-08]
- Anthropic. Claude 2. URL: https://www.anthropic.com/index/claude-2 [accessed 2023-11-08]
- Chernyak V, Fowler KJ, Kamaya A, Kielar AZ, Elsayes KM, Bashir MR, et al. Liver Imaging Reporting and Data System (LI-RADS) Version 2018: imaging of hepatocellular carcinoma in at-risk patients. Radiology. Dec 2018;289(3):816-830. [FREE Full text] [CrossRef] [Medline]
- Martin MD, Kanne JP, Broderick LS, Kazerooni EA, Meyer CA. Update: Lung-RADS 2022. Radiographics. Nov 2023;43(11):e230037. [CrossRef] [Medline]
- Sadowski EA, Thomassin-Naggara I, Rockall A, Maturen KE, Forstner R, Jha P, et al. O-RADS MRI risk stratification system: guide for assessing adnexal lesions from the ACR O-RADS Committee. Radiology. Apr 2022;303(1):35-47. [FREE Full text] [CrossRef] [Medline]
- AskYourPDF. URL: https://askyourpdf.com [accessed 2023-11-08]
- ChatGPT plugins. URL: https://openai.com/blog/chatgpt-plugins [accessed 2023-11-08]
- Jones J, Rasuli B, Vadera S. Bronchopulmonary segmental anatomy. URL: https://doi.org/10.53347/rID-13644 [accessed 2023-11-08]
- Mitchell DG, Bruix J, Sherman M, Sirlin CB. LI-RADS (Liver Imaging Reporting and Data System): summary, discussion, and consensus of the LI-RADS Management Working Group and future directions. Hepatology. Mar 2015;61(3):1056-1065. [CrossRef] [Medline]
- Elsayes K, Hooker J, Agrons M, Kielar A, Tang A, Fowler K, et al. 2017 Version of LI-RADS for CT and MR imaging: an update. Radiographics. 2017;37(7):1994-2017. [CrossRef] [Medline]
- Suarez-Weiss KE, Sadowski EA, Zhang M, Burk KS, Tran VT, Shinagare AB. Practical tips for reporting adnexal lesions using O-RADS MRI. Radiographics. 2023;43(7):e220142. [CrossRef] [Medline]
- American College of Radiology. Breast Imaging Reporting & Data System. URL: https://www.acr.org/Clinical-Resources/Reporting-and-Data-Systems/Bi-Rads [accessed 2023-11-03]
- Strachowski LM, Jha P, Phillips CH, Blanchette Porter MM, Froyman W, Glanc P, et al. O-RADS US v2022: an update from the American College of Radiology's Ovarian-Adnexal Reporting and Data System US Committee. Radiology. 2023;308(3):e230685. [CrossRef] [Medline]
- Quaia E. State of the art: LI-RADS for contrast-enhanced US. Radiology. 2019;293(1):4-14. [CrossRef] [Medline]
- New models and developer products announced at DevDay. URL: https://openai.com/blog/new-models-and-developer-products-announced-at-devday [accessed 2023-11-09]
Abbreviations
CT: computed tomography |
E: error |
LI-RADS: Liver Imaging Reporting & Data System |
LLM: large language model |
Lung-RADS: Lung CT Screening Reporting & Data System |
MRI: magnetic resonance imaging |
O-RADS: Ovarian-Adnexal Reporting & Data System |
OR: odds ratio |
RADS: Reporting and Data Systems |
US: ultrasound |
Edited by C Lovis; submitted 25.12.23; peer-reviewed by Z Liu, D Bu, TAR Sure, S Nuthakki, L Zhu; comments to author 14.01.24; revised version received 02.02.24; accepted 25.05.24; published 17.07.24.
Copyright©Qingxia Wu, Qingxia Wu, Huali Li, Yan Wang, Yan Bai, Yaping Wu, Xuan Yu, Xiaodong Li, Pei Dong, Jon Xue, Dinggang Shen, Meiyun Wang. Originally published in JMIR Medical Informatics (https://medinform.jmir.org), 17.07.2024.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on https://medinform.jmir.org/, as well as this copyright and license information must be included.