<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="review-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Med Inform</journal-id><journal-id journal-id-type="publisher-id">medinform</journal-id><journal-id journal-id-type="index">7</journal-id><journal-title>JMIR Medical Informatics</journal-title><abbrev-journal-title>JMIR Med Inform</abbrev-journal-title><issn pub-type="epub">2291-9694</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v13i1e68707</article-id><article-id pub-id-type="doi">10.2196/68707</article-id><article-categories><subj-group subj-group-type="heading"><subject>Review</subject></subj-group></article-categories><title-group><article-title>Performance of Natural Language Processing for Information Extraction From Electronic Health Records Within Cancer: Systematic Review</article-title></title-group><contrib-group><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Dahl</surname><given-names>Simon</given-names></name><degrees>MSc</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>B&#x00F8;gsted</surname><given-names>Martin</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff3">3</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Sagi</surname><given-names>Tomer</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Vesteghem</surname><given-names>Charles</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff3">3</xref></contrib></contrib-group><aff id="aff1"><institution>Center for Clinical Data Science, Department of Clinical Medicine, Aalborg University</institution><addr-line>Selma Lagerl&#x00F8;fs Vej 249</addr-line><addr-line>Gistrup</addr-line><country>Denmark</country></aff><aff id="aff2"><institution>Center for Clinical Data Science, Research, Education and Innovation, Aalborg University Hospital</institution><addr-line>Aalborg</addr-line><country>Denmark</country></aff><aff id="aff3"><institution>Clinical Cancer Research Centre, Department of Clinical Medicine, Aalborg University Hospital</institution><addr-line>Aalborg</addr-line><country>Denmark</country></aff><aff id="aff4"><institution>Department of Computer Science, Aalborg University</institution><addr-line>Aalborg</addr-line><country>Denmark</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Castonguay</surname><given-names>Alexandre</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Chaturvedi</surname><given-names>Akhil</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Chrimes</surname><given-names>Dillon</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Simon Dahl, MSc, Center for Clinical Data Science, Department of Clinical Medicine, Aalborg University, Selma Lagerl&#x00F8;fs Vej 249, Gistrup, 9260, Denmark, +45 99407244; <email>scd@rn.dk</email></corresp></author-notes><pub-date pub-type="collection"><year>2025</year></pub-date><pub-date pub-type="epub"><day>12</day><month>9</month><year>2025</year></pub-date><volume>13</volume><elocation-id>e68707</elocation-id><history><date date-type="received"><day>12</day><month>11</month><year>2024</year></date><date date-type="rev-recd"><day>16</day><month>06</month><year>2025</year></date><date date-type="accepted"><day>17</day><month>06</month><year>2025</year></date></history><copyright-statement>&#x00A9; Simon Dahl, Martin B&#x00F8;gsted, Tomer Sagi, Charles Vesteghem. Originally published in JMIR Medical Informatics (<ext-link ext-link-type="uri" xlink:href="https://medinform.jmir.org">https://medinform.jmir.org</ext-link>), 12.9.2025. </copyright-statement><copyright-year>2025</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://medinform.jmir.org/">https://medinform.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://medinform.jmir.org/2025/1/e68707"/><abstract><sec><title>Background</title><p>Over the last decade, natural language processing (NLP) has provided various solutions for information extraction (IE) from textual clinical data. In recent years, the use of NLP in cancer research has gained considerable attention, with numerous studies exploring the effectiveness of various NLP techniques for identifying and extracting cancer-related entities from clinical text data.</p></sec><sec><title>Objective</title><p>We aimed to summarize the performance differences between various NLP models for IE within the context of cancer to provide an overview of the relative performance of existing models.</p></sec><sec sec-type="methods"><title>Methods</title><p>This systematic literature review was conducted using 3 databases (PubMed, Scopus, and Web of Science) to search for articles extracting cancer-related entities from clinical texts. In total, 33 articles were eligible for inclusion. We extracted NLP models and their performance by <italic>F</italic><sub>1</sub>-scores. Each model was categorized into the following categories: rule-based, traditional machine learning, conditional random field-based, neural network, and bidirectional transformer (BT). The average of the performance difference for each combination of categorizations was calculated across all articles.</p></sec><sec sec-type="results"><title>Results</title><p>The articles covered various scenarios, with the best performance for each article ranging from 0.355 to 0.985 in <italic>F</italic><sub>1</sub>-score. Examining the overall relative performances, the BT category outperformed every other category (average <italic>F</italic><sub>1</sub>-score between 0.2335 and 0.0439). The percentage of articles on implementing BTs has increased over the years.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>NLP has demonstrated the ability to identify and extract cancer-related entities from unstructured textual data. Generally, more advanced models outperform less advanced ones. The BT category performed the best.</p></sec></abstract><kwd-group><kwd>natural language processing</kwd><kwd>information extraction</kwd><kwd>clinical textual data</kwd><kwd>performance</kwd><kwd>F1-score</kwd><kwd>review</kwd><kwd>rule-based solutions</kwd><kwd>traditional machine learning</kwd><kwd>neural network</kwd><kwd>bidirectional transformer</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>Electronic health records (EHRs) are increasingly being adopted by health care providers worldwide, as they offer numerous benefits [<xref ref-type="bibr" rid="ref1">1</xref>]. This has led to an increase in the quantity of data stored in EHRs, consisting of both structured and unstructured data (eg, text, images, and time series). Unstructured textual data from discharge summaries, radiology reports, clinical notes, and patient histories provide valuable information about patients that may not be captured by structured data alone [<xref ref-type="bibr" rid="ref2">2</xref>]. The extraction of clinical parameters from unstructured textual data, also known as information extraction (IE), has proven to be valuable in health care, such as in clinical research (eg, epidemiology) and decision support systems [<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref4">4</xref>].</p><p>However, working with unstructured textual data presents several challenges to health care providers and researchers. The volume of free text makes manual extraction and analysis time-consuming and resource-heavy, thereby limiting their utility and requiring automated solutions. Moreover, the lack of standardization and consistency in formatting and terminology makes it difficult to accurately identify and extract the relevant information in an automated manner. Furthermore, free text is prone to spelling errors, resulting in inaccurate or harder-to-find patient information for methods that rely on keyword extraction or other exact-match search techniques.</p><p>Natural language processing (NLP) techniques are well suited for extracting information from free text because of their ability to process, comprehend, and generate human language in a manner that allows for automatic extraction of structured information from free text. In recent years, NLP has gained considerable attention, with numerous studies exploring the effectiveness of various NLP techniques, notably for identifying and extracting cancer-related entities, such as smoking history [<xref ref-type="bibr" rid="ref2">2</xref>], toxicities [<xref ref-type="bibr" rid="ref5">5</xref>], and Gleason scores [<xref ref-type="bibr" rid="ref6">6</xref>], which are only recorded as free text in clinical notes. These techniques are known as named entity recognition, or more generally, IE [<xref ref-type="bibr" rid="ref7">7</xref>-<xref ref-type="bibr" rid="ref9">9</xref>].</p><p>A variety of techniques and pipelines have been developed for IE from medical free texts, ranging from simple rule-based solutions to advanced machine learning approaches. Rule-based solutions allow domain experts to define a set of linguistic rules and patterns to be implemented to identify and extract relevant information from clinical notes and medical free-form texts. Studies have shown that rule-based approaches can outperform machine-learning models [<xref ref-type="bibr" rid="ref10">10</xref>-<xref ref-type="bibr" rid="ref13">13</xref>]. However, rule-based approaches are custom-made for specific datasets and use cases, require manual specification of rules by medical experts, and are therefore difficult to generalize [<xref ref-type="bibr" rid="ref3">3</xref>]. Moreover, machine learning models allow for multiple methodologies and applications for different IE problems, solved by training specific traditional machine learning models such as support vector machines, decision trees, and neural networks (NNs). Recently, bidirectional transformers (BTs) such as large language models (LLM) have been identified as a possible tool for IE because of their strengths in pattern recognition, text summarization, and generation [<xref ref-type="bibr" rid="ref14">14</xref>]. LLMs allow for pretraining on large text corpora, which enables them to learn linguistic patterns applicable to different IE tasks. Furthermore, LLMs show promising results for specific tasks because of the domain-specific fine-tuning of pretrained models [<xref ref-type="bibr" rid="ref15">15</xref>].</p><p>Over the years, numerous models have been developed for various IE tasks, but their relative performance across different datasets remains largely unknown. Both rule-based and machine learning approaches often exhibit limited generalizability between tasks, and their results are frequently inconsistent when applied to different datasets. This inconsistency highlights the need for studies that investigate and compare the relative performance of these models.</p><p>To the best of our knowledge, no review has been conducted that summarizes the differences in the performance of various types of NLP models for IE within the context of cancer. This review provides an overview of the various NLP methods used for IE and compares them in terms of their performance.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><p>This systematic review followed the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines.</p><p>We searched 3 databases&#x2014;PubMed, Scopus, and Web of Science&#x2014;for relevant literature published between January 1, 2014, and April 19, 2024 (<xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>). The following search criteria were used for the titles and abstracts:</p><p>(&#x201C;information extraction&#x201D; OR &#x201C;natural language processing&#x201D; OR &#x201C;nlp&#x201D;) AND (EHR OR notes OR reports) AND (cancer OR tumor OR oncology)</p><p>The inclusion criterion for articles was the application of 2 or more NLP models to extract identical cancer-related entities from the same unstructured medical text in EHR. Articles were excluded during title and abstract screening if they were as follows: (1) reviews; (2) articles that do not compare 2 or more NLP models; (3) articles whose purpose is not to extract information from free text from EHR; and (4) articles whose purpose is not to develop an NLP cancer-related application.</p><p>For further exclusion during the full-text screening, articles were excluded if they were defined as follows: (1) abstract only; (2) text classification without cancer entity extraction; (3) results within the article were not compatible; (4) no NLP application development; (5) not related to cancer IE from EHR; and (6) no comparison with other NLP methods within the article.</p><p>Using the exclusion criteria, one author (SCD) performed 2 rounds of article selection: title and abstract screening, followed by a full-text review. A second reviewer (CV) was consulted for unclear cases during the screening.</p><p>Data from each of the included articles were extracted by 2 authors (SCD and CV). Both authors independently categorized the NLP models and extracted their performance metrics. Any discrepancies in categorization were resolved through consensus guided by consideration of the primary architectural components of the model. Each model was categorized into the following groups: rule-based, traditional machine learning (ML), conditional random field (CRF)&#x2013;based, NN, and BT.</p><p>The rule-based category includes IE models that use regular expressions (Regex), keywords, and dictionary matching. The CRF-based category includes linear CRF, except bidirectional long short-term memory-CRF, which is in the NN category. The NN category includes NNs, except for BTs, that belong to the BT category. Ensemble models are categorized as the most advanced part of the ensemble. For example, a rule-based model combined with a BT is categorized as a BT model (see <xref ref-type="table" rid="table1">Table 1</xref>). For articles that included both strict and relaxed keyword matching, the strict <italic>F</italic><sub>1</sub>-scores were extracted as the performance metric. For articles presenting both macro- and micro-averaged <italic>F</italic><sub>1</sub>-scores, macro-averaged <italic>F</italic><sub>1</sub>-scores were extracted.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Method categorization of models.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Category</td><td align="left" valign="bottom">Included models</td><td align="left" valign="bottom">Articles using category</td><td align="left" valign="bottom">Total number of models implemented</td></tr></thead><tbody><tr><td align="left" valign="top">Rule-based</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Regular expressions</p></list-item><list-item><p>keyword, term, and dictionary matching</p></list-item></list></td><td align="char" char="." valign="top">[<xref ref-type="bibr" rid="ref10">10</xref>-<xref ref-type="bibr" rid="ref13">13</xref>,<xref ref-type="bibr" rid="ref16">16</xref>-<xref ref-type="bibr" rid="ref22">22</xref>] (n=11)</td><td align="char" char="." valign="top">12</td></tr><tr><td align="left" valign="top">CRF<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup>-based</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Linear CRF</p></list-item><list-item><p>CRF + Rule-based</p></list-item></list></td><td align="char" char="." valign="top">[<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref12">12</xref>,<xref ref-type="bibr" rid="ref16">16</xref>-<xref ref-type="bibr" rid="ref18">18</xref>,<xref ref-type="bibr" rid="ref20">20</xref>,<xref ref-type="bibr" rid="ref23">23</xref>,<xref ref-type="bibr" rid="ref24">24</xref>] (n=8)</td><td align="char" char="." valign="top">26</td></tr><tr><td align="left" valign="top">Bidirectional transformer</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>BERT<sup><xref ref-type="table-fn" rid="table1fn2">b</xref></sup></p></list-item><list-item><p>BlueBERT</p></list-item><list-item><p>BioBERT</p></list-item><list-item><p>CharBERT</p></list-item><list-item><p>Character-BERT</p></list-item><list-item><p>CancerBERT</p></list-item><list-item><p>RoBERTa</p></list-item><list-item><p>MBERT (multilingual BERT)</p></list-item><list-item><p>BETO</p></list-item><list-item><p>XLM-R</p></list-item><list-item><p>ClinicalBERT</p></list-item><list-item><p>XLNet</p></list-item><list-item><p>Bidirectional Transformer + Rule-based</p></list-item><list-item><p>Bidirectional Transformer + BiLSTM<sup><xref ref-type="table-fn" rid="table1fn3">c</xref></sup>-CRF</p></list-item></list></td><td align="char" char="." valign="top">[<xref ref-type="bibr" rid="ref11">11</xref>,<xref ref-type="bibr" rid="ref15">15</xref>-<xref ref-type="bibr" rid="ref17">17</xref>,<xref ref-type="bibr" rid="ref21">21</xref>,<xref ref-type="bibr" rid="ref23">23</xref>-<xref ref-type="bibr" rid="ref33">33</xref>] (n=16)</td><td align="char" char="." valign="top">60</td></tr><tr><td align="left" valign="top">Neural network</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>BiGRU</p></list-item><list-item><p>BiRNN</p></list-item><list-item><p>CNN</p></list-item><list-item><p>LSTM</p></list-item><list-item><p>BiLSTM-CRF</p></list-item><list-item><p>RNN</p></list-item><list-item><p>MLP,</p></list-item><list-item><p>HAN,</p></list-item><list-item><p>SLA,</p></list-item><list-item><p>CNN + Rule-based</p></list-item></list></td><td align="char" char="." valign="top">[<xref ref-type="bibr" rid="ref15">15</xref>-<xref ref-type="bibr" rid="ref20">20</xref>,<xref ref-type="bibr" rid="ref24">24</xref>-<xref ref-type="bibr" rid="ref26">26</xref>,<xref ref-type="bibr" rid="ref28">28</xref>-<xref ref-type="bibr" rid="ref43">43</xref>] (n=25)</td><td align="char" char="." valign="top">83</td></tr><tr><td align="left" valign="top">Traditional machine learning</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>SVM<sup><xref ref-type="table-fn" rid="table1fn4">d</xref></sup></p></list-item><list-item><p>Random forest</p></list-item><list-item><p>Na&#x00EF;ve Bayes</p></list-item><list-item><p>Extreme Gradient Boosting</p></list-item><list-item><p>AdaBoost</p></list-item></list></td><td align="char" char="." valign="top">[<xref ref-type="bibr" rid="ref13">13</xref>,<xref ref-type="bibr" rid="ref22">22</xref>,<xref ref-type="bibr" rid="ref27">27</xref>,<xref ref-type="bibr" rid="ref30">30</xref>,<xref ref-type="bibr" rid="ref34">34</xref>-<xref ref-type="bibr" rid="ref43">43</xref>] (n=14)</td><td align="char" char="." valign="top">39</td></tr></tbody></table><table-wrap-foot><fn id="table1fn1"><p><sup>a</sup>CRF: conditional random field.</p></fn><fn id="table1fn2"><p><sup>b</sup>BERT: Bidirectional Encoder Representations from Transformers.</p></fn><fn id="table1fn3"><p><sup>c</sup>BiLSTM: bidirectional long short-term memory.</p></fn><fn id="table1fn4"><p><sup>d</sup>SVM: support vector machine.</p></fn></table-wrap-foot></table-wrap><p>To calculate the performance differences for all categories across the included articles, the following steps were executed for all categories.</p><p>The best-performing model for each category within each article was selected. The best-performing model within category <italic>c</italic> for article <italic>a</italic> is given by <inline-formula><mml:math id="ieqn1"><mml:msub><mml:mrow><mml:mi>m</mml:mi><mml:mi>a</mml:mi><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mo>,</mml:mo><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>:</p><disp-formula id="equWL1"><mml:math id="eqn1"><mml:msub><mml:mrow><mml:mi>m</mml:mi><mml:mi>a</mml:mi><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mo>,</mml:mo><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mrow><mml:mrow><mml:mi mathvariant="normal">max</mml:mi></mml:mrow><mml:mo>&#x2061;</mml:mo><mml:mrow><mml:mo>(</mml:mo></mml:mrow></mml:mrow><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mi>c</mml:mi><mml:mo>,</mml:mo><mml:mi>m</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi> </mml:mi><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mi>c</mml:mi><mml:mo>,</mml:mo><mml:mi>m</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi> </mml:mi><mml:mo>&#x2026;</mml:mo><mml:mo>,</mml:mo><mml:mi> </mml:mi><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:math></disp-formula><p>where <italic>P</italic> is the <italic>F</italic><sub>1</sub> performance score of method <italic>m</italic> within category <italic>c. n</italic> is the number of methods within category <italic>c</italic>.</p><p>Having the best-performing categories within an article allows for calculation of the category difference for each combination of categories. Category differences for categories <italic>c<sup>1</sup></italic> and <italic>c<sup>2</sup></italic> within article <italic>a</italic> are given as <inline-formula><mml:math id="ieqn2"><mml:msub><mml:mrow><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>g</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>y</mml:mi><mml:mo>_</mml:mo><mml:mi>d</mml:mi><mml:mi>i</mml:mi><mml:mi>f</mml:mi><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mi>c</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:mi>c</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>:</p><disp-formula id="equWL2"><mml:math id="eqn2"><mml:msub><mml:mrow><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>g</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>y</mml:mi><mml:mo>_</mml:mo><mml:mi>d</mml:mi><mml:mi>i</mml:mi><mml:mi>f</mml:mi><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mi>c</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:mi>c</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>m</mml:mi><mml:mi>a</mml:mi><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mi>c</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>m</mml:mi><mml:mi>a</mml:mi><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mi>c</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:math></disp-formula><p>where  <inline-formula><mml:math id="ieqn3"><mml:msub><mml:mrow><mml:mi>m</mml:mi><mml:mi>a</mml:mi><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mo>,</mml:mo><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is the best performing model of category <italic>c</italic> in article <italic>a</italic>.</p><p>All performance differences of the same combination of categories were averaged across all the articles. The average of the category difference for all articles with combination <italic>c<sup>1</sup></italic> and <italic>c<sup>2</sup></italic> is given as <inline-formula><mml:math id="ieqn4"><mml:msub><mml:mrow><mml:mi>p</mml:mi><mml:mi>e</mml:mi><mml:mi>r</mml:mi><mml:mi>f</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>m</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi><mml:mi>c</mml:mi><mml:mi>e</mml:mi><mml:mo>_</mml:mo><mml:mi>d</mml:mi><mml:mi>i</mml:mi><mml:mi>f</mml:mi><mml:mi>f</mml:mi><mml:mi>e</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi><mml:mi>c</mml:mi><mml:mi>e</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula> :</p><disp-formula id="equWL3"><mml:math id="eqn3"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>p</mml:mi><mml:mi>e</mml:mi><mml:mi>r</mml:mi><mml:mi>f</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>m</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi><mml:mi>c</mml:mi><mml:mi>e</mml:mi><mml:mi mathvariant="normal">_</mml:mi><mml:mi>d</mml:mi><mml:mi>i</mml:mi><mml:mi>f</mml:mi><mml:mi>f</mml:mi><mml:mi>e</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi><mml:mi>c</mml:mi><mml:msub><mml:mi>e</mml:mi><mml:mrow><mml:mi>c</mml:mi><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>g</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>y</mml:mi><mml:mi mathvariant="normal">_</mml:mi><mml:mi>d</mml:mi><mml:mi>i</mml:mi><mml:mi>f</mml:mi><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msup><mml:mi>c</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mi>c</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mi>a</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>g</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>y</mml:mi><mml:mi mathvariant="normal">_</mml:mi><mml:mi>d</mml:mi><mml:mi>i</mml:mi><mml:mi>f</mml:mi><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msup><mml:mi>c</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mi>c</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mi>a</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>+</mml:mo><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>g</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>y</mml:mi><mml:mi mathvariant="normal">_</mml:mi><mml:mi>d</mml:mi><mml:mi>i</mml:mi><mml:mi>f</mml:mi><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msup><mml:mi>c</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mi>c</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mi>a</mml:mi><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:msub></mml:mrow><mml:mi>n</mml:mi></mml:mfrac></mml:mrow></mml:mstyle></mml:math></disp-formula><p>where  <inline-formula><mml:math id="ieqn5"><mml:msub><mml:mrow><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>g</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>y</mml:mi><mml:mo>_</mml:mo><mml:mi>d</mml:mi><mml:mi>i</mml:mi><mml:mi>f</mml:mi><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mi>c</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:mi>c</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is the category difference between <italic>c<sup>1</sup></italic> and <italic>c<sup>2</sup></italic> in article <italic>a. n</italic> is the number of articles with a specific category combination.</p><p>Statistical significance between categories for each category combination was determined using a <italic>t</italic> test (<italic>P</italic>&#x003C;.05).</p></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Overview</title><p>The article selection process is detailed in a PRISMA flowchart, shown in <xref ref-type="fig" rid="figure1">Figure 1</xref>. A total of 2032 articles were identified through searches in Web of Science, Scopus, and PubMed.</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>The PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) flowchart for article selection.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v13i1e68707_fig01.png"/></fig><p>In total, 33 articles were included in this review. The articles were published between 2018 and 2024. They compared at least 2 NLP models for cancer-related IE from unstructured medical texts in EHRs. The articles contained a total of 220 implementations of NLP models. Selecting only the best-performing models within each category of each article summarizes 74 implementations.</p></sec><sec id="s3-2"><title>Models</title><p>We categorized each NLP model as rule-based, CRF-based, BT, NN, and ML. <xref ref-type="table" rid="table1">Table 1</xref> shows how each model was categorized and the articles in which the categories are contained.</p><p><xref ref-type="table" rid="table1">Table 1</xref> shows the categorization of the models, which articles contain the specific categories, and the total number of implemented models within each category.</p><p>The most frequently used category was NN, with 25 occurrences, followed by BT and ML with 16 and 11 occurrences, respectively. The most frequently implemented category was NN, with 83 implementations. The distribution of unique categorizations per year shows the variety of models that have been used throughout the years (see <xref ref-type="fig" rid="figure2">Figure 2</xref>). Notably, the percentage of articles on the implementation of BTs has increased over the years.</p><fig position="float" id="figure2"><label>Figure 2.</label><caption><p>Distribution of unique categorizations per article per year. CRF: conditional random field.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v13i1e68707_fig02.png"/></fig></sec><sec id="s3-3"><title>Performance</title><p>The performance varied a lot according to specific use cases. Inspecting the best-performing models of the specific articles shows that a total of 5 rule-based models performed best in their articles, with <italic>F</italic><sub>1</sub>-scores in the range of 0.73-0.887 (see <xref ref-type="table" rid="table2">Table 2</xref>). ML did not perform best in any article despite being compared in 14 articles with a total of 39 different model implementations, and neither did CRF-based. In total, 14 articles showed that NN performed the best, with <italic>F</italic><sub>1</sub>-scores ranging from 0.3539 to 0.972. BT performed the best in 13 articles, with <italic>F</italic><sub>1</sub>-scores ranging from 0.6023 to 0.97. Looking at the raw <italic>F</italic><sub>1</sub>-scores, more advanced models outperformed less advanced ones.</p><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>Article overview.</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Article</td><td align="left" valign="bottom">Year</td><td align="left" valign="bottom">Title</td><td align="left" valign="bottom">Number of tested models (best performing <italic>F</italic><sub>1</sub>-score within category)</td></tr></thead><tbody><tr><td align="left" valign="top">AAlAbdulsalam et al [<xref ref-type="bibr" rid="ref10">10</xref>]</td><td align="char" char="." valign="top">2018</td><td align="left" valign="top">Automated extraction and classification of cancer stage mentions from unstructured text fields in a central cancer registry</td><td align="left" valign="top">Rule-based=1 (0.887)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup>;<break/>CRF<sup><xref ref-type="table-fn" rid="table2fn2">b</xref></sup>-based=1 (0.882)</td></tr><tr><td align="left" valign="top">Alawad et al [<xref ref-type="bibr" rid="ref36">36</xref>]</td><td align="char" char="." valign="top">2018</td><td align="left" valign="top">Coarse-to-fine multi-task training of convolutional neural networks for automated information extraction from cancer pathology reports</td><td align="left" valign="top">ML<sup><xref ref-type="table-fn" rid="table2fn3">c</xref></sup>=1 (0.626);<break/>NN<sup><xref ref-type="table-fn" rid="table2fn4">d</xref></sup>=2 (0.752)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">Miao et al [<xref ref-type="bibr" rid="ref20">20</xref>]</td><td align="char" char="." valign="top">2018</td><td align="left" valign="top">Extraction of BI-RADS findings from breast ultrasound reports in Chinese using deep learning approaches</td><td align="left" valign="top">Rule-based=1 (0.848);<break/>CRF-based=1 (0.881);<break/>NN=2 (0.904)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">Qiu et al [<xref ref-type="bibr" rid="ref35">35</xref>]</td><td align="char" char="." valign="top">2018</td><td align="left" valign="top">Deep learning for automated extraction of primary sites from cancer pathology reports</td><td align="left" valign="top">ML=3 (0.640);<break/>NN=3 (0.701)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">Chen et al [<xref ref-type="bibr" rid="ref12">12</xref>]</td><td align="char" char="." valign="top">2019</td><td align="left" valign="top">Using natural language processing to extract clinically useful information from Chinese electronic medical records</td><td align="left" valign="top">Rule-based=1 (0.83)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup>;<break/>CRF=1 (0.8)</td></tr><tr><td align="left" valign="top">Coquet et al [<xref ref-type="bibr" rid="ref19">19</xref>]</td><td align="char" char="." valign="top">2019</td><td align="left" valign="top">Comparison of orthogonal NLP methods for clinical phenotyping and assessment of bone scan utilization among prostate cancer patients</td><td align="left" valign="top">Rule-based=1 (0.897);<break/>NN=3 (0.918)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">Dubey et al [<xref ref-type="bibr" rid="ref37">37</xref>]</td><td align="char" char="." valign="top">2019</td><td align="left" valign="top">Inverse regression for extraction of tumor site from cancer pathology reports</td><td align="left" valign="top">ML=5 (0.759)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup>;<break/>NN=2 (0.701)</td></tr><tr><td align="left" valign="top">Kim et al [<xref ref-type="bibr" rid="ref18">18</xref>]</td><td align="char" char="." valign="top">2019</td><td align="left" valign="top">A study of medical problem extraction for better disease management</td><td align="left" valign="top">Rule-based=2 (0.883);<break/>CRF-based=4 (0.926);<break/>NN=5 (0.929)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">Thompson et al [<xref ref-type="bibr" rid="ref34">34</xref>]</td><td align="char" char="." valign="top">2019</td><td align="left" valign="top">Relevant word order vectorization for improved natural language processing in electronic health records</td><td align="left" valign="top">ML=7 (0.788);<break/>NN=7 (0.858)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">Zhang et al [<xref ref-type="bibr" rid="ref17">17</xref>]</td><td align="char" char="." valign="top">2019</td><td align="left" valign="top">Extracting comprehensive clinical information for breast cancer using deep learning methods</td><td align="left" valign="top">Rule-based=1 (0.484);<break/>NN=1 (0.887);<break/>CRF-based=1 (0.837);<break/>BT<sup><xref ref-type="table-fn" rid="table2fn5">e</xref></sup>=1 (0.935)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">Alawad et al [<xref ref-type="bibr" rid="ref40">40</xref>]</td><td align="char" char="." valign="top">2020</td><td align="left" valign="top">Automatic extraction of cancer registry reportable information from free-text pathology reports using multitask convolutional neural networks</td><td align="left" valign="top">ML=2 (0.615);<break/>NN=3 (0.752)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">Odisho et al [<xref ref-type="bibr" rid="ref28">28</xref>]</td><td align="char" char="." valign="top">2020</td><td align="left" valign="top">Natural language processing systems for pathology parsing in limited data environments with uncertainty estimation</td><td align="left" valign="top">ML=4 (0.948);<break/>NN=2 (0.972)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">Osborne et al [<xref ref-type="bibr" rid="ref11">11</xref>]</td><td align="char" char="." valign="top">2020</td><td align="left" valign="top">Identification of cancer entities in clinical text combining transformers with dictionary features</td><td align="left" valign="top">Rule-based=1 (0.73)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup>;<break/>BT=7 (0.7)</td></tr><tr><td align="left" valign="top">Wu et al [<xref ref-type="bibr" rid="ref38">38</xref>]</td><td align="char" char="." valign="top">2020</td><td align="left" valign="top">Structured information extraction of pathology reports with attention-based graph convolutional network</td><td align="left" valign="top">ML=1 (0.74);<break/>NN=6 (0.803)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">Hu et al [<xref ref-type="bibr" rid="ref28">28</xref>]</td><td align="char" char="." valign="top">2021</td><td align="left" valign="top">Automatic extraction of lung cancer staging information from computed tomography reports: deep learning approach</td><td align="left" valign="top">NN=2 (0.773);<break/>BT=1 (0.81)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">Liu et al [<xref ref-type="bibr" rid="ref24">24</xref>]</td><td align="char" char="." valign="top">2021</td><td align="left" valign="top">Use of BERT (Bidirectional Encoder Representations from Transformers)-based deep learning method for extracting evidences in Chinese radiology reports: development of a computer-aided liver cancer diagnosis framework</td><td align="left" valign="top">CRF-based=1 (0.729);<break/>NN=1 (0.832);<break/>BT=1 (0.857)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">L&#x00F3;pez-Garc&#x00ED;a et al [<xref ref-type="bibr" rid="ref23">23</xref>]</td><td align="char" char="." valign="top">2021</td><td align="left" valign="top">Detection of tumor morphology mentions in clinical reports in Spanish using transformers</td><td align="left" valign="top">CRF-based=1 (0.794);<break/>BT=18 (0.89)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">Lu et al [<xref ref-type="bibr" rid="ref27">27</xref>]</td><td align="char" char="." valign="top">2021</td><td align="left" valign="top">Natural language processing and machine learning methods to characterize unstructured patient-reported outcomes: validation study</td><td align="left" valign="top">ML=2 (0.365);<break/>BT=1 (0.602)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">Park et al [<xref ref-type="bibr" rid="ref43">43</xref>]</td><td align="char" char="." valign="top">2021</td><td align="left" valign="top">Improving natural language information extraction from cancer pathology reports using transfer learning and zero-shot string similarity</td><td align="left" valign="top">ML=4 (0.484);<break/>NN=5 (0.502)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">Rios et al [<xref ref-type="bibr" rid="ref42">42</xref>]</td><td align="char" char="." valign="top">2021</td><td align="left" valign="top">Assigning ICD-O-3 codes to pathology reports using neural multi-task training with hierarchical regularization</td><td align="left" valign="top">ML=3 (0.276);<break/>NN=12 (0.355)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">Wu et al [<xref ref-type="bibr" rid="ref41">41</xref>]</td><td align="char" char="." valign="top">2021</td><td align="left" valign="top">BioIE: biomedical information extraction with multi-head attention enhanced graph convolutional network</td><td align="left" valign="top">ML=1 (0.444);<break/>NN=5 (0.613)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">Yu et al [<xref ref-type="bibr" rid="ref26">26</xref>]</td><td align="char" char="." valign="top">2021</td><td align="left" valign="top">A study of social and behavioral determinants of health in lung cancer patients using transformers-based natural language processing models</td><td align="left" valign="top">BT=4 (0.879)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup>;<break/>NN=2 (0.844)</td></tr><tr><td align="left" valign="top">Bozkurt et al [<xref ref-type="bibr" rid="ref13">13</xref>]</td><td align="char" char="." valign="top">2022</td><td align="left" valign="top">Expanding the secondary use of prostate cancer real world data: automated classifiers for clinical and pathological stage</td><td align="left" valign="top">Rule-based=1 (0.87)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup>;<break/>ML=1 (0.723)</td></tr><tr><td align="left" valign="top">Fang et al [<xref ref-type="bibr" rid="ref16">16</xref>]</td><td align="char" char="." valign="top">2022</td><td align="left" valign="top">Extracting clinical named entity for pituitary adenomas from Chinese electronic medical records</td><td align="left" valign="top">Rule-based=1 (0.431);<break/>CRF-based=16 (0.904);<break/>NN=1 (0.899);<break/>BT=1 (0.913)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">Hu et al [<xref ref-type="bibr" rid="ref29">29</xref>]</td><td align="char" char="." valign="top">2022</td><td align="left" valign="top">Using natural language processing and machine learning to preoperatively predict lymph node metastasis for non-small cell lung cancer with electronic medical records: development and validation study</td><td align="left" valign="top">NN=1 (0.701);<break/>BT=2 (0.948)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">Pab&#x00F3;n et al [<xref ref-type="bibr" rid="ref25">25</xref>]</td><td align="char" char="." valign="top">2022</td><td align="left" valign="top">Negation and uncertainty detection in clinical texts written in Spanish: a deep learning-based approach</td><td align="left" valign="top">NN=2 (0.788);<break/>BT=1 (0.823)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">Zhou et al [<xref ref-type="bibr" rid="ref15">15</xref>]</td><td align="char" char="." valign="top">2022</td><td align="left" valign="top">CancerBERT: a cancer domain-specific language model for extracting breast cancer phenotypes from electronic health records</td><td align="left" valign="top">NN=1 (0.834);<break/>BT=8 (0.876)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">Ansoborlo et al [<xref ref-type="bibr" rid="ref22">22</xref>]</td><td align="char" char="." valign="top">2023</td><td align="left" valign="top">Prescreening in oncology trials using medical records. Natural language processing applied on lung cancer multidisciplinary team meeting reports</td><td align="left" valign="top">Rule-based=1 (0.932)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup>;<break/>ML=1 (0.68)</td></tr><tr><td align="left" valign="top">Rohanian et al [<xref ref-type="bibr" rid="ref32">32</xref>]</td><td align="char" char="." valign="top">2023</td><td align="left" valign="top">Using bottleneck adapters to identify cancer in clinical notes under low-resource constraints</td><td align="left" valign="top">NN=3 (0.83);<break/>BT=8 (0.97)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">Seong et al [<xref ref-type="bibr" rid="ref31">31</xref>]</td><td align="char" char="." valign="top">2023</td><td align="left" valign="top">Deep learning approach to detection of colonoscopic information from unstructured reports</td><td align="left" valign="top">NN=3 (0.985)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup>;<break/>BT=2 (0.982)</td></tr><tr><td align="left" valign="top">Zitu et al [<xref ref-type="bibr" rid="ref30">30</xref>]</td><td align="char" char="." valign="top">2023</td><td align="left" valign="top">Generalizability of machine learning methods in detecting adverse drug events from clinical narratives in electronic medical records</td><td align="left" valign="top">ML=1 (0.69);<break/>NN=2 (0.763);<break/>BT=2 (0.778)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">Mart&#x00ED;n-Noguerol et al [<xref ref-type="bibr" rid="ref33">33</xref>]</td><td align="char" char="." valign="top">2024</td><td align="left" valign="top">Natural language processing deep learning models for the differential between high-grade gliomas and metastasis: what if the key is how we report them?</td><td align="left" valign="top">NN=3 (0.872)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup>;<break/>BT=1 (0.766)</td></tr><tr><td align="left" valign="top">Hu et al [<xref ref-type="bibr" rid="ref21">21</xref>]</td><td align="char" char="." valign="top">2024</td><td align="left" valign="top">Zero-shot information extraction from radiological reports using ChatGPT</td><td align="left" valign="top">Rule-based=1 (0.926);<break/>BT=2 (0.957)<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td></tr></tbody></table><table-wrap-foot><fn id="table2fn1"><p><sup>a</sup>This is the best <italic>F</italic><sub>1</sub>-score for the article.</p></fn><fn id="table2fn2"><p><sup>b</sup>CRF: conditional random field.</p></fn><fn id="table2fn3"><p><sup>c</sup>ML: traditional machine learning.</p></fn><fn id="table2fn4"><p><sup>d</sup>NN: neural network.</p></fn><fn id="table2fn5"><p><sup>e</sup>BT: bidirectional transformer. </p></fn></table-wrap-foot></table-wrap><p><xref ref-type="table" rid="table2">Table 2</xref> shows each article and the number of models in each category within the article. Parentheses show the best <italic>F</italic><sub>1</sub>-score for each category. The best <italic>F</italic><sub>1</sub>-score for each article is marked by a footnote.</p><p>Some variations between the average <italic>F</italic><sub>1</sub>-score performance differences were observed (see <xref ref-type="fig" rid="figure3">Figure 3</xref>).</p><p>Our results show that more advanced models outperform less advanced ones. The largest difference between the category performance <italic>F</italic><sub>1</sub>-scores was observed between the BT category and the rule-based category. BT models were compared with rule-based models in 4 studies, yielding an average performance difference of 0.2335 in terms of <italic>F</italic><sub>1</sub>-score. BT was the best-performing category. NN outperformed CRF-based, ML, and rule-based models, while CRF-based outperformed rule-based models, and rule-based outperformed ML models. The only statistically significant difference between categories is observed when comparing rule-based and ML; see <xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref> for <italic>P</italic> values.</p><fig position="float" id="figure3"><label>Figure 3.</label><caption><p>Illustration of the average <italic>F</italic><sub>1</sub>-score performance differences for all combinations of the best model within each category. The parentheses indicate the number of comparisons between the categories. CRF: conditional random field. *<italic>P</italic>&#x003C;.05.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v13i1e68707_fig03.png"/></fig></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Principal Findings</title><p>This study provides an overview of the models used for IE in cancer and their performance in terms of the <italic>F</italic><sub>1</sub>-score. By including only articles with 2 or more NLP models for IE, we were able to evaluate the relative performance of each NLP within categories: rule-based, CRF-based, BT, NN, and ML.</p><p>The search string for this review combined keywords for techniques (IE and NLP), data sources (EHR, notes, reports), and the domain (cancer, tumor, and oncology) using Boolean operators to limit irrelevant results. The initial yield of 2032 articles suggests a reasonable balance, considering the stringent inclusion criteria. The &#x201C;AND&#x201D; clauses effectively limit the search while still including the relevant articles for the screening process. Although our search strategy included articles published from 01/01/2014, no articles prior to 2018 were included in the analysis. The reason for this discrepancy is not addressed within the scope of this review, which focused on quantifying performance differences between our categories. Notably, the most frequent reason for full-text exclusion was &#x201C;No comparison with other NLP methods within the article (185 articles).&#x201D; Arguing for common benchmark testing of the implemented NLP models.</p><p>Without considering a dataset or specific extraction entities, our results show that BT is the best performing category, followed by NN, CRF-based, rule-based, and ML in written order. We observed an increasing number of transformer-based models developed in recent years, with promising results. Our results highlight a pivotal moment in which BTs, such as language models, are on the verge of demonstrating their full potential in IE. Although transformers [<xref ref-type="bibr" rid="ref44">44</xref>] and BERT [<xref ref-type="bibr" rid="ref45">45</xref>] were introduced in 2017 and 2018, respectively, our literature review includes no articles using these technologies until 2019. This delay in time may reflect the time required for these models to become integrated into clinical research workflows. Surprisingly, rule-based solutions perform better than machine learning [<xref ref-type="bibr" rid="ref13">13</xref>,<xref ref-type="bibr" rid="ref22">22</xref>]. One explanation could be that rule-based solutions allow for the implementation of expert knowledge. The lowest-performing articles in terms of <italic>F</italic><sub>1</sub>-score do not aim to show the best possible method for extraction, but rather how <italic>F</italic><sub>1</sub>-scores increase using hierarchical regularization when extracting ICD-O-3 codes [<xref ref-type="bibr" rid="ref42">42</xref>]. Similarly, the study of Park et al [<xref ref-type="bibr" rid="ref43">43</xref>] aims to show how to increase the <italic>F</italic><sub>1</sub>-score, using transfer learning and zero-shot string similarity, when the number of annotated pathology reports is limited.</p><p>Multiple reviews have been conducted within the scope of NLP in a clinical context with different aims. The review by Kreimeyer et al [<xref ref-type="bibr" rid="ref46">46</xref>] aims to identify NLP systems capable of processing clinical free text and generating structured output, thereby compiling a list of NLP solutions in use. The review by Datta et al [<xref ref-type="bibr" rid="ref47">47</xref>] defines relevant linguistic terms by organizing unstructured clinical text related to cancer into structured data using frame semantics. The review by Bilal et al [<xref ref-type="bibr" rid="ref48">48</xref>] examines the current state-of-the-art literature on NLP applications in analyzing EHRs and clinical notes for cancer research, quantifying the number of studies for each cancer type and outlining the research challenges and future directions for NLP when analyzing EHRs and clinical notes in cancer research. However, no review has been conducted comparing the performance of NLP models for IE of cancer-related entities from clinical text, a gap relevant to clinical informatics and crucial for improving the accuracy of cancer-related data IE within EHRs. This is the first review to summarize and compare the performance of NLP models for IE of different cancer entities from unstructured text, offering insights for clinical researchers focused on leveraging EHR data for cancer care and research.</p></sec><sec id="s4-2"><title>Strengths</title><p>One strength of our study was its ability to overcome the challenge of comparing low-performing models. By including only articles with 2 or more categories, we can determine the relative performance for each paper while neglecting low-performing models from papers that do not aim to beat state-of-the-art <italic>F</italic><sub>1</sub>-score. Our review shows how models can be categorized and how the categorizations perform compared to each other through different datasets and extraction entities. The performance differences observed in our included articles highlight the importance of selecting the appropriate NLP model for each health care application. Our categorizations allow all models to be included, even ensemble and hybrid models. Furthermore, our performance calculation uses the best-performing model for each category reported within each included article. This approach allows for the addition of multiple new categories to support the desired level of model performance granularity.</p></sec><sec id="s4-3"><title>Limitations</title><p>A categorization strategy was required to categorize all models. Most models assign into well-defined and distinct categories. However, some could be assigned to multiple categories, notably bidirectional long short-term memory-CRF models. To present intelligible results, the number of categories had to be kept relatively low, neglecting model specificities. Increasing the number of categories would reduce the number of models in that category, making the results too anecdotal. Decreasing the number of categories would increase the number of times each categorization was compared, making the averaged <italic>F</italic><sub>1</sub>-scores less distinctive. Ideally, we would have wished for multiple studies implementing the same set of models and categorizations to avoid certain categorizations not being compared with every other category and to avoid certain combinations of categorizations occurring only once.</p><p>We selected the <italic>F</italic><sub>1</sub>-score as a metric for performance; precision or recall could also be used. However, extracting specific numbers from the confusion matrix can provide deeper insights. The included studies reported <italic>F</italic><sub>1</sub>-scores as a measure of performance. Although this is a practical method to generate 1 performance metric, its use has some limitations. In medical IE, one could argue that false negatives are worse than false positives, potentially leading to missed diagnoses or inappropriate treatment decisions, which is not considered in the <italic>F</italic><sub>1</sub>-score. While metrics such as AUC-ROC, precision-recall tradeoff, or specificity offer complementary insights, their calculation was limited by the inconsistent reporting of the necessary data. Furthermore, given the sensitive nature of EHR data and the need for clinical trust, future research should also prioritize evaluating the interpretability of IE models alongside traditional performance measures to allow clinicians to understand how cancer-related entities are being extracted and validated from EHR data.</p><p>Furthermore, our included studies neglected to address the handling of negation and spelling errors. Giorgia et al [<xref ref-type="bibr" rid="ref49">49</xref>] showed that negations account for 66% of the errors. Another study stated that BERT fails completely to show a generalizable understanding of negation, raising questions about the aptitude of language models to learn this type of meaning [<xref ref-type="bibr" rid="ref50">50</xref>]. In this study, BTs performed well; one could wish for a general approach to analyze the errors of each model instead of the general performance derived from the confusion matrix. Negation errors pose a significant challenge in EHR data and are critical in oncology, as a misidentified negated symptom or finding can alter clinical interpretation, treatment planning, and patient care.</p></sec><sec id="s4-4"><title>Perspectives</title><p>The field of IE has evolved rapidly, and models, such as LLMs, have been successfully applied in the context of cancer IE, both in terms of model performance and operational efficiency [<xref ref-type="bibr" rid="ref51">51</xref>]. LLM could allow for enhanced transferability and utility for different IE tasks on unstructured textual data. Using LLMs for IE on unstructured textual data seems feasible because of the variety of available pretrained models in different versions. Some might perform well out of the box or with minor domain-specific fine-tuning [<xref ref-type="bibr" rid="ref15">15</xref>]. Generally, the evaluation of LLMs is challenging because of the lack of clarity regarding whether a public benchmark dataset has been used for training. However, when using data from EHRs, it is certain that they have not been used for training a public model.</p></sec><sec id="s4-5"><title>Conclusions</title><p>NLP has demonstrated the ability to identify and extract cancer-related entities from unstructured medical textual data. Generally, most of the reviewed models showed excellent performance in terms of the <italic>F</italic><sub>1</sub>-score, and more advanced models outperformed less advanced ones. The BT category performed the best, followed by NN. The use of BTs has increased in recent years. Rule-based applications for IE remain competitive in terms of performance in this specific context.</p></sec></sec></body><back><ack><p>This study was supported solely by the institutional resources of the Center for Clinical Data Science at Aalborg University Hospital. We acknowledge the use of Covidence systematic review software (Veritas Health Innovation, Melbourne, Australia) for automatic identification and removal of duplicate articles.</p></ack><notes><sec><title>Data Availability</title><p>The datasets used and analyzed in this study are available upon reasonable request from the corresponding author.</p></sec></notes><fn-group><fn fn-type="con"><p>All authors were involved in the conception and design of the study. SCD contributed to the collection and assembly of data. SCD and CV were involved in data analysis and interpretation. All authors contributed to manuscript writing, and all authors approved the final version of the manuscript.</p></fn><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">BERT</term><def><p>Bidirectional Encoder Representations from Transformers</p></def></def-item><def-item><term id="abb2">BT</term><def><p>bidirectional transformer</p></def></def-item><def-item><term id="abb3">CRF</term><def><p>conditional random field</p></def></def-item><def-item><term id="abb4">EHR</term><def><p>electronic health record</p></def></def-item><def-item><term id="abb5">IE</term><def><p>information extraction</p></def></def-item><def-item><term id="abb6">LLM</term><def><p>large language model</p></def></def-item><def-item><term id="abb7">ML</term><def><p>machine learning</p></def></def-item><def-item><term id="abb8">NLP</term><def><p>natural language processing</p></def></def-item><def-item><term id="abb9">NN</term><def><p>neural network</p></def></def-item><def-item><term id="abb10">PRISMA</term><def><p>Preferred Reporting Items for Systematic Reviews and Meta-Analyses</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Evans</surname><given-names>RS</given-names> </name></person-group><article-title>Electronic health records: then, now, and in the future</article-title><source>Yearb Med Inform</source><year>2016</year><month>05</month><day>20</day><volume>Suppl 1</volume><issue>Suppl 1</issue><fpage>S48</fpage><lpage>61</lpage><pub-id pub-id-type="doi">10.15265/IYS-2016-s006</pub-id><pub-id pub-id-type="medline">27199197</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ruckdeschel</surname><given-names>JC</given-names> </name><name name-style="western"><surname>Riley</surname><given-names>M</given-names> </name><name name-style="western"><surname>Parsatharathy</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Unstructured data are superior to structured data for eliciting quantitative smoking history from the electronic health record</article-title><source>JCO Clin Cancer Inform</source><year>2023</year><month>02</month><volume>7</volume><fpage>e2200155</fpage><pub-id pub-id-type="doi">10.1200/CCI.22.00155</pub-id><pub-id pub-id-type="medline">36809022</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Landolsi</surname><given-names>MY</given-names> </name><name name-style="western"><surname>Hlaoua</surname><given-names>L</given-names> </name><name name-style="western"><surname>Ben Romdhane</surname><given-names>L</given-names> </name></person-group><article-title>Information extraction from electronic medical documents: state of the art and future research directions</article-title><source>Knowl Inf Syst</source><year>2023</year><volume>65</volume><issue>2</issue><fpage>463</fpage><lpage>516</lpage><pub-id pub-id-type="doi">10.1007/s10115-022-01779-1</pub-id><pub-id pub-id-type="medline">36405956</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Spasi&#x0107;</surname><given-names>I</given-names> </name><name name-style="western"><surname>Livsey</surname><given-names>J</given-names> </name><name name-style="western"><surname>Keane</surname><given-names>JA</given-names> </name><name name-style="western"><surname>Nenadi&#x0107;</surname><given-names>G</given-names> </name></person-group><article-title>Text mining of cancer-related information: review of current status and future directions</article-title><source>Int J Med Inform</source><year>2014</year><month>09</month><volume>83</volume><issue>9</issue><fpage>605</fpage><lpage>623</lpage><pub-id pub-id-type="doi">10.1016/j.ijmedinf.2014.06.009</pub-id><pub-id pub-id-type="medline">25008281</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hong</surname><given-names>JC</given-names> </name><name name-style="western"><surname>Fairchild</surname><given-names>AT</given-names> </name><name name-style="western"><surname>Tanksley</surname><given-names>JP</given-names> </name><name name-style="western"><surname>Palta</surname><given-names>M</given-names> </name><name name-style="western"><surname>Tenenbaum</surname><given-names>JD</given-names> </name></person-group><article-title>Natural language processing for abstraction of cancer treatment toxicities: accuracy versus human experts</article-title><source>JAMIA Open</source><year>2021</year><month>02</month><day>15</day><volume>3</volume><issue>4</issue><fpage>513</fpage><lpage>517</lpage><pub-id pub-id-type="doi">10.1093/jamiaopen/ooaa064</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yu</surname><given-names>S</given-names> </name><name name-style="western"><surname>Le</surname><given-names>A</given-names> </name><name name-style="western"><surname>Feld</surname><given-names>E</given-names> </name><etal/></person-group><article-title>A natural language processing-assisted extraction system for Gleason scores: development and usability study</article-title><source>JMIR Cancer</source><year>2021</year><month>07</month><day>2</day><volume>7</volume><issue>3</issue><fpage>e27970</fpage><pub-id pub-id-type="doi">10.2196/27970</pub-id><pub-id pub-id-type="medline">34255641</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Alkaitis</surname><given-names>MS</given-names> </name><name name-style="western"><surname>Agrawal</surname><given-names>MN</given-names> </name><name name-style="western"><surname>Riely</surname><given-names>GJ</given-names> </name><name name-style="western"><surname>Razavi</surname><given-names>P</given-names> </name><name name-style="western"><surname>Sontag</surname><given-names>D</given-names> </name></person-group><article-title>Automated NLP extraction of clinical rationale for treatment discontinuation in breast cancer</article-title><source>JCO Clin Cancer Inform</source><year>2021</year><month>05</month><volume>5</volume><issue>5</issue><fpage>550</fpage><lpage>560</lpage><pub-id pub-id-type="doi">10.1200/CCI.20.00139</pub-id><pub-id pub-id-type="medline">33989016</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Benson</surname><given-names>R</given-names> </name><name name-style="western"><surname>Winterton</surname><given-names>C</given-names> </name><name name-style="western"><surname>Winn</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Leveraging natural language processing to extract features of colorectal polyps from pathology reports for epidemiologic study</article-title><source>JCO Clin Cancer Inform</source><year>2023</year><month>01</month><volume>7</volume><fpage>e2200131</fpage><pub-id pub-id-type="doi">10.1200/CCI.22.00131</pub-id><pub-id pub-id-type="medline">36753686</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Si</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Roberts</surname><given-names>K</given-names> </name></person-group><article-title>A frame-based NLP system for cancer-related information extraction</article-title><source>AMIA Annu Symp Proc</source><year>2018</year><volume>2018</volume><fpage>1524</fpage><lpage>1533</lpage><pub-id pub-id-type="medline">30815198</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>AAlAbdulsalam</surname><given-names>AK</given-names> </name><name name-style="western"><surname>Garvin</surname><given-names>JH</given-names> </name><name name-style="western"><surname>Redd</surname><given-names>A</given-names> </name><name name-style="western"><surname>Carter</surname><given-names>ME</given-names> </name><name name-style="western"><surname>Sweeny</surname><given-names>C</given-names> </name><name name-style="western"><surname>Meystre</surname><given-names>SM</given-names> </name></person-group><article-title>Automated extraction and classification of cancer stage mentions from unstructured text fields in a central cancer registry</article-title><source>AMIA Jt Summits Transl Sci Proc</source><year>2018</year><volume>2017</volume><fpage>16</fpage><lpage>25</lpage><pub-id pub-id-type="medline">29888032</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Osborne</surname><given-names>JD</given-names> </name><name name-style="western"><surname>O&#x2019;Leary</surname><given-names>T</given-names> </name><name name-style="western"><surname>Monte</surname><given-names>JD</given-names> </name><name name-style="western"><surname>Sasse</surname><given-names>K</given-names> </name><name name-style="western"><surname>Liang</surname><given-names>WH</given-names> </name></person-group><article-title>Identification of cancer entities in clinical text combining transformers with dictionary features</article-title><source>CEUR Workshop Proceedings</source><year>2020</year><volume>2664</volume><fpage>458</fpage><lpage>467</lpage></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chen</surname><given-names>L</given-names> </name><name name-style="western"><surname>Song</surname><given-names>L</given-names> </name><name name-style="western"><surname>Shao</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Li</surname><given-names>D</given-names> </name><name name-style="western"><surname>Ding</surname><given-names>K</given-names> </name></person-group><article-title>Using natural language processing to extract clinically useful information from Chinese electronic medical records</article-title><source>Int J Med Inform</source><year>2019</year><month>04</month><volume>124</volume><fpage>6</fpage><lpage>12</lpage><pub-id pub-id-type="doi">10.1016/j.ijmedinf.2019.01.004</pub-id><pub-id pub-id-type="medline">30784428</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bozkurt</surname><given-names>S</given-names> </name><name name-style="western"><surname>Magnani</surname><given-names>CJ</given-names> </name><name name-style="western"><surname>Seneviratne</surname><given-names>MG</given-names> </name><name name-style="western"><surname>Brooks</surname><given-names>JD</given-names> </name><name name-style="western"><surname>Hernandez-Boussard</surname><given-names>T</given-names> </name></person-group><article-title>Expanding the secondary use of prostate cancer real world data: automated classifiers for clinical and pathological stage</article-title><source>Front Digit Health</source><year>2022</year><volume>4</volume><fpage>793316</fpage><pub-id pub-id-type="doi">10.3389/fdgth.2022.793316</pub-id><pub-id pub-id-type="medline">35721793</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Iannantuono</surname><given-names>GM</given-names> </name><name name-style="western"><surname>Bracken-Clarke</surname><given-names>D</given-names> </name><name name-style="western"><surname>Floudas</surname><given-names>CS</given-names> </name><name name-style="western"><surname>Roselli</surname><given-names>M</given-names> </name><name name-style="western"><surname>Gulley</surname><given-names>JL</given-names> </name><name name-style="western"><surname>Karzai</surname><given-names>F</given-names> </name></person-group><article-title>Applications of large language models in cancer care: current evidence and future perspectives</article-title><source>Front Oncol</source><year>2023</year><volume>13</volume><fpage>1268915</fpage><pub-id pub-id-type="doi">10.3389/fonc.2023.1268915</pub-id><pub-id pub-id-type="medline">37731643</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhou</surname><given-names>S</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>N</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>L</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>H</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>R</given-names> </name></person-group><article-title>CancerBERT: a cancer domain-specific language model for extracting breast cancer phenotypes from electronic health records</article-title><source>J Am Med Inform Assoc</source><year>2022</year><month>06</month><day>14</day><volume>29</volume><issue>7</issue><fpage>1208</fpage><lpage>1216</lpage><pub-id pub-id-type="doi">10.1093/jamia/ocac040</pub-id><pub-id pub-id-type="medline">35333345</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fang</surname><given-names>A</given-names> </name><name name-style="western"><surname>Hu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Zhao</surname><given-names>W</given-names> </name><etal/></person-group><article-title>Extracting clinical named entity for pituitary adenomas from Chinese electronic medical records</article-title><source>BMC Med Inform Decis Mak</source><year>2022</year><month>03</month><day>23</day><volume>22</volume><issue>1</issue><fpage>72</fpage><pub-id pub-id-type="doi">10.1186/s12911-022-01810-z</pub-id><pub-id pub-id-type="medline">35321705</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>X</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>Q</given-names> </name><etal/></person-group><article-title>Extracting comprehensive clinical information for breast cancer using deep learning methods</article-title><source>Int J Med Inform</source><year>2019</year><month>12</month><volume>132</volume><fpage>103985</fpage><pub-id pub-id-type="doi">10.1016/j.ijmedinf.2019.103985</pub-id><pub-id pub-id-type="medline">31627032</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kim</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Meystre</surname><given-names>SM</given-names> </name></person-group><article-title>A study of medical problem extraction for better disease management</article-title><source>Stud Health Technol Inform</source><year>2019</year><month>08</month><day>21</day><volume>264</volume><fpage>193</fpage><lpage>197</lpage><pub-id pub-id-type="doi">10.3233/SHTI190210</pub-id><pub-id pub-id-type="medline">31437912</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Coquet</surname><given-names>J</given-names> </name><name name-style="western"><surname>Bozkurt</surname><given-names>S</given-names> </name><name name-style="western"><surname>Kan</surname><given-names>KM</given-names> </name><etal/></person-group><article-title>Comparison of orthogonal NLP methods for clinical phenotyping and assessment of bone scan utilization among prostate cancer patients</article-title><source>J Biomed Inform</source><year>2019</year><month>06</month><volume>94</volume><fpage>103184</fpage><pub-id pub-id-type="doi">10.1016/j.jbi.2019.103184</pub-id><pub-id pub-id-type="medline">31014980</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Miao</surname><given-names>S</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>T</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>Extraction of BI-RADS findings from breast ultrasound reports in Chinese using deep learning approaches</article-title><source>Int J Med Inform</source><year>2018</year><month>11</month><volume>119</volume><fpage>17</fpage><lpage>21</lpage><pub-id pub-id-type="doi">10.1016/j.ijmedinf.2018.08.009</pub-id><pub-id pub-id-type="medline">30342682</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hu</surname><given-names>D</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>B</given-names> </name><name name-style="western"><surname>Zhu</surname><given-names>X</given-names> </name><name name-style="western"><surname>Lu</surname><given-names>X</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>N</given-names> </name></person-group><article-title>Zero-shot information extraction from radiological reports using ChatGPT</article-title><source>Int J Med Inform</source><year>2024</year><month>03</month><volume>183</volume><fpage>105321</fpage><pub-id pub-id-type="doi">10.1016/j.ijmedinf.2023.105321</pub-id><pub-id pub-id-type="medline">38157785</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ansoborlo</surname><given-names>M</given-names> </name><name name-style="western"><surname>Gaborit</surname><given-names>C</given-names> </name><name name-style="western"><surname>Grammatico-Guillon</surname><given-names>L</given-names> </name><name name-style="western"><surname>Cuggia</surname><given-names>M</given-names> </name><name name-style="western"><surname>Bouzille</surname><given-names>G</given-names> </name></person-group><article-title>Prescreening in oncology trials using medical records. Natural language processing applied on lung cancer multidisciplinary team meeting reports</article-title><source>Health Informatics J</source><year>2023</year><volume>29</volume><issue>1</issue><fpage>146045822211467</fpage><pub-id pub-id-type="doi">10.1177/14604582221146709</pub-id><pub-id pub-id-type="medline">36964666</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>L&#x00F3;pez-Garc&#x00ED;a</surname><given-names>G</given-names> </name><name name-style="western"><surname>Jerez</surname><given-names>JM</given-names> </name><name name-style="western"><surname>Ribelles</surname><given-names>N</given-names> </name><name name-style="western"><surname>Alba</surname><given-names>E</given-names> </name><name name-style="western"><surname>Veredas</surname><given-names>FJ</given-names> </name></person-group><article-title>Detection of tumor morphology mentions in clinical reports in Spanish using transformers</article-title><source>Advances in Computational Intelligence</source><year>2021</year><fpage>24</fpage><lpage>35</lpage><pub-id pub-id-type="doi">10.1007/978-3-030-85030-2_3</pub-id><pub-id pub-id-type="other">9783030850296</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>H</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>Use of BERT (Bidirectional Encoder Representations from Transformers)-based deep learning method for extracting evidences in chinese radiology reports: development of a computer-aided liver cancer diagnosis framework</article-title><source>J Med Internet Res</source><year>2021</year><month>01</month><day>12</day><volume>23</volume><issue>1</issue><fpage>e19689</fpage><pub-id pub-id-type="doi">10.2196/19689</pub-id><pub-id pub-id-type="medline">33433395</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Solarte Pab&#x00F3;n</surname><given-names>O</given-names> </name><name name-style="western"><surname>Montenegro</surname><given-names>O</given-names> </name><name name-style="western"><surname>Torrente</surname><given-names>M</given-names> </name><name name-style="western"><surname>Rodr&#x00ED;guez Gonz&#x00E1;lez</surname><given-names>A</given-names> </name><name name-style="western"><surname>Provencio</surname><given-names>M</given-names> </name><name name-style="western"><surname>Menasalvas</surname><given-names>E</given-names> </name></person-group><article-title>Negation and uncertainty detection in clinical texts written in Spanish: a deep learning-based approach</article-title><source>PeerJ Comput Sci</source><year>2022</year><volume>8</volume><fpage>e913</fpage><pub-id pub-id-type="doi">10.7717/peerj-cs.913</pub-id><pub-id pub-id-type="medline">35494817</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yu</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>X</given-names> </name><name name-style="western"><surname>Dang</surname><given-names>C</given-names> </name><etal/></person-group><article-title>A study of social and behavioral determinants of health in lung cancer patients using transformers-based natural language processing models</article-title><source>AMIA Annu Symp Proc</source><year>2021</year><volume>2021</volume><fpage>1225</fpage><lpage>1233</lpage><pub-id pub-id-type="medline">35309014</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lu</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Sim</surname><given-names>JA</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>JX</given-names> </name><etal/></person-group><article-title>Natural language processing and machine learning methods to characterize unstructured patient-reported outcomes: validation study</article-title><source>J Med Internet Res</source><year>2021</year><month>11</month><day>3</day><volume>23</volume><issue>11</issue><fpage>e26777</fpage><pub-id pub-id-type="doi">10.2196/26777</pub-id><pub-id pub-id-type="medline">34730546</pub-id></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hu</surname><given-names>D</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>H</given-names> </name><name name-style="western"><surname>Li</surname><given-names>S</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>N</given-names> </name><name name-style="western"><surname>Lu</surname><given-names>X</given-names> </name></person-group><article-title>Automatic extraction of lung cancer staging information from computed tomography reports: deep learning approach</article-title><source>JMIR Med Inform</source><year>2021</year><month>07</month><day>21</day><volume>9</volume><issue>7</issue><fpage>e27955</fpage><pub-id pub-id-type="doi">10.2196/27955</pub-id><pub-id pub-id-type="medline">34287213</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hu</surname><given-names>D</given-names> </name><name name-style="western"><surname>Li</surname><given-names>S</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>H</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>N</given-names> </name><name name-style="western"><surname>Lu</surname><given-names>X</given-names> </name></person-group><article-title>Using natural language processing and machine learning to preoperatively predict lymph node metastasis for non-small cell lung cancer with electronic medical records: development and validation study</article-title><source>JMIR Med Inform</source><year>2022</year><month>04</month><day>25</day><volume>10</volume><issue>4</issue><fpage>e35475</fpage><pub-id pub-id-type="doi">10.2196/35475</pub-id><pub-id pub-id-type="medline">35468085</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zitu</surname><given-names>MM</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>S</given-names> </name><name name-style="western"><surname>Owen</surname><given-names>DH</given-names> </name><name name-style="western"><surname>Chiang</surname><given-names>C</given-names> </name><name name-style="western"><surname>Li</surname><given-names>L</given-names> </name></person-group><article-title>Generalizability of machine learning methods in detecting adverse drug events from clinical narratives in electronic medical records</article-title><source>Front Pharmacol</source><year>2023</year><volume>14</volume><fpage>1218679</fpage><pub-id pub-id-type="doi">10.3389/fphar.2023.1218679</pub-id><pub-id pub-id-type="medline">37502211</pub-id></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Seong</surname><given-names>D</given-names> </name><name name-style="western"><surname>Choi</surname><given-names>YH</given-names> </name><name name-style="western"><surname>Shin</surname><given-names>SY</given-names> </name><name name-style="western"><surname>Yi</surname><given-names>BK</given-names> </name></person-group><article-title>Deep learning approach to detection of colonoscopic information from unstructured reports</article-title><source>BMC Med Inform Decis Mak</source><year>2023</year><month>02</month><day>7</day><volume>23</volume><issue>1</issue><fpage>28</fpage><pub-id pub-id-type="doi">10.1186/s12911-023-02121-7</pub-id><pub-id pub-id-type="medline">36750932</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Rohanian</surname><given-names>O</given-names> </name><name name-style="western"><surname>Jauncey</surname><given-names>H</given-names> </name><name name-style="western"><surname>Nouriborji</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Using bottleneck adapters to identify cancer in clinical notes under low-resource constraints</article-title><source>Proceedings of the 22nd Workshop on Biomedical Natural Language Processing and BioNLP Shared Tasks</source><year>2023</year><publisher-name>Association for Computational Linguistics</publisher-name><fpage>62</fpage><lpage>78</lpage><pub-id pub-id-type="doi">10.18653/v1/2023.bionlp-1.5</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mart&#x00ED;n-Noguerol</surname><given-names>T</given-names> </name><name name-style="western"><surname>L&#x00F3;pez-&#x00DA;beda</surname><given-names>P</given-names> </name><name name-style="western"><surname>Pons-Escoda</surname><given-names>A</given-names> </name><name name-style="western"><surname>Luna</surname><given-names>A</given-names> </name></person-group><article-title>Natural language processing deep learning models for the differential between high-grade gliomas and metastasis: what if the key is how we report them?</article-title><source>Eur Radiol</source><year>2024</year><month>03</month><volume>34</volume><issue>3</issue><fpage>2113</fpage><lpage>2120</lpage><pub-id pub-id-type="doi">10.1007/s00330-023-10202-4</pub-id><pub-id pub-id-type="medline">37665389</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Thompson</surname><given-names>J</given-names> </name><name name-style="western"><surname>Hu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Mudaranthakam</surname><given-names>DP</given-names> </name><etal/></person-group><article-title>Relevant word order vectorization for improved natural language processing in electronic health records</article-title><source>Sci Rep</source><year>2019</year><month>06</month><day>25</day><volume>9</volume><issue>1</issue><fpage>9253</fpage><pub-id pub-id-type="doi">10.1038/s41598-019-45705-y</pub-id><pub-id pub-id-type="medline">31239489</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Qiu</surname><given-names>JX</given-names> </name><name name-style="western"><surname>Yoon</surname><given-names>HJ</given-names> </name><name name-style="western"><surname>Fearn</surname><given-names>PA</given-names> </name><name name-style="western"><surname>Tourassi</surname><given-names>GD</given-names> </name></person-group><article-title>Deep learning for automated extraction of primary sites from cancer pathology reports</article-title><source>IEEE J Biomed Health Inform</source><year>2018</year><month>01</month><volume>22</volume><issue>1</issue><fpage>244</fpage><lpage>251</lpage><pub-id pub-id-type="doi">10.1109/JBHI.2017.2700722</pub-id><pub-id pub-id-type="medline">28475069</pub-id></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Alawad</surname><given-names>M</given-names> </name><name name-style="western"><surname>Yoon</surname><given-names>HJ</given-names> </name><name name-style="western"><surname>Tourassi</surname><given-names>GD</given-names> </name></person-group><article-title>Coarse-to-fine multi-task training of convolutional neural networks for automated information extraction from cancer pathology reports</article-title><conf-name>2018 IEEE EMBS International Conference on Biomedical &#x0026; Health Informatics (BHI)</conf-name><conf-date>Mar 4-7, 2018</conf-date><conf-loc>Las Vegas, NV, USA</conf-loc><pub-id pub-id-type="doi">10.1109/BHI.2018.8333408</pub-id></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Dubey</surname><given-names>AK</given-names> </name><name name-style="western"><surname>Yoon</surname><given-names>HJ</given-names> </name><name name-style="western"><surname>Tourassi</surname><given-names>GD</given-names> </name></person-group><article-title>Inverse regression for extraction of tumor site from cancer pathology reports</article-title><conf-name>2019 IEEE EMBS International Conference on Biomedical &#x0026; Health Informatics (BHI)</conf-name><conf-date>May 19-22, 2019</conf-date><conf-loc>Chicago, IL, USA</conf-loc><pub-id pub-id-type="doi">10.1109/BHI.2019.8834527</pub-id></nlm-citation></ref><ref id="ref38"><label>38</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Wu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Tang</surname><given-names>K</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>H</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>C</given-names> </name><name name-style="western"><surname>Li</surname><given-names>C</given-names> </name></person-group><article-title>Structured information extraction of pathology reports with attention-based graph convolutional network</article-title><conf-name>2020 IEEE International Conference on Bioinformatics and Biomedicine (BIBM)</conf-name><conf-date>Dec 16-19, 2020</conf-date><conf-loc>Seoul, South Korea</conf-loc><pub-id pub-id-type="doi">10.1109/BIBM49941.2020.9313347</pub-id></nlm-citation></ref><ref id="ref39"><label>39</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Odisho</surname><given-names>AY</given-names> </name><name name-style="western"><surname>Park</surname><given-names>B</given-names> </name><name name-style="western"><surname>Altieri</surname><given-names>N</given-names> </name><etal/></person-group><article-title>Natural language processing systems for pathology parsing in limited data environments with uncertainty estimation</article-title><source>JAMIA Open</source><year>2020</year><month>10</month><volume>3</volume><issue>3</issue><fpage>431</fpage><lpage>438</lpage><pub-id pub-id-type="doi">10.1093/jamiaopen/ooaa029</pub-id><pub-id pub-id-type="medline">33381748</pub-id></nlm-citation></ref><ref id="ref40"><label>40</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Alawad</surname><given-names>M</given-names> </name><name name-style="western"><surname>Gao</surname><given-names>S</given-names> </name><name name-style="western"><surname>Qiu</surname><given-names>JX</given-names> </name><etal/></person-group><article-title>Automatic extraction of cancer registry reportable information from free-text pathology reports using multitask convolutional neural networks</article-title><source>J Am Med Inform Assoc</source><year>2020</year><month>01</month><day>1</day><volume>27</volume><issue>1</issue><fpage>89</fpage><lpage>98</lpage><pub-id pub-id-type="doi">10.1093/jamia/ocz153</pub-id><pub-id pub-id-type="medline">31710668</pub-id></nlm-citation></ref><ref id="ref41"><label>41</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Wu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>R</given-names> </name><name name-style="western"><surname>Gong</surname><given-names>T</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>C</given-names> </name><name name-style="western"><surname>Li</surname><given-names>C</given-names> </name></person-group><article-title>BioIE: biomedical information extraction with multi-head attention enhanced graph convolutional network</article-title><conf-name>2021 IEEE International Conference on Bioinformatics and Biomedicine (BIBM)</conf-name><conf-date>Dec 9-12, 2021</conf-date><conf-loc>Houston, TX, USA</conf-loc><pub-id pub-id-type="doi">10.1109/BIBM52615.2021.9669650</pub-id></nlm-citation></ref><ref id="ref42"><label>42</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Rios</surname><given-names>A</given-names> </name><name name-style="western"><surname>Durbin</surname><given-names>EB</given-names> </name><name name-style="western"><surname>Hands</surname><given-names>I</given-names> </name><name name-style="western"><surname>Kavuluru</surname><given-names>R</given-names> </name></person-group><article-title>Assigning ICD-o-3 codes to pathology reports using neural multi-task training with hierarchical regularization</article-title><year>2021</year><month>08</month><conf-name>BCB &#x2019;21</conf-name><conf-loc>Gainesville Florida</conf-loc><publisher-name>ACM</publisher-name><fpage>1</fpage><lpage>10</lpage><pub-id pub-id-type="doi">10.1145/3459930.3469541</pub-id></nlm-citation></ref><ref id="ref43"><label>43</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Park</surname><given-names>B</given-names> </name><name name-style="western"><surname>Altieri</surname><given-names>N</given-names> </name><name name-style="western"><surname>DeNero</surname><given-names>J</given-names> </name><name name-style="western"><surname>Odisho</surname><given-names>AY</given-names> </name><name name-style="western"><surname>Yu</surname><given-names>B</given-names> </name></person-group><article-title>Improving natural language information extraction from cancer pathology reports using transfer learning and zero-shot string similarity</article-title><source>JAMIA Open</source><year>2021</year><month>07</month><volume>4</volume><issue>3</issue><fpage>ooab085</fpage><pub-id pub-id-type="doi">10.1093/jamiaopen/ooab085</pub-id><pub-id pub-id-type="medline">34604711</pub-id></nlm-citation></ref><ref id="ref44"><label>44</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Vaswani</surname><given-names>A</given-names> </name><name name-style="western"><surname>Shazeer</surname><given-names>N</given-names> </name><name name-style="western"><surname>Parmar</surname><given-names>N</given-names> </name><etal/></person-group><article-title>Attention is all you need</article-title><source>arXiv</source><comment>Preprint posted online on  Jun 12, 2017</comment><pub-id pub-id-type="doi">10.48550/arXiv.1706.03762</pub-id></nlm-citation></ref><ref id="ref45"><label>45</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Devlin</surname><given-names>J</given-names> </name><name name-style="western"><surname>Chang</surname><given-names>MW</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>K</given-names> </name><name name-style="western"><surname>Toutanova</surname><given-names>K</given-names> </name></person-group><article-title>BERT: pre-training of deep bidirectional transformers for language understanding</article-title><source>arXiv</source><comment>Preprint posted online on  Oct 11, 2018</comment><pub-id pub-id-type="doi">10.48550/arXiv.1810.04805</pub-id></nlm-citation></ref><ref id="ref46"><label>46</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kreimeyer</surname><given-names>K</given-names> </name><name name-style="western"><surname>Foster</surname><given-names>M</given-names> </name><name name-style="western"><surname>Pandey</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Natural language processing systems for capturing and standardizing unstructured clinical information: a systematic review</article-title><source>J Biomed Inform</source><year>2017</year><month>09</month><volume>73</volume><fpage>14</fpage><lpage>29</lpage><pub-id pub-id-type="doi">10.1016/j.jbi.2017.07.012</pub-id><pub-id pub-id-type="medline">28729030</pub-id></nlm-citation></ref><ref id="ref47"><label>47</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Datta</surname><given-names>S</given-names> </name><name name-style="western"><surname>Bernstam</surname><given-names>EV</given-names> </name><name name-style="western"><surname>Roberts</surname><given-names>K</given-names> </name></person-group><article-title>A frame semantic overview of NLP-based information extraction for cancer-related EHR notes</article-title><source>J Biomed Inform</source><year>2019</year><month>12</month><volume>100</volume><fpage>103301</fpage><pub-id pub-id-type="doi">10.1016/j.jbi.2019.103301</pub-id><pub-id pub-id-type="medline">31589927</pub-id></nlm-citation></ref><ref id="ref48"><label>48</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bilal</surname><given-names>M</given-names> </name><name name-style="western"><surname>Hamza</surname><given-names>A</given-names> </name><name name-style="western"><surname>Malik</surname><given-names>N</given-names> </name></person-group><article-title>NLP for analyzing electronic health records and clinical notes in cancer research: a review</article-title><source>J Pain Symptom Manage</source><year>2025</year><month>05</month><volume>69</volume><issue>5</issue><fpage>e374</fpage><lpage>e394</lpage><pub-id pub-id-type="doi">10.1016/j.jpainsymman.2025.01.019</pub-id><pub-id pub-id-type="medline">39894080</pub-id></nlm-citation></ref><ref id="ref49"><label>49</label><nlm-citation citation-type="report"><person-group person-group-type="author"><name name-style="western"><surname>Giorgia</surname><given-names>T</given-names> </name><name name-style="western"><surname>Johannes</surname><given-names>CS</given-names> </name><name name-style="western"><surname>Gerasimos</surname><given-names>S</given-names> </name></person-group><article-title>A study of BERT&#x2019;s processing of negations to determine sentiment</article-title><year>2021</year><publisher-name>BNAIC/BeneLearn</publisher-name></nlm-citation></ref><ref id="ref50"><label>50</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ettinger</surname><given-names>A</given-names> </name></person-group><article-title>What BERT is not: lessons from a new suite of psycholinguistic diagnostics for language models</article-title><source>Trans Assoc Comput Linguist</source><year>2020</year><month>12</month><volume>8</volume><fpage>34</fpage><lpage>48</lpage><pub-id pub-id-type="doi">10.1162/tacl_a_00298</pub-id></nlm-citation></ref><ref id="ref51"><label>51</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Choi</surname><given-names>HS</given-names> </name><name name-style="western"><surname>Song</surname><given-names>JY</given-names> </name><name name-style="western"><surname>Shin</surname><given-names>KH</given-names> </name><name name-style="western"><surname>Chang</surname><given-names>JH</given-names> </name><name name-style="western"><surname>Jang</surname><given-names>BS</given-names> </name></person-group><article-title>Developing prompts from large language model for extracting clinical information from pathology and ultrasound reports in breast cancer</article-title><source>Radiat Oncol J</source><year>2023</year><month>09</month><volume>41</volume><issue>3</issue><fpage>209</fpage><lpage>216</lpage><pub-id pub-id-type="doi">10.3857/roj.2023.00633</pub-id><pub-id pub-id-type="medline">37793630</pub-id></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>Search strategy.</p><media xlink:href="medinform_v13i1e68707_app1.docx" xlink:title="DOCX File, 12 KB"/></supplementary-material><supplementary-material id="app2"><label>Multimedia Appendix 2</label><p>Statistical significance of <italic>t</italic> test results.</p><media xlink:href="medinform_v13i1e68707_app2.docx" xlink:title="DOCX File, 14 KB"/></supplementary-material><supplementary-material id="app3"><label>Checklist 1</label><p>PRISMA 2020 checklist.</p><media xlink:href="medinform_v13i1e68707_app3.pdf" xlink:title="PDF File, 67 KB"/></supplementary-material></app-group></back></article>