<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMI</journal-id>
      <journal-id journal-id-type="nlm-ta">JMIR Med Inform</journal-id>
      <journal-title>JMIR Medical Informatics</journal-title>
      <issn pub-type="epub">2291-9694</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v8i11e19735</article-id>
      <article-id pub-id-type="pmid">33226350</article-id>
      <article-id pub-id-type="doi">10.2196/19735</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>Measurement of Semantic Textual Similarity in Clinical Texts: Comparison of Transformer-Based Models</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Wang</surname>
            <given-names>Yanshan</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Li</surname>
            <given-names>Fang</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Lei</surname>
            <given-names>Jianbo</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Mavragani</surname>
            <given-names>Amaryllis</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author">
          <name name-style="western">
            <surname>Yang</surname>
            <given-names>Xi</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-2981-3972</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author">
          <name name-style="western">
            <surname>He</surname>
            <given-names>Xing</given-names>
          </name>
          <degrees>MSc</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-0290-8058</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author">
          <name name-style="western">
            <surname>Zhang</surname>
            <given-names>Hansi</given-names>
          </name>
          <degrees>MSc</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-4585-0888</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author">
          <name name-style="western">
            <surname>Ma</surname>
            <given-names>Yinghan</given-names>
          </name>
          <degrees>BSc</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-4914-641X</ext-link>
        </contrib>
        <contrib id="contrib5" contrib-type="author">
          <name name-style="western">
            <surname>Bian</surname>
            <given-names>Jiang</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-2238-5429</ext-link>
        </contrib>
        <contrib id="contrib6" contrib-type="author" corresp="yes">
          <name name-style="western">
            <surname>Wu</surname>
            <given-names>Yonghui</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <address>
            <institution>Department of Health Outcomes and Biomedical Informatics</institution>
            <institution>University of Florida</institution>
            <addr-line>2004 Mowry Road</addr-line>
            <addr-line>Gainesville, FL, 32610</addr-line>
            <country>United States</country>
            <phone>1 352 294 8436</phone>
            <email>yonghui.wu@ufl.edu</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-6780-6135</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>Department of Health Outcomes and Biomedical Informatics</institution>
        <institution>University of Florida</institution>
        <addr-line>Gainesville, FL</addr-line>
        <country>United States</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Yonghui Wu <email>yonghui.wu@ufl.edu</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <month>11</month>
        <year>2020</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>23</day>
        <month>11</month>
        <year>2020</year>
      </pub-date>
      <volume>8</volume>
      <issue>11</issue>
      <elocation-id>e19735</elocation-id>
      <history>
        <date date-type="received">
          <day>27</day>
          <month>7</month>
          <year>2020</year>
        </date>
        <date date-type="rev-request">
          <day>6</day>
          <month>10</month>
          <year>2020</year>
        </date>
        <date date-type="rev-recd">
          <day>19</day>
          <month>10</month>
          <year>2020</year>
        </date>
        <date date-type="accepted">
          <day>26</day>
          <month>10</month>
          <year>2020</year>
        </date>
      </history>
      <copyright-statement>©Xi Yang, Xing He, Hansi Zhang, Yinghan Ma, Jiang Bian, Yonghui Wu. Originally published in JMIR Medical Informatics (http://medinform.jmir.org), 23.11.2020.</copyright-statement>
      <copyright-year>2020</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on http://medinform.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="http://medinform.jmir.org/2020/11/e19735/" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>Semantic textual similarity (STS) is one of the fundamental tasks in natural language processing (NLP). Many shared tasks and corpora for STS have been organized and curated in the general English domain; however, such resources are limited in the biomedical domain. In 2019, the National NLP Clinical Challenges (n2c2) challenge developed a comprehensive clinical STS dataset and organized a community effort to solicit state-of-the-art solutions for clinical STS.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>This study presents our transformer-based clinical STS models developed during this challenge as well as new models we explored after the challenge. This project is part of the 2019 n2c2/Open Health NLP shared task on clinical STS.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>In this study, we explored 3 transformer-based models for clinical STS: Bidirectional Encoder Representations from Transformers (BERT), XLNet, and Robustly optimized BERT approach (RoBERTa). We examined transformer models pretrained using both general English text and clinical text. We also explored using a general English STS dataset as a supplementary corpus in addition to the clinical training set developed in this challenge. Furthermore, we investigated various ensemble methods to combine different transformer models.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>Our best submission based on the XLNet model achieved the third-best performance (Pearson correlation of 0.8864) in this challenge. After the challenge, we further explored other transformer models and improved the performance to 0.9065 using a RoBERTa model, which outperformed the best-performing system developed in this challenge (Pearson correlation of 0.9010).</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>This study demonstrated the efficiency of utilizing transformer-based models to measure semantic similarity for clinical text. Our models can be applied to clinical applications such as clinical text deduplication and summarization.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>clinical semantic textual similarity</kwd>
        <kwd>deep learning</kwd>
        <kwd>natural language processing</kwd>
        <kwd>transformers</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <p>Semantic textual similarity (STS) is a natural language processing (NLP) task to quantitatively assess the semantic similarity between two text snippets. STS is usually approached as a regression task where a real-value score is used to quantify the similarity between two text snippets. STS is a fundamental NLP task for many text-related applications, including text deduplication, paraphrasing detection, semantic searching, and question answering. In the general English domain, semantic evaluation (SemEval) STS shared tasks have been organized annually from 2012 to 2017 [<xref ref-type="bibr" rid="ref1">1</xref>-<xref ref-type="bibr" rid="ref6">6</xref>], and STS benchmark datasets were developed for evaluation [<xref ref-type="bibr" rid="ref6">6</xref>]. Previous work on STS often used machine learning models [<xref ref-type="bibr" rid="ref7">7</xref>-<xref ref-type="bibr" rid="ref9">9</xref>] such as support vector machine [<xref ref-type="bibr" rid="ref10">10</xref>], random forest [<xref ref-type="bibr" rid="ref11">11</xref>], convolutional neural networks [<xref ref-type="bibr" rid="ref12">12</xref>], and recurrent neural networks [<xref ref-type="bibr" rid="ref13">13</xref>] and topic modeling techniques [<xref ref-type="bibr" rid="ref8">8</xref>] such as latent semantic analysis [<xref ref-type="bibr" rid="ref14">14</xref>] and latent Dirichlet allocation [<xref ref-type="bibr" rid="ref15">15</xref>]. Recently, deep learning models based on transformer architectures such as Bidirectional Encoder Representations from Transformers (BERT) [<xref ref-type="bibr" rid="ref16">16</xref>], XLNet [<xref ref-type="bibr" rid="ref17">17</xref>], and Robustly optimized BERT approach (RoBERTa) [<xref ref-type="bibr" rid="ref18">18</xref>] have demonstrated state-of-the-art performances on the STS benchmark dataset [<xref ref-type="bibr" rid="ref19">19</xref>] and remarkably outperformed the previous models. More recently, the Text-to-Text Transfer Transformer model [<xref ref-type="bibr" rid="ref20">20</xref>] and the StructBERT model [<xref ref-type="bibr" rid="ref21">21</xref>] have further improved the performance on the STS benchmark. These studies demonstrated the efficiency of transformer-based models for STS tasks.</p>
      <p>Rapid adoption of electronic health record (EHR) systems has made longitudinal health information of patients available electronically [<xref ref-type="bibr" rid="ref22">22</xref>,<xref ref-type="bibr" rid="ref23">23</xref>]. EHRs consist of structured, coded data and clinical narratives. The structured EHR data are typically stored as predefined medical codes (eg, International Classification of Diseases, 9th/10th Revision, codes for diagnoses) in relational databases. Various common data models were used to standardize EHR data to facilitate downstream research and clinical studies [<xref ref-type="bibr" rid="ref24">24</xref>]. However, clinical narratives are often documented in a free-text format, which contains many types of detailed patient information, such as family history, adverse drug events, and medical imaging result interpretations, that are not well captured in the structured medical codes [<xref ref-type="bibr" rid="ref25">25</xref>]. As free text, the clinical notes may contain a considerable amount of duplication, error, and incompleteness for various reasons (eg, copy-and-paste or using templates and inconsistent modifications) [<xref ref-type="bibr" rid="ref26">26</xref>,<xref ref-type="bibr" rid="ref27">27</xref>]. STS can be applied to assess the quality of the clinical notes and reduce redundancy to support downstream NLP tasks [<xref ref-type="bibr" rid="ref28">28</xref>]. However, up until now, only a few studies [<xref ref-type="bibr" rid="ref29">29</xref>-<xref ref-type="bibr" rid="ref31">31</xref>] have explored STS in the clinical domain due to the limited data resources for developing and benchmarking clinical STS tasks. Recently, a team at the Mayo Clinic developed a clinical STS dataset, MedSTS [<xref ref-type="bibr" rid="ref32">32</xref>], which consists of more than 1000 annotated sentence pairs extracted from clinical notes. Based on the MedSTS dataset, the 2018 BioCreative/Open Health NLP (OHNLP) challenge [<xref ref-type="bibr" rid="ref33">33</xref>] was organized as the first shared task examining advanced NLP methods for STS in the clinical domain. In this challenge, two different teams explored various machine learning approaches, including several deep learning models [<xref ref-type="bibr" rid="ref30">30</xref>,<xref ref-type="bibr" rid="ref31">31</xref>]. Later, more teams competed in the 2019 National NLP Clinical Challenges (n2c2)/OHNLP STS challenge with a larger clinical STS dataset [<xref ref-type="bibr" rid="ref34">34</xref>]. During this challenge, many new emerging NLP techniques, such as transformer-based models, were explored.</p>
      <p>This study presents our machine learning models developed for the 2019 n2c2/OHNLP STS challenge. We explored state-of-the-art transformer-based models (BERT, XLNet, and RoBERTa) for clinical STS. We systematically examined transformer models pretrained using general English corpora and compared them with clinical transformer models pretrained using clinical corpora. We also proposed a representation fusion method to ensemble the transformer-based models. In this challenge, our clinical STS system based on the XLNet model achieved a Pearson correlation score of 0.8864, ranked as the third-best performance among all participants. After the challenge, we further explored a new transformer-based model, RoBERTa, which improved the performance to 0.9065 and outperformed the best performance (0.9010) reported in this challenge. This study demonstrated the efficiency of transformer-based models for STS in the clinical domain.</p>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <sec>
        <title>Dataset</title>
        <p>The 2019 n2c2 organizers developed a corpus of 2054 sentence pairs derived from over 300 million deidentified clinical notes from the Mayo Clinic’s EHR data warehouse. The sentence pairs were divided into a training set of 1642 sentence pairs for model development and a test set of 412 sentence pairs for evaluation. Similar to the annotation scheme in the general English domain, the challenge corpus was annotated by assigning a similarity score for each sentence pair as a number on a scale from 0.0 to 5.0, where 0.0 indicates that the semantics of the two sentences are entirely independent (ie, no overlap in their meanings), and 5.0 signifies that two sentences are semantically equivalent. Annotators used arbitrary similarity scores between 0.0 and 5.0, such as 2.5 or 3.5, to reflect different levels of equality. <xref ref-type="table" rid="table1">Table 1</xref> presents the descriptive statistics of the datasets. The distribution of similarity scores is quite different between the training and test datasets. In the training set, the range with the most cases (509/1642, 31.0%) was (3.0, 4.0], whereas in the test set, most scores (238/412, 57.8%) were distributed in the range (0.0, 1.0]. In this study, we denoted this challenge dataset as STS-Clinic. In addition to the STS-Clinic, we also used a general English domain STS benchmark dataset from the SemEval 2017 [<xref ref-type="bibr" rid="ref6">6</xref>] as an external source. We merged the original training and development datasets to create a unique dataset of 7249 annotated sentence pairs. We denoted this combined general English domain dataset as STS-General and used it as a complementary training set for model development in this study. Compared to the STS-Clinic, the similarity scores in STS-General were more evenly distributed in different ranges (<xref ref-type="table" rid="table1">Table 1</xref>).</p>
        <table-wrap position="float" id="table1">
          <label>Table 1</label>
          <caption>
            <p>Descriptive statistics of the datasets.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="240"/>
            <col width="160"/>
            <col width="120"/>
            <col width="120"/>
            <col width="120"/>
            <col width="120"/>
            <col width="120"/>
            <thead>
              <tr valign="top">
                <td>Dataset</td>
                <td>Sentence pairs, n</td>
                <td colspan="5">Annotation distribution, n (%)</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>[0.0, 1.0]</td>
                <td>(1.0, 2.0]</td>
                <td>(2.0, 3.0]</td>
                <td>(3.0, 4.0]</td>
                <td>(4.0, 5.0]</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>STS-Clinic<sup>a</sup> Training</td>
                <td>1642</td>
                <td>312 (19.0)</td>
                <td>154 (9.4)</td>
                <td>394 (24.0)</td>
                <td>509 (31.0)</td>
                <td>273 (16.6)</td>
              </tr>
              <tr valign="top">
                <td>STS-Clinic Test</td>
                <td>412</td>
                <td>238 (57.8)</td>
                <td>46 (11.2)</td>
                <td>32 (7.8)</td>
                <td>62 (15.0)</td>
                <td>34 (8.3)</td>
              </tr>
              <tr valign="top">
                <td>STS-General Training</td>
                <td>7249</td>
                <td>1492 (20.6)</td>
                <td>1122 (15.5)</td>
                <td>1413 (19.5)</td>
                <td>1260 (17.4)</td>
                <td>1962 (27.1)</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table1fn1">
              <p><sup>a</sup>STS: semantic textual similarity.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Preprocessing of Sentence Pairs</title>
        <p>We developed a preprocessing pipeline to normalize each sentence pair, including (1) converting all words to lower case; (2) inserting white spaces to separate words from punctuation (eg, “[ab/cd]” → “[ ab / cd ]”; “abc,def” → “abc , def”); and (3) replacing two or more spaces or tabs (“\t”) with a single space. We did not remove any stop-words from the sentences and kept the original formats of the numbers without any conversion. Since different transformer models adopted different tokenization strategies (eg, WordPiece for BERT, byte pair encoding for RoBERTa, and SentencePiece for XLNet), our preprocessing automatically picked the appropriate tokenizer according to the transformer model in use.</p>
        <sec>
          <title>Transformer Model-Based STS System</title>
          <p>In this study, we investigated three transformer models (BERT, XLNet, and RoBERTa) for clinical STS. BERT is a bidirectional transformer-based encoder model pretrained with a combination of masked language modeling (MLM) and next sentence prediction. RoBERTa has the same architecture as BERT but pretrained with a robust optimizing strategy. The RoBERTa pretraining procedure used dynamic MLM but removed the next sentence prediction task. XLNet is a transformer-based model pretrained with the bidirectional autoregressive language modeling method. Unlike the MLM used by BERT and RoBERTa, the autoregressive language model uses data permutation instead of data corruption and reconstruction. All three transformer models provided two different settings: a “BASE” setting and a “LARGE” setting. The main difference between the BASE model and the LARGE model is the number of layers. For example, the BERT-base model features 12 layers of transformer encoder layers, 768 hidden units in each layer, and 12 attention heads, while the BERT-large consists of 24 transformer blocks with a hidden size of 1024 and 16 attention heads. The total number of parameters for the BERT-large model is approximately 340 million, which is about 3 times more than the BERT-base model. In this study, we explored general transformers (pretrained using general English corpora) using both the BASE model and the LARGE model. We also examined clinical transformers pretrained using clinical notes from the MIMIC-III database. For clinical transformers, we adopted the BASE settings as we did not observe additional benefits from using the LARGE setting.</p>
          <p>As shown in <xref rid="figure1" ref-type="fig">Figure 1</xref>, our STS system has two modules: (1) a transformer model–based feature learning module and (2) a regression-based similarity score learning module. In the feature learning module, transformer-based models were applied to learn distributed sentence-level representations from sentence pairs. In the similarity score learning module, we adopted a linear regression layer to calculate a similarity score between 0.0 and 5.0 according to the distributed representations derived from the transformers. We explored both single-model and ensemble solutions. <xref rid="figure1" ref-type="fig">Figure 1</xref>A shows the single-model solution where only one transformer-based model was used for feature representation learning. <xref rid="figure1" ref-type="fig">Figure 1</xref>B shows the ensemble solution where different transformer models were integrated. Ensemble learning is an efficient approach to aggregate different machine learning models to achieve better performance [<xref ref-type="bibr" rid="ref35">35</xref>]. In this work, we tried different strategies to combine the distributed representations from two or three transformers as a new input layer for the similarity score learning module. We explored several methods to combine the distributed representations from different transformers, including (1) simple head-to-tail concatenation, (2) pooling, and (3) convolution.</p>
          <fig id="figure1" position="float">
            <label>Figure 1</label>
            <caption>
              <p>An overview of our single-model and ensemble solutions for clinical STS. STS: semantic textual similarity.</p>
            </caption>
            <graphic xlink:href="medinform_v8i11e19735_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>Training Strategy</title>
          <p>As shown in <xref rid="figure2" ref-type="fig">Figure 2</xref>, we adopted a two-phase procedure to train our clinical STS models. In the first phase, an intermediate STS model was fine-tuned using the STS-General corpus. Subsequently, the intermediate model was further fine-tuned using the STS-Clinic corpus in phase 2. The fine-tuned model from the second phase was used for final testing. We used 5-fold cross-validation for hyperparameter optimization in both phase 1 and phase 2 training. We optimized the epoch number, batch size, and learning rate according to the cross-validation results.</p>
          <fig id="figure2" position="float">
            <label>Figure 2</label>
            <caption>
              <p>The two-stage procedure for clinical STS model development. STS: semantic textual similarity.</p>
            </caption>
            <graphic xlink:href="medinform_v8i11e19735_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>Experiments and Evaluations</title>
          <p>In this study, we implemented our STS system using the Transformers library developed by the HuggingFace team [<xref ref-type="bibr" rid="ref36">36</xref>]. We also used the PyTorch-based general transformer models trained using general English corpora maintained by the HuggingFace team. The clinical transformer models were derived by further pretraining these general transformer models with clinical notes from the MIMIC-III database [<xref ref-type="bibr" rid="ref37">37</xref>]. <xref ref-type="table" rid="table2">Table 2</xref> shows the hyperparameters used for each transformer model. For evaluation, the results were calculated as the Pearson correlation scores using the official evaluation script provided by the 2019 n2c2/OHNLP challenge organizers. To report the <italic>P</italic> value for each Pearson correlation score, we adopted the SciPy package [<xref ref-type="bibr" rid="ref38">38</xref>].</p>
          <table-wrap position="float" id="table2">
            <label>Table 2</label>
            <caption>
              <p>Hyperparameters for transformer models.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="400"/>
              <col width="230"/>
              <col width="180"/>
              <col width="190"/>
              <thead>
                <tr valign="bottom">
                  <td>Model</td>
                  <td>Number of epochs</td>
                  <td>Batch size</td>
                  <td>Learning rate<sup>a</sup></td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>BERT-base<sup>b</sup></td>
                  <td>4</td>
                  <td>8</td>
                  <td>1.00E-05</td>
                </tr>
                <tr valign="top">
                  <td>BERT-mimic</td>
                  <td>3</td>
                  <td>8</td>
                  <td>1.00E-05</td>
                </tr>
                <tr valign="top">
                  <td>BERT-large</td>
                  <td>3</td>
                  <td>8</td>
                  <td>1.00E-05</td>
                </tr>
                <tr valign="top">
                  <td>XLNet-base</td>
                  <td>3</td>
                  <td>4</td>
                  <td>1.00E-05</td>
                </tr>
                <tr valign="top">
                  <td>XLNet-mimic</td>
                  <td>3</td>
                  <td>4</td>
                  <td>1.00E-05</td>
                </tr>
                <tr valign="top">
                  <td>XLNet-large</td>
                  <td>4</td>
                  <td>4</td>
                  <td>1.00E-05</td>
                </tr>
                <tr valign="top">
                  <td>RoBERTa-base<sup>c</sup></td>
                  <td>3</td>
                  <td>4</td>
                  <td>1.00E-05</td>
                </tr>
                <tr valign="top">
                  <td>RoBERTa-mimic</td>
                  <td>3</td>
                  <td>4</td>
                  <td>1.00E-05</td>
                </tr>
                <tr valign="top">
                  <td>RoBERTa-large</td>
                  <td>3</td>
                  <td>4</td>
                  <td>1.00E-05</td>
                </tr>
                <tr valign="top">
                  <td>BERT-large + XLNet-large</td>
                  <td>4</td>
                  <td>8</td>
                  <td>1.00E-05</td>
                </tr>
                <tr valign="top">
                  <td>BERT-large + RoBERTa-large</td>
                  <td>3</td>
                  <td>4</td>
                  <td>1.00E-05</td>
                </tr>
                <tr valign="top">
                  <td>RoBERTa-large + XLNet-large</td>
                  <td>4</td>
                  <td>4</td>
                  <td>1.00E-05</td>
                </tr>
                <tr valign="top">
                  <td>BERT-large + XLNet-large + RoBERTa-large</td>
                  <td>3</td>
                  <td>2</td>
                  <td>1.00E-05</td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table2fn1">
                <p><sup>a</sup>The learning rate is a tuning parameter in an optimization algorithm that determines the step size at each iteration while moving toward a minimum of a loss function [<xref ref-type="bibr" rid="ref39">39</xref>].</p>
              </fn>
              <fn id="table2fn2">
                <p><sup>b</sup>BERT: Bidirectional Encoder Representations from Transformers.</p>
              </fn>
              <fn id="table2fn3">
                <p><sup>c</sup>RoBERTa: Robustly optimized BERT approach.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
        </sec>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <p><xref ref-type="table" rid="table3">Table 3</xref> compares the performance of the different transformer models on the test dataset. The RoBERTa-large model achieved the best Pearson correlation of 0.9065 among all models, which outperformed the two models we developed and submitted during the challenge, including the XLNet-large (a Pearson correlation score of 0.8864) and the BERT-large models (a Pearson correlation score of 0.8549). For RoBERTa and XLNet, the models developed using the LARGE setting pretrained using general English corpora achieved better performances than their BASE settings (0.9065 vs 0.8778 for RoBERTa; 0.8864 vs 0.8470 for XLNet, respectively), whereas the BERT-base achieved a Pearson correlation score of 0.8615 that outperformed the BERT-large model’s score of 0.8549. For all transformers, the models pretrained using general English corpora (in both LARGE settings and BASE settings) outperformed their corresponding clinical models pretrained using clinical notes from the MIMIC-III database. Among the ensemble models, the BERT-large + RoBERTa-large model achieved the best Pearson correlation score of 0.8914, which is remarkably lower than the best model, RoBERTa-large. We also observed that the performances of ensemble models were often in between the two individual models (eg, BERT-large + RoBERTa-large achieved 0.8914, which is between the BERT-large score of 0.8549 and RoBERTa-large score of 0.9065). The ensemble model of all three transformers achieved a Pearson correlation of 0.8452, which was even worse.</p>
      <table-wrap position="float" id="table3">
        <label>Table 3</label>
        <caption>
          <p>Performances of the Pearson correlation on the test set.</p>
        </caption>
        <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
          <col width="500"/>
          <col width="400"/>
          <col width="100"/>
          <thead>
            <tr valign="top">
              <td>Model</td>
              <td>Pearson correlation on test set</td>
              <td><italic>P</italic> value</td>
            </tr>
          </thead>
          <tbody>
            <tr valign="top">
              <td>BERT-base<sup>a</sup></td>
              <td>0.8615</td>
              <td>&#60;.001</td>
            </tr>
            <tr valign="top">
              <td>BERT-mimic</td>
              <td>0.8521</td>
              <td>&#60;.001</td>
            </tr>
            <tr valign="top">
              <td>BERT-large<sup>b</sup></td>
              <td>0.8549</td>
              <td>&#60;.001</td>
            </tr>
            <tr valign="top">
              <td>XLNet-base</td>
              <td>0.8470</td>
              <td>&#60;.001</td>
            </tr>
            <tr valign="top">
              <td>XLNet-mimic</td>
              <td>0.8286</td>
              <td>&#60;.001</td>
            </tr>
            <tr valign="top">
              <td>XLNet-large<sup>b,c</sup></td>
              <td>0.8864</td>
              <td>&#60;.001</td>
            </tr>
            <tr valign="top">
              <td>RoBERTa-base<sup>d</sup></td>
              <td>0.8778</td>
              <td>&#60;.001</td>
            </tr>
            <tr valign="top">
              <td>RoBERTa-mimic</td>
              <td>0.8705</td>
              <td>&#60;.001</td>
            </tr>
            <tr valign="top">
              <td>RoBERTa-large</td>
              <td>
                <italic>0.9065</italic>
              </td>
              <td>&#60;.001</td>
            </tr>
            <tr valign="top">
              <td>BERT-large + XLNet-large<sup>b</sup></td>
              <td>0.8764</td>
              <td>&#60;.001</td>
            </tr>
            <tr valign="top">
              <td>BERT-large + RoBERTa-large</td>
              <td>0.8914</td>
              <td>&#60;.001</td>
            </tr>
            <tr valign="top">
              <td>RoBERTa-large + XLNet-large</td>
              <td>0.8854</td>
              <td>&#60;.001</td>
            </tr>
            <tr valign="top">
              <td>BERT-large + XLNet-large + RoBERTa-large</td>
              <td>0.8452</td>
              <td>&#60;.001</td>
            </tr>
          </tbody>
        </table>
        <table-wrap-foot>
          <fn id="table3fn1">
            <p><sup>a</sup>BERT: Bidirectional Encoder Representations from Transformers.</p>
          </fn>
          <fn id="table3fn2">
            <p><sup>b</sup>The challenge submissions.</p>
          </fn>
          <fn id="table3fn3">
            <p><sup>c</sup>The best challenge submission (ranked 3rd).</p>
          </fn>
          <fn id="table3fn4">
            <p><sup>d</sup>RoBERTa: Robustly optimized BERT approach.</p>
          </fn>
        </table-wrap-foot>
      </table-wrap>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <sec>
        <title>Principal Results</title>
        <p>Clinical STS is a fundamental task in biomedical NLP. The 2019 n2c2/OHNLP shared task was organized to solicit state-of-the-art STS algorithms in the clinical domain. We participated in this challenge and developed a deep learning–based system using transformer-based models. Our best submission (XLNet-large) achieved the third-best performance (a Pearson correlation score of 0.8864) among the 33 teams. Based on our participation, we further explored RoBERTa models and improved the performance to 0.9065 (RoBERTa-large), demonstrating the efficiency of transformer models for clinical STS. We also further explored three different ensemble strategies to develop ensembled models using transformers. Our experimental results show that the ensemble methods did not outperform the unified individual models. Another interesting finding is that the transformers pretrained using the clinical notes from the MIMIC-III database did not outperform the general transformers pretrained using general English corpora on clinical STS. One possible reason might be that the clinical corpora we used for training are relatively small compared with the general English corpus. Further investigation examining these findings is warranted.</p>
      </sec>
      <sec>
        <title>Experiment Findings</title>
        <p>Although previous studies [<xref ref-type="bibr" rid="ref40">40</xref>-<xref ref-type="bibr" rid="ref44">44</xref>] have shown that pretraining transformer models with domain-specific corpora could enhance their performances in domain-related downstream tasks (such as clinical concept extraction), our results in this study indicated that this strategy might not be helpful for clinical STS. For all three types of transformers explored in this study, the models pretrained using general English text consistently obtained higher scores than the corresponding models pretrained using clinical text. For example, the Pearson correlation score achieved by the RoBERTa-mimic was 0.8705; however, the RoBERTa-base yielded a higher performance of 0.8778. Tawfik et al [<xref ref-type="bibr" rid="ref45">45</xref>] have similarly observed that the PubMed pretrained BioBERT did not outperform the corresponding general BERT model pretrained using English text on clinical STS.</p>
        <p>In the clinical STS task, using STS-General (an STS corpus annotated in the general English domain) as an extra training set in addition to STS-Clinic could efficiently improve performances for transformer-based models. Taking the RoBERTa model as an example, the RoBERTa-large fine-tuned using only the clinical text (ie, STS-Clinic) achieved a Pearson correlation score of 0.8720; however, the same model fine-tuned with both the general English text (ie, STS-General) and clinical text (ie, STS-Clinic) achieved a score of 0.9065 (approximately 0.035 higher). We observed similar results for BERT and XLNet. Without Phase 1 (<xref rid="figure2" ref-type="fig">Figure 2</xref>), the BERT-large and XLNet-large models achieved Pearson correlation scores of 0.8413 and 0.8626, respectively, which are lower than the results we submitted (0.8549 and 0.8864) using two-phase training. We looked into the training datasets for possible reasons. Although the STS-General and STS-Clinic were extracted from different domains, there are common contents shared between them. First, the annotation guidelines between the two datasets were highly aligned. For both datasets, the annotation scale is from 0.0 to 5.0, and each score reflects the same similarity level. Since the two STS datasets were annotated by different annotators, subjective annotation bias might be introduced (eg, the judgement and agreement of semantic similarity among annotators might be different in the two datasets). However, our experiment results showed that training with both datasets improved the performance despite the potential annotation bias. Second, a considerable portion of STS-Clinic sentence pairs are common descriptions that do not require comprehensive clinical knowledge to interpret the semantics. Typical examples include sentences extracted from Consultation Note or Discharge Summary as follows:</p>
        <disp-quote>
          <p>Plan: the patient stated an understanding of the program, and agrees to continue independently with a home management program.</p>
        </disp-quote>
        <disp-quote>
          <p>Thank you for choosing the name M.D. care team for your health care needs!</p>
        </disp-quote>
        <p>On the other hand, there are many sentences in the STS-General associated with healthcare. An example is exhibited below:</p>
        <disp-quote>
          <p>Although obesity can increase the risk of health problems, skeptics argue, so do smoking and high cholesterol.</p>
        </disp-quote>
        <p>Tang et al [<xref ref-type="bibr" rid="ref30">30</xref>] have demonstrated that combining representations derived from different models is an efficient strategy in clinical STS. We explored similar strategies to combine sentence-level distributed representations, including vector concatenation, average pooling, max pooling, and convolution. Surprisingly, our results showed that such ensemble strategies did not help transformer-based STS systems. For example, for the ensemble model derived from the BERT-large and the XLNet-large models (ie, BERT-large + XLNet-large), the achieved Pearson correlation scores for vector concatenation, average pooling, max pooling, and convolution were 0.8764, 0.8760, 0.8799, and 0.8803, respectively. All the results were approximately 0.01 lower than that for XLNet-large (0.8864). We also observed that ensemble models’ performances were consistently in between the two individual models (0.8549 for BERT-large and 0.8864 for XLNet-large). Future studies should examine this finding.</p>
        <p>To examine the statistical significance among different models’ results, we used a 1-tailed parametric test based on the Fisher Z-transformation [<xref ref-type="bibr" rid="ref46">46</xref>], adopted in the previous SemEval STS shared tasks [<xref ref-type="bibr" rid="ref2">2</xref>-<xref ref-type="bibr" rid="ref4">4</xref>]. Our best model (ie, RoBERTa-large) achieved a statistically significant higher performance than most of our other solutions (see <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>) but was not significantly better than the models XLNet-large (<italic>P</italic>=.07), BERT-large + RoBERTa-large (<italic>P</italic>=.13), and RoBERTa-large + XLNet-large (<italic>P</italic>=.06). The significance analysis indicated that these four models performed very similarly to each other.</p>
      </sec>
      <sec>
        <title>Error Analysis</title>
        <p>We compared the system prediction from our best model (ie, RoBERTa-large) with the gold standards and identified sentence pairs with the largest discrepancy in terms of the similarity score. Among the top 50 sentence pairs, 26 of them had labeled scores in the range of 0.0 to 1.0, and only 6 sentence pairs had gold standard STS scores over 3.0. We further split the testing results into two subsets using a threshold score of 2.5 on gold standards and calculated the mean and median of the differences between the gold standards and predictions. For the subgroup consisting of sentence pairs with gold standard scores over 2.5, the mean and median of difference were 0.46 and 0.37. For the other subset (difference≤2.5), the mean and median of difference were 0.69 and 0.66. Therefore, it was more challenging for the system to predict appropriate STS scores for sentence pairs with low similarity (gold standard score≤2.5) than for those with high similarity.</p>
        <p>We also observed that sentence pairs with high similarity scores usually have a similar sentence structure where many words occur in both sentences. Therefore, we hypothesized that the STS models will assign higher scores to sentence pairs that share a large portion of their lexicons and similar syntax. To test our hypothesis, we adopted the BertViz package [<xref ref-type="bibr" rid="ref47">47</xref>] to profile the attention pattern of the RoBERTa-large model (ie, our best STS model). BertViz can generate the attention pattern between two sentences by linking words via lines, where the line weights reflect the attention weights; higher line weights indicate higher attention weights between the two words. <xref ref-type="table" rid="table4">Table 4</xref> and <xref rid="figure3" ref-type="fig">Figure 3</xref> show an example for two sentence pairs on a similar topic from the training and test sets. In the first example from the training set, the attention pattern has three dominant attention weights (eg, “questions-questions”) and the similarity score for this sentence pair is labeled as 5.0. However, the attention pattern for the sentence pair from the test set also has similar dominant attention weights (such as “questions-questions”) but was labeled with a similarity score of 0.0.</p>
        <table-wrap position="float" id="table4">
          <label>Table 4</label>
          <caption>
            <p>Transformer model attention visualization on two examples from STS-Clinic.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="150"/>
            <col width="600"/>
            <col width="150"/>
            <col width="100"/>
            <thead>
              <tr valign="top">
                <td>Category</td>
                <td>Sentence pair</td>
                <td>Gold standard</td>
                <td>Prediction</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Training</td>
                <td>
                  <list list-type="bullet">
                    <list-item>
                      <p>S1<sup>a</sup>: advised to contact us with questions or concerns.</p>
                    </list-item>
                    <list-item>
                      <p>S2: please do not hesitate to contact me with any further questions.</p>
                    </list-item>
                  </list>
                </td>
                <td>5</td>
                <td>N/A<sup>b</sup></td>
              </tr>
              <tr valign="top">
                <td>Test</td>
                <td>
                  <list list-type="bullet">
                    <list-item>
                      <p>S1: patient discharged ambulatory without further questions or concerns noted.</p>
                    </list-item>
                    <list-item>
                      <p>S2: please contact location at phone number with any questions or concerns regarding this patient.</p>
                    </list-item>
                  </list>
                </td>
                <td>0</td>
                <td>2.5</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table4fn1">
              <p><sup>a</sup>S: sentence.</p>
            </fn>
            <fn id="table4fn2">
              <p><sup>b</sup>N/A: not applicable.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <fig id="figure3" position="float">
          <label>Figure 3</label>
          <caption>
            <p>Transformer model attention visualization on two examples from STS-Clinic. STS: semantic textual similarity.</p>
          </caption>
          <graphic xlink:href="medinform_v8i11e19735_fig3.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Limitations</title>
        <p>This study has limitations. First, it is worth exploring methods to effectively integrate clinical resources with general English resources in transformer-based models. In this study, we explored an approach by pretraining transformer-based models with a clinical corpus (ie, MIMIC-III corpus). However, our results showed that this approach was not efficient. Therefore, new strategies to better integrate medical resources are needed. Second, our clinical STS systems performed better for sentence pairs with high similarity scores (ie, similarity score≥3 in gold standard) whereas, for the sentence pairs with low similarity scores (ie, similarity score&#60;2 in gold standard), our systems still need to be improved. How to address this issue is one of our future focuses.</p>
      </sec>
      <sec>
        <title>Conclusions</title>
        <p>In this study, we demonstrated transformer-based models for measuring clinical STS and developed a system that can use various transformer algorithms. Our experiment results show that the RoBERTa model achieved the best performance compared to other transformer models. Our study demonstrated the efficiency of transformer-based models for assessing the semantic similarity for clinical text. Our models and system could be applied to various downstream clinical NLP applications. The source code, system, and pretrained models can be accessed on GitHub [<xref ref-type="bibr" rid="ref48">48</xref>].</p>
      </sec>
    </sec>
  </body>
  <back>
    <app-group>
      <supplementary-material id="app1">
        <label>Multimedia Appendix 1</label>
        <p>The 1-tailed parametric test results based on Fisher Z-transformation.</p>
        <media xlink:href="medinform_v8i11e19735_app1.xlsx" xlink:title="XLSX File  (Microsoft Excel File), 10 KB"/>
      </supplementary-material>
    </app-group>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">BERT</term>
          <def>
            <p>Bidirectional Encoder Representations from Transformers</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">MIMIC-III</term>
          <def>
            <p>Medical Information Mart for Intensive Care</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">MLM</term>
          <def>
            <p>masked language modeling</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">n2c2</term>
          <def>
            <p>National Natural Language Processing Clinical Challenges</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">NIH</term>
          <def>
            <p>National Institutes of Health</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">NLP</term>
          <def>
            <p>natural language processing</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb7">OHNLP</term>
          <def>
            <p>Open Health Natural Language Processing</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb8">RoBERTa</term>
          <def>
            <p>Robustly optimized BERT approach</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb9">SemEval</term>
          <def>
            <p>semantic evaluation</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb10">STS</term>
          <def>
            <p>semantic textual similarity</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <ack>
      <p>Research reported in this publication was supported by (1) the University of Florida Clinical and Translational Science Institute, which is supported in part by the National Institutes of Health (NIH) National Center for Advancing Translational Sciences under award number UL1TR001427; (2) the Patient-Centered Outcomes Research Institute under award number ME-2018C3-14754; (3) the Centers for Disease Control and Prevention under award number U18DP006512; (4) the NIH National Cancer Institute under award number R01CA246418; and (5) the NIH National Institute on Aging under award number R21AG061431-02S1. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health and Patient-Centered Outcomes Research Institute.</p>
      <p>We would like to thank the n2c2 organizers for providing the annotated corpus and the guidance for this challenge. We gratefully acknowledge the support of NVIDIA Corporation for the donation of the graphics processing units used for this research.</p>
    </ack>
    <fn-group>
      <fn fn-type="con">
        <p>XY, JB, and YW were responsible for the overall design, development, and evaluation of this study. YM, HZ, and XH were involved in conducting experiments and result analysis. XY, JB, and YW wrote and edited this manuscript. All authors reviewed the manuscript critically for scientific content, and all authors gave final approval of the manuscript for publication.</p>
      </fn>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Agirre</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Cer</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Diab</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Gonzalez-Agirre</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>SemEval-2012 Task 6: A Pilot on Semantic Textual Similarity</article-title>
          <year>2012</year>
          <conf-name>*SEM 2012: The First Joint Conference on Lexical and Computational Semantics</conf-name>
          <conf-date>June 7-8, 2012</conf-date>
          <conf-loc>Montréal, Canada</conf-loc>
          <publisher-name>Association for Computational Linguistics</publisher-name>
          <fpage>385</fpage>
          <lpage>393</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.aclweb.org/anthology/S12-1051"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Agirre</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Cer</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Diab</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Gonzalez-Agirre</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Guo</surname>
              <given-names>W</given-names>
            </name>
          </person-group>
          <article-title>*SEM 2013 shared taskmantic Textual Similarity</article-title>
          <year>2013</year>
          <conf-name>*SEM 2013: The Second Joint Conference on Lexical and Computational Semantics</conf-name>
          <conf-date>June 13-14, 2013</conf-date>
          <conf-loc>Atlanta, USA</conf-loc>
          <publisher-name>Association for Computational Linguistics</publisher-name>
          <fpage>32</fpage>
          <lpage>43</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.aclweb.org/anthology/S13-1004"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Agirre</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Banea</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Cardie</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Cer</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Diab</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Gonzalez-Agirre</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Guo</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Mihalcea</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Rigau</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Wiebe</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>SemEval-2014 Task 10: Multilingual Semantic Textual Similarity</article-title>
          <year>2014</year>
          <conf-name>The 8th International Workshop on Semantic Evaluation (SemEval 2014)</conf-name>
          <conf-date>Aug 23-24, 2014</conf-date>
          <conf-loc>Dublin, Ireland</conf-loc>
          <fpage>81</fpage>
          <lpage>91</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.aclweb.org/anthology/S14-2010"/>
          </comment>
          <pub-id pub-id-type="doi">10.3115/v1/s14-2010</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Agirre</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Banea</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Cardie</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Cer</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Diab</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Gonzalez-Agirre</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Guo</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Lopez-Gazpio</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Maritxalar</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Mihalcea</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Rigau</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Uria</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Wiebe</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>SemEval-2015 Task 2: Semantic Textual Similarity, English, Spanish and Pilot on Interpretability</article-title>
          <year>2015</year>
          <conf-name>The 9th International Workshop on Semantic Evaluation (SemEval 2015)</conf-name>
          <conf-date>June 4-5, 2015</conf-date>
          <conf-loc>Denver, USA</conf-loc>
          <fpage>252</fpage>
          <lpage>263</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.aclweb.org/anthology/S15-2045/"/>
          </comment>
          <pub-id pub-id-type="doi">10.18653/v1/s15-2045</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Agirre</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Banea</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Cer</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Diab</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Gonzalez-Agirre</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Mihalcea</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Rigau</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Wiebe</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>SemEval-2016 Task 1: Semantic Textual Similarity, Monolingual and Cross-Lingual Evaluation</article-title>
          <year>2016</year>
          <conf-name>The 10th International Workshop on Semantic Evaluation (SemEval-2016)</conf-name>
          <conf-date>June 16-17, 2016</conf-date>
          <conf-loc>San Diego, USA</conf-loc>
          <fpage>497</fpage>
          <lpage>511</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.aclweb.org/anthology/S16-1081/"/>
          </comment>
          <pub-id pub-id-type="doi">10.18653/v1/S16-1081</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Cer</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Diab</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Agirre</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Lopez-Gazpio</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Specia</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>SemEval-2017 Task 1: Semantic Textual Similarity Multilingual and Crosslingual Focused Evaluation</article-title>
          <year>2017</year>
          <conf-name>The 11th International Workshop on Semantic Evaluation (SemEval-2017)</conf-name>
          <conf-date>Aug 3-4, 2017</conf-date>
          <conf-loc>Vancouver, Canada</conf-loc>
          <fpage>1</fpage>
          <lpage>14</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.aclweb.org/anthology/S17-2001/"/>
          </comment>
          <pub-id pub-id-type="doi">10.18653/v1/s17-2001</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Farouk</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Measuring Sentences Similarity: A Survey</article-title>
          <source>ArXiv</source>
          <year>2019</year>
          <month>10</month>
          <day>06</day>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/1910.03940"/>
          </comment>
          <pub-id pub-id-type="doi">10.17485/ijst/2019/v12i25/143977</pub-id>
          <pub-id pub-id-type="pii">arXiv:1910.03940</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ramaprabha</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Das</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Mukerjee</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>Survey on Sentence Similarity Evaluation using Deep Learning</article-title>
          <source>J. Phys.: Conf. Ser</source>
          <year>2018</year>
          <month>04</month>
          <day>25</day>
          <conf-name>National Conference on Mathematical Techniques and its Applications (NCMTA 18)</conf-name>
          <conf-date>Jan 5–6, 2018</conf-date>
          <conf-loc>Kattankulathur, India</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://iopscience.iop.org/article/10.1088/1742-6596/1000/1/012070"/>
          </comment>
          <pub-id pub-id-type="doi">10.1088/1742-6596/1000/1/012070</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Gomaa</surname>
              <given-names>WH</given-names>
            </name>
            <name name-style="western">
              <surname>Fahmy</surname>
              <given-names>AA</given-names>
            </name>
          </person-group>
          <article-title>A Survey of Text Similarity Approaches</article-title>
          <source>IJCA</source>
          <year>2013</year>
          <month>04</month>
          <day>18</day>
          <volume>68</volume>
          <issue>13</issue>
          <fpage>13</fpage>
          <lpage>18</lpage>
          <pub-id pub-id-type="doi">10.5120/11638-7118</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Béchara</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Costa</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Taslimipoor</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Gupta</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Orasan</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Corpas</surname>
              <given-names>PG</given-names>
            </name>
            <name name-style="western">
              <surname>Mitkov</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>MiniExperts: An SVM Approach for Measuring Semantic Textual Similarity</article-title>
          <year>2015</year>
          <conf-name>The 9th International Workshop on Semantic Evaluation (SemEval 2015)</conf-name>
          <conf-date>Jun 4-5, 2015</conf-date>
          <conf-loc>Denver, USA</conf-loc>
          <publisher-loc>Proceedings of the 9th International Workshop on Semantic Evaluation (SemEval 2015) Denver, Colorado</publisher-loc>
          <publisher-name>Association for Computational Linguistics</publisher-name>
          <fpage>96</fpage>
          <lpage>101</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.aclweb.org/anthology/S15-2017/"/>
          </comment>
          <pub-id pub-id-type="doi">10.18653/v1/S15-2017</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Buscaldi</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Flores</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Ruiz</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Rodriguez</surname>
              <given-names>I</given-names>
            </name>
          </person-group>
          <article-title>SOPA: Random Forests Regression for the Semantic Textual Similarity task</article-title>
          <year>2015</year>
          <conf-name>The 9th International Workshop on Semantic Evaluation (SemEval 2015)</conf-name>
          <conf-date>Jun 4-5, 2015</conf-date>
          <conf-loc>Denver, USA</conf-loc>
          <fpage>132</fpage>
          <lpage>137</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.aclweb.org/anthology/S15-2024/"/>
          </comment>
          <pub-id pub-id-type="doi">10.18653/v1/s15-2024</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>He</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Gimpel</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Multi-perspective sentence similarity modeling with convolutional neural networks</article-title>
          <year>2015</year>
          <conf-name>The 2015 Conference on Empirical Methods in Natural Language Processing</conf-name>
          <conf-date>Sept 19 – 21, 2015</conf-date>
          <conf-loc>Lisbon, Portugal</conf-loc>
          <fpage>1576</fpage>
          <lpage>1586</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.aclweb.org/anthology/D15-1181/"/>
          </comment>
          <pub-id pub-id-type="doi">10.18653/v1/d15-1181</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Mueller</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Thyagarajan</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Siamese recurrent architectures for learning sentence similarity</article-title>
          <year>2016</year>
          <conf-name>The Thirtieth AAAI Conference on Artificial Intelligence</conf-name>
          <conf-date>Feb 12–17, 2016</conf-date>
          <conf-loc>Phoenix, USA</conf-loc>
          <publisher-name>AAAI Press</publisher-name>
          <fpage>2786</fpage>
          <lpage>2792</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://dl.acm.org/doi/10.5555/3016100.3016291"/>
          </comment>
          <pub-id pub-id-type="doi">10.5555/3016100.3016291</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kashyap</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Han</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Yus</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Sleeman</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Satyapanich</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Gandhi</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Finin</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Robust semantic text similarity using LSA, machine learning, and linguistic resources</article-title>
          <source>Lang Resources &#38; Evaluation</source>
          <year>2015</year>
          <month>10</month>
          <day>30</day>
          <volume>50</volume>
          <issue>1</issue>
          <fpage>125</fpage>
          <lpage>161</lpage>
          <pub-id pub-id-type="doi">10.1007/s10579-015-9319-2</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Niraula</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Banjade</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Ştefănescu</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Rus</surname>
              <given-names>V</given-names>
            </name>
          </person-group>
          <article-title>Experiments with Semantic Similarity Measures Based on LDA and LSA</article-title>
          <year>2013</year>
          <conf-name>The First International Conference on Statistical Language and Speech Processing (SLSP 2013)</conf-name>
          <conf-date>July 29-31, 2013</conf-date>
          <conf-loc>Tarragona, Spain</conf-loc>
          <fpage>188</fpage>
          <lpage>199</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://link.springer.com/chapter/10.1007/978-3-642-39593-2_17"/>
          </comment>
          <pub-id pub-id-type="doi">10.1007/978-3-642-39593-2_17</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Devlin</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Chang</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Toutanova</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Bert: Pre-training of deep bidirectional transformers for language understanding</article-title>
          <source>arXiv</source>
          <year>2018</year>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/1810.04805"/>
          </comment>
          <pub-id pub-id-type="pii">arXiv:1810.04805</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Dai</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Carbonell</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Salakhutdinov</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Le</surname>
              <given-names>Q</given-names>
            </name>
          </person-group>
          <article-title>XLNet: Generalized Autoregressive Pretraining for Language Understanding</article-title>
          <source>arXiv</source>
          <year>2019</year>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/1906.08237"/>
          </comment>
          <pub-id pub-id-type="pii">arXiv:1906.08237</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Ott</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Goyal</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Du</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Joshi</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Levy</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Lewis</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Zettlemoyer</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Stoyanov</surname>
              <given-names>V</given-names>
            </name>
          </person-group>
          <article-title>RoBERTa: A Robustly Optimized BERT Pretraining Approach</article-title>
          <source>arXiv</source>
          <year>2019</year>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/1907.11692"/>
          </comment>
          <pub-id pub-id-type="pii">arXiv:1907.11692</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Singh</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Michael</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Hill</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Levy</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Bowman</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding</article-title>
          <source>arXiv</source>
          <year>2019</year>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1804.07461"/>
          </comment>
          <pub-id pub-id-type="pii">arXiv:1804.07461</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Raffel</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Shazeer</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Roberts</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Narang</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Matena</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Zhou</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer</article-title>
          <source>arXiv</source>
          <year>2019</year>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1910.10683"/>
          </comment>
          <pub-id pub-id-type="pii">arXiv:1910.10683</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Bi</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Yan</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Bao</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Xia</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Peng</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Si</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>StructBERT: Incorporating Language Structures into Pre-training for Deep Language Understanding</article-title>
          <source>arXiv</source>
          <year>2019</year>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1908.04577"/>
          </comment>
          <pub-id pub-id-type="pii">arxiv:1908.04577</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Birkhead</surname>
              <given-names>GS</given-names>
            </name>
            <name name-style="western">
              <surname>Klompas</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Shah</surname>
              <given-names>NR</given-names>
            </name>
          </person-group>
          <article-title>Uses of electronic health records for public health surveillance to advance public health</article-title>
          <source>Annu Rev Public Health</source>
          <year>2015</year>
          <month>03</month>
          <day>18</day>
          <volume>36</volume>
          <fpage>345</fpage>
          <lpage>59</lpage>
          <pub-id pub-id-type="doi">10.1146/annurev-publhealth-031914-122747</pub-id>
          <pub-id pub-id-type="medline">25581157</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Obeid</surname>
              <given-names>JS</given-names>
            </name>
            <name name-style="western">
              <surname>Beskow</surname>
              <given-names>LM</given-names>
            </name>
            <name name-style="western">
              <surname>Rape</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Gouripeddi</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Black</surname>
              <given-names>RA</given-names>
            </name>
            <name name-style="western">
              <surname>Cimino</surname>
              <given-names>JJ</given-names>
            </name>
            <name name-style="western">
              <surname>Embi</surname>
              <given-names>PJ</given-names>
            </name>
            <name name-style="western">
              <surname>Weng</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Marnocha</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Buse</surname>
              <given-names>JB</given-names>
            </name>
          </person-group>
          <article-title>A survey of practices for the use of electronic health records to support research recruitment</article-title>
          <source>J Clin Transl Sci</source>
          <year>2017</year>
          <month>08</month>
          <volume>1</volume>
          <issue>4</issue>
          <fpage>246</fpage>
          <lpage>252</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/29657859"/>
          </comment>
          <pub-id pub-id-type="doi">10.1017/cts.2017.301</pub-id>
          <pub-id pub-id-type="medline">29657859</pub-id>
          <pub-id pub-id-type="pii">00301</pub-id>
          <pub-id pub-id-type="pmcid">PMC5890320</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Garza</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Del Fiol</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Tenenbaum</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Walden</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Zozus</surname>
              <given-names>MN</given-names>
            </name>
          </person-group>
          <article-title>Evaluating common data models for use with a longitudinal community registry</article-title>
          <source>J Biomed Inform</source>
          <year>2016</year>
          <month>12</month>
          <volume>64</volume>
          <fpage>333</fpage>
          <lpage>341</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S1532-0464(16)30153-8"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.jbi.2016.10.016</pub-id>
          <pub-id pub-id-type="medline">27989817</pub-id>
          <pub-id pub-id-type="pii">S1532-0464(16)30153-8</pub-id>
          <pub-id pub-id-type="pmcid">PMC6810649</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Rosenbloom</surname>
              <given-names>ST</given-names>
            </name>
            <name name-style="western">
              <surname>Denny</surname>
              <given-names>JC</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Lorenzi</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Stead</surname>
              <given-names>WW</given-names>
            </name>
            <name name-style="western">
              <surname>Johnson</surname>
              <given-names>KB</given-names>
            </name>
          </person-group>
          <article-title>Data from clinical notes: a perspective on the tension between structure and flexible documentation</article-title>
          <source>J Am Med Inform Assoc</source>
          <year>2011</year>
          <volume>18</volume>
          <issue>2</issue>
          <fpage>181</fpage>
          <lpage>6</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://jamia.oxfordjournals.org/cgi/pmidlookup?view=long&#38;pmid=21233086"/>
          </comment>
          <pub-id pub-id-type="doi">10.1136/jamia.2010.007237</pub-id>
          <pub-id pub-id-type="medline">21233086</pub-id>
          <pub-id pub-id-type="pii">jamia.2010.007237</pub-id>
          <pub-id pub-id-type="pmcid">PMC3116264</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Pakhomov</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>McInnes</surname>
              <given-names>BT</given-names>
            </name>
            <name name-style="western">
              <surname>Melton</surname>
              <given-names>GB</given-names>
            </name>
          </person-group>
          <article-title>Evaluating measures of redundancy in clinical texts</article-title>
          <source>AMIA Annu Symp Proc</source>
          <year>2011</year>
          <volume>2011</volume>
          <fpage>1612</fpage>
          <lpage>20</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/22195227"/>
          </comment>
          <pub-id pub-id-type="medline">22195227</pub-id>
          <pub-id pub-id-type="pmcid">PMC3243221</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref27">
        <label>27</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>MD</given-names>
            </name>
            <name name-style="western">
              <surname>Khanna</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Najafi</surname>
              <given-names>N</given-names>
            </name>
          </person-group>
          <article-title>Characterizing the Source of Text in Electronic Health Record Progress Notes</article-title>
          <source>JAMA Intern Med</source>
          <year>2017</year>
          <month>08</month>
          <day>01</day>
          <volume>177</volume>
          <issue>8</issue>
          <fpage>1212</fpage>
          <lpage>1213</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/28558106"/>
          </comment>
          <pub-id pub-id-type="doi">10.1001/jamainternmed.2017.1548</pub-id>
          <pub-id pub-id-type="medline">28558106</pub-id>
          <pub-id pub-id-type="pii">2629493</pub-id>
          <pub-id pub-id-type="pmcid">PMC5818790</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref28">
        <label>28</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Botsis</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Hartvigsen</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Weng</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Secondary Use of EHR: Data Quality Issues and Informatics Opportunities</article-title>
          <source>AMIA Jt Summits Transl Sci Proc</source>
          <year>2010</year>
          <month>03</month>
          <day>01</day>
          <volume>2010</volume>
          <fpage>1</fpage>
          <lpage>5</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/21347133"/>
          </comment>
          <pub-id pub-id-type="medline">21347133</pub-id>
          <pub-id pub-id-type="pmcid">PMC3041534</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref29">
        <label>29</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sogancioglu</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Öztürk</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Özgür</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>BIOSSES: a semantic sentence similarity estimation system for the biomedical domain</article-title>
          <source>Bioinformatics</source>
          <year>2017</year>
          <month>07</month>
          <day>15</day>
          <volume>33</volume>
          <issue>14</issue>
          <fpage>i49</fpage>
          <lpage>i58</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/28881973"/>
          </comment>
          <pub-id pub-id-type="doi">10.1093/bioinformatics/btx238</pub-id>
          <pub-id pub-id-type="medline">28881973</pub-id>
          <pub-id pub-id-type="pii">3953954</pub-id>
          <pub-id pub-id-type="pmcid">PMC5870675</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref30">
        <label>30</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Xiong</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Qin</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Cao</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Shen</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Yan</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Tang</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>Distributed representation and one-hot representation fusion with gated network for clinical semantic textual similarity</article-title>
          <source>BMC Med Inform Decis Mak</source>
          <year>2020</year>
          <month>04</month>
          <day>30</day>
          <volume>20</volume>
          <issue>Suppl 1</issue>
          <fpage>72</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://bmcmedinformdecismak.biomedcentral.com/articles/10.1186/s12911-020-1045-z"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/s12911-020-1045-z</pub-id>
          <pub-id pub-id-type="medline">32349764</pub-id>
          <pub-id pub-id-type="pii">10.1186/s12911-020-1045-z</pub-id>
          <pub-id pub-id-type="pmcid">PMC7191689</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref31">
        <label>31</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Du</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Wilbur</surname>
              <given-names>WJ</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>Z</given-names>
            </name>
          </person-group>
          <article-title>Deep learning with sentence embeddings pre-trained on biomedical corpora improves the performance of finding similar sentences in electronic medical records</article-title>
          <source>BMC Med Inform Decis Mak</source>
          <year>2020</year>
          <month>04</month>
          <day>30</day>
          <volume>20</volume>
          <issue>Suppl 1</issue>
          <fpage>73</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://bmcmedinformdecismak.biomedcentral.com/articles/10.1186/s12911-020-1044-0"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/s12911-020-1044-0</pub-id>
          <pub-id pub-id-type="medline">32349758</pub-id>
          <pub-id pub-id-type="pii">10.1186/s12911-020-1044-0</pub-id>
          <pub-id pub-id-type="pmcid">PMC7191680</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref32">
        <label>32</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Afzal</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Fu</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Shen</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Rastegar-Mojarad</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>MedSTS: a resource for clinical semantic textual similarity</article-title>
          <source>Lang Resources &#38; Evaluation</source>
          <year>2018</year>
          <month>10</month>
          <day>24</day>
          <volume>54</volume>
          <issue>1</issue>
          <fpage>57</fpage>
          <lpage>72</lpage>
          <pub-id pub-id-type="doi">10.1007/s10579-018-9431-1</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref33">
        <label>33</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Rastegar-Mojarad</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Afzal</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Shen</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Fu</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>BioCreative/OHNLP Challenge 2018</article-title>
          <year>2018</year>
          <conf-name>The 9th ACM Conference on Bioinformatics, Computational Biology, and Health Informatics (ACM BCB)</conf-name>
          <conf-date>Aug 29 - Sept 1, 2018</conf-date>
          <conf-loc>Washington DC, USA</conf-loc>
          <fpage>575</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1145/3233547.3233672"/>
          </comment>
          <pub-id pub-id-type="doi">10.1145/3233547.3233672</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref34">
        <label>34</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Fu</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Shen</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Henry</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Uzuner</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>Overview of the 2019 n2c2/OHNLP Track on Clinical Semantic Textual Similarity</article-title>
          <source>JMIR Medical Informatics</source>
          <year>2020</year>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://preprints.jmir.org/preprint/23375"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/23375</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref35">
        <label>35</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="editor">
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Ma</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Ensemble Learning</article-title>
          <source>Ensemble machine learning: methods and applications</source>
          <year>2012</year>
          <publisher-loc>New York, USA</publisher-loc>
          <publisher-name>Springer</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref36">
        <label>36</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wolf</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Debut</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Sanh</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Chaumond</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Delangue</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Moi</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Cistac</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Rault</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Louf</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Funtowicz</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Brew</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>HuggingFace's Transformers: State-of-the-art Natural Language Processing</article-title>
          <source>arXiv</source>
          <year>2019</year>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/1910.03771"/>
          </comment>
          <pub-id pub-id-type="pii">arXiv:1910.03771</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref37">
        <label>37</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Johnson</surname>
              <given-names>AEW</given-names>
            </name>
            <name name-style="western">
              <surname>Pollard</surname>
              <given-names>TJ</given-names>
            </name>
            <name name-style="western">
              <surname>Shen</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Lehman</surname>
              <given-names>LH</given-names>
            </name>
            <name name-style="western">
              <surname>Feng</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Ghassemi</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Moody</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Szolovits</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Celi</surname>
              <given-names>LA</given-names>
            </name>
            <name name-style="western">
              <surname>Mark</surname>
              <given-names>RG</given-names>
            </name>
          </person-group>
          <article-title>MIMIC-III, a freely accessible critical care database</article-title>
          <source>Sci Data</source>
          <year>2016</year>
          <volume>3</volume>
          <fpage>160035</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/27219127"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/sdata.2016.35</pub-id>
          <pub-id pub-id-type="medline">27219127</pub-id>
          <pub-id pub-id-type="pii">sdata201635</pub-id>
          <pub-id pub-id-type="pmcid">PMC4878278</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref38">
        <label>38</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Virtanen</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Gommers</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Oliphant</surname>
              <given-names>TE</given-names>
            </name>
            <name name-style="western">
              <surname>Haberland</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Reddy</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Cournapeau</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Burovski</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Peterson</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Weckesser</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Bright</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>van der Walt</surname>
              <given-names>SJ</given-names>
            </name>
            <name name-style="western">
              <surname>Brett</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Wilson</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Millman</surname>
              <given-names>KJ</given-names>
            </name>
            <name name-style="western">
              <surname>Mayorov</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Nelson</surname>
              <given-names>ARJ</given-names>
            </name>
            <name name-style="western">
              <surname>Jones</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Kern</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Larson</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Carey</surname>
              <given-names>CJ</given-names>
            </name>
            <name name-style="western">
              <surname>Polat</surname>
              <given-names>İ</given-names>
            </name>
            <name name-style="western">
              <surname>Feng</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Moore</surname>
              <given-names>EW</given-names>
            </name>
            <name name-style="western">
              <surname>VanderPlas</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Laxalde</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Perktold</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Cimrman</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Henriksen</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Quintero</surname>
              <given-names>EA</given-names>
            </name>
            <name name-style="western">
              <surname>Harris</surname>
              <given-names>CR</given-names>
            </name>
            <name name-style="western">
              <surname>Archibald</surname>
              <given-names>AM</given-names>
            </name>
            <name name-style="western">
              <surname>Ribeiro</surname>
              <given-names>AH</given-names>
            </name>
            <name name-style="western">
              <surname>Pedregosa</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>van Mulbregt</surname>
              <given-names>P</given-names>
            </name>
            <collab>SciPy 1.0 Contributors</collab>
          </person-group>
          <article-title>SciPy 1.0: fundamental algorithms for scientific computing in Python</article-title>
          <source>Nat Methods</source>
          <year>2020</year>
          <month>03</month>
          <volume>17</volume>
          <issue>3</issue>
          <fpage>261</fpage>
          <lpage>272</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/32015543"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41592-019-0686-2</pub-id>
          <pub-id pub-id-type="medline">32015543</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41592-019-0686-2</pub-id>
          <pub-id pub-id-type="pmcid">PMC7056644</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref39">
        <label>39</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Murphy</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <source>Machine Learning: A Probabilistic Perspective</source>
          <year>2012</year>
          <publisher-loc>Cambridge, USA</publisher-loc>
          <publisher-name>MIT Press</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref40">
        <label>40</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Yoon</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>So</surname>
              <given-names>CH</given-names>
            </name>
            <name name-style="western">
              <surname>Kang</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>BioBERT: a pre-trained biomedical language representation model for biomedical text mining</article-title>
          <source>Bioinformatics</source>
          <year>2020</year>
          <month>02</month>
          <day>15</day>
          <volume>36</volume>
          <issue>4</issue>
          <fpage>1234</fpage>
          <lpage>1240</lpage>
          <pub-id pub-id-type="doi">10.1093/bioinformatics/btz682</pub-id>
          <pub-id pub-id-type="medline">31501885</pub-id>
          <pub-id pub-id-type="pii">5566506</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref41">
        <label>41</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Huang</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Altosaar</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Ranganath</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>ClinicalBERT: Modeling Clinical Notes and Predicting Hospital Readmission</article-title>
          <source>arXiv</source>
          <year>2019</year>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/1904.05342"/>
          </comment>
          <pub-id pub-id-type="pii">arXiv:1904.05342</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref42">
        <label>42</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Alsentzer</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Murphy</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Boag</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Weng</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Jin</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Naumann</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>McDermott</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Publicly Available Clinical BERT Embeddings</article-title>
          <source>arXiv</source>
          <year>2019</year>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1904.03323"/>
          </comment>
          <pub-id pub-id-type="pii">arXiv:1904.03323</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref43">
        <label>43</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Peng</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Yan</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>Z</given-names>
            </name>
          </person-group>
          <article-title>Transfer Learning in Biomedical Natural Language Processing: An Evaluation of BERT and ELMo on Ten Benchmarking Datasets</article-title>
          <year>2019</year>
          <conf-name>The 18th BioNLP Workshop and Shared Task (BioBLP 2019)</conf-name>
          <conf-date>Aug 1, 2019</conf-date>
          <conf-loc>Florence, Italy</conf-loc>
          <fpage>58</fpage>
          <lpage>65</lpage>
          <pub-id pub-id-type="doi">10.18653/v1/w19-5006</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref44">
        <label>44</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Si</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Roberts</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Enhancing clinical concept extraction with contextual embeddings</article-title>
          <source>J Am Med Inform Assoc</source>
          <year>2019</year>
          <month>11</month>
          <day>01</day>
          <volume>26</volume>
          <issue>11</issue>
          <fpage>1297</fpage>
          <lpage>1304</lpage>
          <pub-id pub-id-type="doi">10.1093/jamia/ocz096</pub-id>
          <pub-id pub-id-type="medline">31265066</pub-id>
          <pub-id pub-id-type="pii">5527248</pub-id>
          <pub-id pub-id-type="pmcid">PMC6798561</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref45">
        <label>45</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tawfik</surname>
              <given-names>NS</given-names>
            </name>
            <name name-style="western">
              <surname>Spruit</surname>
              <given-names>MR</given-names>
            </name>
          </person-group>
          <article-title>Evaluating sentence representations for biomedical text: Methods and experimental results</article-title>
          <source>J Biomed Inform</source>
          <year>2020</year>
          <month>04</month>
          <volume>104</volume>
          <fpage>103396</fpage>
          <pub-id pub-id-type="doi">10.1016/j.jbi.2020.103396</pub-id>
          <pub-id pub-id-type="medline">32147441</pub-id>
          <pub-id pub-id-type="pii">S1532-0464(20)30025-3</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref46">
        <label>46</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Fisher</surname>
              <given-names>RA</given-names>
            </name>
          </person-group>
          <article-title>Frequency Distribution of the Values of the Correlation Coefficient in Samples from an Indefinitely Large Population</article-title>
          <source>Biometrika</source>
          <year>1915</year>
          <month>05</month>
          <volume>10</volume>
          <issue>4</issue>
          <fpage>507</fpage>
          <pub-id pub-id-type="doi">10.2307/2331838</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref47">
        <label>47</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Vig</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>A Multiscale Visualization of Attention in the Transformer Model</article-title>
          <source>arXiv</source>
          <year>2019</year>
          <month>06</month>
          <day>12</day>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1906.05714"/>
          </comment>
          <pub-id pub-id-type="doi">10.18653/v1/p19-3007</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref48">
        <label>48</label>
        <nlm-citation citation-type="web">
          <article-title>2019 N2C2 Track-1 Clinical Semantic Textual Similarity</article-title>
          <source>GitHub</source>
          <access-date>2020-11-02</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://github.com/uf-hobi-informatics-lab/2019_N2C2_Track1_ClinicalSTS.git">https://github.com/uf-hobi-informatics-lab/2019_N2C2_Track1_ClinicalSTS.git</ext-link>
          </comment>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
