<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMI</journal-id>
      <journal-id journal-id-type="nlm-ta">JMIR Med Inform</journal-id>
      <journal-title>JMIR Medical Informatics</journal-title>
      <issn pub-type="epub">2291-9694</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v10i2e32875</article-id>
      <article-id pub-id-type="pmid">35142635</article-id>
      <article-id pub-id-type="doi">10.2196/32875</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Viewpoint</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Viewpoint</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>Operationalizing and Implementing Pretrained, Large Artificial Intelligence Linguistic Models in the US Health Care System: Outlook of Generative Pretrained Transformer 3 (GPT-3) as a Service Model</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Lovis</surname>
            <given-names>Christian</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Trojan</surname>
            <given-names>Andreas</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Caze</surname>
            <given-names>Todd</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Carot-Sans</surname>
            <given-names>Gerard</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Ayre</surname>
            <given-names>Julie</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Zhang</surname>
            <given-names>Wei</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author" corresp="yes" equal-contrib="yes">
          <name name-style="western">
            <surname>Sezgin</surname>
            <given-names>Emre</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <address>
            <institution>The Abigail Wexner Research Institute</institution>
            <institution>Nationwide Children's Hospital</institution>
            <addr-line>700 Children's Drive</addr-line>
            <addr-line>Columbus, OH, 43205</addr-line>
            <country>United States</country>
            <phone>1 6143556814</phone>
            <email>esezgin1@gmail.com</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-8798-9605</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Sirrianni</surname>
            <given-names>Joseph</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-2952-5818</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author">
          <name name-style="western">
            <surname>Linwood</surname>
            <given-names>Simon L</given-names>
          </name>
          <degrees>MBA, MD</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-2876-2042</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>The Abigail Wexner Research Institute</institution>
        <institution>Nationwide Children's Hospital</institution>
        <addr-line>Columbus, OH</addr-line>
        <country>United States</country>
      </aff>
      <aff id="aff2">
        <label>2</label>
        <institution>School of Medicine</institution>
        <institution>University of California Riverside</institution>
        <addr-line>Riverside, CA</addr-line>
        <country>United States</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Emre Sezgin <email>esezgin1@gmail.com</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <month>2</month>
        <year>2022</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>10</day>
        <month>2</month>
        <year>2022</year>
      </pub-date>
      <volume>10</volume>
      <issue>2</issue>
      <elocation-id>e32875</elocation-id>
      <history>
        <date date-type="received">
          <day>12</day>
          <month>8</month>
          <year>2021</year>
        </date>
        <date date-type="rev-request">
          <day>24</day>
          <month>10</month>
          <year>2021</year>
        </date>
        <date date-type="rev-recd">
          <day>14</day>
          <month>12</month>
          <year>2021</year>
        </date>
        <date date-type="accepted">
          <day>9</day>
          <month>1</month>
          <year>2022</year>
        </date>
      </history>
      <copyright-statement>©Emre Sezgin, Joseph Sirrianni, Simon L Linwood. Originally published in JMIR Medical Informatics (https://medinform.jmir.org), 10.02.2022.</copyright-statement>
      <copyright-year>2022</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on https://medinform.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://medinform.jmir.org/2022/2/e32875" xlink:type="simple"/>
      <abstract>
        <p>Generative pretrained transformer models have been popular recently due to their enhanced capabilities and performance. In contrast to many existing artificial intelligence models, generative pretrained transformer models can perform with very limited training data. Generative pretrained transformer 3 (GPT-3) is one of the latest releases in this pipeline, demonstrating human-like logical and intellectual responses to prompts. Some examples include writing essays, answering complex questions, matching pronouns to their nouns, and conducting sentiment analyses. However, questions remain with regard to its implementation in health care, specifically in terms of operationalization and its use in clinical practice and research. In this viewpoint paper, we briefly introduce GPT-3 and its capabilities and outline considerations for its implementation and operationalization in clinical practice through a use case. The implementation considerations include (1) processing needs and information systems infrastructure, (2) operating costs, (3) model biases, and (4) evaluation metrics. In addition, we outline the following three major operational factors that drive the adoption of GPT-3 in the US health care system: (1) ensuring Health Insurance Portability and Accountability Act compliance, (2) building trust with health care providers, and (3) establishing broader access to the GPT-3 tools. This viewpoint can inform health care practitioners, developers, clinicians, and decision makers toward understanding the use of the powerful artificial intelligence tools integrated into hospital systems and health care.</p>
      </abstract>
      <kwd-group>
        <kwd>natural language processing</kwd>
        <kwd>artificial intelligence</kwd>
        <kwd>generative pretrained transformer</kwd>
        <kwd>clinical informatics</kwd>
        <kwd>chatbot</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <p>In 2020, OpenAI unveiled their third-generation language generation model, which is known as the generative pretrained transformer 3 (GPT-3) model [<xref ref-type="bibr" rid="ref1">1</xref>]. This model was the latest in a line of large pretrained models designed for understanding and producing natural language by using the transformer architecture, which was published only 3 years prior and significantly improved natural language understanding task performance over that of models built on prior architectures [<xref ref-type="bibr" rid="ref2">2</xref>]. However, GPT-3’s development was remarkable because it resulted in a substantial increase in the model’s size; it increased by more than 10-fold in 1 year, reaching 175 billion weights [<xref ref-type="bibr" rid="ref1">1</xref>-<xref ref-type="bibr" rid="ref3">3</xref>]. GPT-3’s increased model size makes it substantially more powerful than prior models; propels its language capabilities to near–human-like levels; and, in some cases, makes it the superior option for several language understanding tasks [<xref ref-type="bibr" rid="ref1">1</xref>].</p>
      <p>Ordinarily, deep learning tasks require large amounts of labeled training data. This requirement usually limits the tasks to which deep learning can be effectively applied. However, with its increased model size, GPT-3 has an enhanced capability for so-called <italic>few-shot</italic>, <italic>one-shot</italic>, and <italic>zero-shot</italic> <italic>learning</italic> when compared to prior models [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref4">4</xref>]. These learning methods involve training a model on significantly smaller amounts of training data. In these methods, the models are given a description of the task and, if applicable, a handful of examples to learn from, with few-shot training on only hundreds to thousands of instances, one-shot training on only 1 example, and zero-shot training on only the task description.</p>
      <p>GPT-3 was designed as a language generation model, focusing on producing appropriate text responses to an input. Although it can be adapted to address more traditional machine learning tasks, such as answering yes-no questions, matching pronouns to their nouns, and conducting sentiment analyses [<xref ref-type="bibr" rid="ref1">1</xref>], GPT-3’s text generation capabilities have attracted much attention as a potential solution for a variety of problems, such as creating enhanced chatbots [<xref ref-type="bibr" rid="ref5">5</xref>], answering complex questions, generating computer code from a design specification [<xref ref-type="bibr" rid="ref6">6</xref>], and writing news articles [<xref ref-type="bibr" rid="ref7">7</xref>]. As such, there is much research on and commercial interest in using GPT-3 for a variety of text generation applications. <xref ref-type="boxed-text" rid="box1">Textbox 1</xref> illustrates the power of GPT-3 in 3 question and answer examples [<xref ref-type="bibr" rid="ref8">8</xref>]. In 3 cases, GPT-3 provided human-like logical and intellectual responses. However, applying GPT-3 in health care is still an emerging topic.</p>
      <p>A plethora of research articles have outlined the potential opportunities for integrating artificial intelligence (AI) and machine learning models in a variety of hospital and health care settings [<xref ref-type="bibr" rid="ref9">9</xref>-<xref ref-type="bibr" rid="ref11">11</xref>]. However, the authors also quickly pointed out the many development and implementation challenges to deploying such solutions, such as the lack of transparency [<xref ref-type="bibr" rid="ref11">11</xref>], potential model biases [<xref ref-type="bibr" rid="ref9">9</xref>], and financing [<xref ref-type="bibr" rid="ref10">10</xref>]. A successful deployment of GPT-3 would need to overcome all of these standard barriers to AI implementation alongside the additional challenges that GPT-3 itself introduces [<xref ref-type="bibr" rid="ref12">12</xref>]. In this viewpoint, we introduce the use of GPT-3 in the health care setting with a use case and discuss implementation and operational considerations for its use in clinical practice.</p>
      <boxed-text id="box1" position="float">
        <title>Generative pretrained transformer 3 (GPT-3) question and answer examples.</title>
        <p>
          <bold>Question and answer</bold>
        </p>
        <list list-type="bullet">
          <list-item>
            <p>Human-generated question: “What is human life expectancy in the United States?”</p>
          </list-item>
          <list-item>
            <p>GPT-3–generated response: “Human life expectancy in the United States is 78 years.”</p>
          </list-item>
        </list>
        <p>
          <bold>Text summary</bold>
        </p>
        <list list-type="bullet">
          <list-item>
            <p>Human-generated prompt: “Jupiter is the fifth planet from the Sun and the largest in the Solar System. It is a gas giant with a mass one-thousandth that of the Sun, but two-and-a-half times that of all the other planets in the Solar System combined. Jupiter is one of the brightest objects visible to the naked eye in the night sky, and has been known to ancient civilizations since before recorded history. It is named after the Roman god Jupiter. When viewed from Earth, Jupiter can be bright enough for its reflected light to cast visible shadows, and is on average the third-brightest natural object in the night sky after the Moon and Venus.”</p>
          </list-item>
          <list-item>
            <p>GPT-3–generated response (a second grader’s language): “Jupiter is a huge planet made of gas. It is the fifth planet from the sun. It is very big and bright. It is the third brightest object in the night sky after the moon and Venus. Jupiter has many moons. One of the moons is bigger than the planet Mercury.”</p>
          </list-item>
        </list>
        <p>
          <bold>A chatbot (sarcastic)</bold>
        </p>
        <list list-type="bullet">
          <list-item>
            <p>Human: “Why is the sky blue?”</p>
          </list-item>
          <list-item>
            <p>GPT-3 chatbot: “I’m not sure. I’ll ask my friend Google.”</p>
          </list-item>
        </list>
      </boxed-text>
    </sec>
    <sec>
      <title>A Use Case: GPT-3 in Health Care</title>
      <p>The power of meaningful text generation by GPT-3 makes it an ideal tool for human-machine conversations, especially those conducted via chatbots. In our case, we illustrate the use of GPT-3 within a hospital network. In this hypothetical example, the hospital is providing a chatbot triaging mechanism for incoming patients to reduce the overhead at clinics and increase the safety and quality of care during the COVID-19 pandemic. The chatbot has to be connected to the hospital network, combined with a triage text summary service that is to be reviewed, and stored in the electronic health record (EHR; <xref rid="figure1" ref-type="fig">Figure 1</xref>). Putting aside the front-end details in this workflow (<xref rid="figure1" ref-type="fig">Figure 1</xref>), this use case outlines a typical implementation of GPT-3 as a service within a health system.</p>
      <fig id="figure1" position="float">
        <label>Figure 1</label>
        <caption>
          <p>GPT-3 use case (chatbot triaging and patient note summarization). API: application programming interface; EHR: electronic health record; GPT-3: generative pretrained transformer 3.</p>
        </caption>
        <graphic xlink:href="medinform_v10i2e32875_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
      </fig>
      <p>In this example, triage could be initiated by a patient or a hospital to conduct a health screening. The front-end application is operationalized through a chatbot mechanism over a front-end application, which could be a patient portal app, voice assistant, phone call, or SMS text messaging. Once a connection is established, the hospital system formulates GPT-3 requests by gathering patient health information and formatting this information to be interpretable with the GPT-3 model. Within the secure hospital network, GPT-3 is located outside of the EHR and provided as the “GPT-3-as-a-Service” platform. The application programming interface enables interoperability and acts as a gatekeeper for the data transfer of requests and responses. Once a request is received, the “GPT-3-as-a-Service” platform preprocesses the data and requests, allocates the tasks to be completed, produces outputs in an interpretable format, and sends the outputs to users. The type of tasks allocated depends on the requests, which, in our case, are question answering, text generation or culturally appropriate language translation, and text summarization. The response is sent back to the EHR system and then to the front-end application. At the end of triage, similar to the after-visit summary, the conversation text is summarized. To reduce the additional clinical burden of reading the whole conversation, GPT-3 summarizes the text (similar to a digital scriber) and stores it in the patient's health records. To avoid or address potential biases [<xref ref-type="bibr" rid="ref12">12</xref>], correct errors, and increase the control over patient data use and the model, the human-in-the-loop model [<xref ref-type="bibr" rid="ref13">13</xref>] can be implemented by using a report back mechanism at the front end, or the clinical team can be given oversight of GPT-3 integrated process in the hospital EHR system at the back end. Furthermore, the error corrections and adjustments in the text can be used to fine-tune the GPT-3 model to increase its accuracy and effectiveness.</p>
      <p>To be able to execute this use case in a real-world setting, health care practitioners and decision makers should consider and address the following operational and implementation challenges.</p>
    </sec>
    <sec>
      <title>Implementation Considerations</title>
      <sec>
        <title>Processing Needs and Information Systems Infrastructure</title>
        <p>Unlike more traditional AI models, GPT-3 is considerably larger in terms of memory requirements and is more computationally intensive. Specialized hardware for model training and execution—either graphics processing units or tensor processing units—is required for a scalable implementation. For any hospital system, additional investments for infrastructure to compensate for processing needs could be required.</p>
        <p>Given its size, dependencies, and hardware requirements, a GPT-3 solution would likely need to be run as a service. For this service, hospital systems would need to submit a service request to the GPT-3 solution service, which would process the request and return its results back to the hospital system. The hospital local network in <xref rid="figure1" ref-type="fig">Figure 1</xref> shows a sample workflow diagram for such an implementation. Such a setup would require diligent and significant provisioning, networking, and monitoring to ensure that the services are accessible and provide meaningful value.</p>
      </sec>
      <sec>
        <title>Operating Cost</title>
        <p>Given the current state of hospital networks and EHR systems, the integration of GPT-3 solutions would require complex systems and high technical knowledge for effective deployment and be costly to operationalize. One possible solution to ease the burden of GPT-3 deployments is integration with cloud computing platforms within hospital systems. Many cloud computing providers offer the specialized hardware needed to run such models and can easily handle off-the-shelf networking and dynamic load balancing. This would ease the burden of the major components of GPT-3 deployment; however, outsourcing cloud computing platforms can potentially increase the operating cost.</p>
      </sec>
      <sec>
        <title>Model Bias</title>
        <p>Several sources of bias can manifest themselves in a GPT-3–powered solution at different levels. At a model level, GPT-3 is trained on a large data set that has many problematic characteristics related to racial and sexist stereotypes, and as a result, the model learns certain biases against marginalized identities [<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref15">15</xref>]. These biases, which are present in GPT-3, can be harmful in clinical settings. Korngiebel and Mooney [<xref ref-type="bibr" rid="ref12">12</xref>] highlight the risks of using GPT-3 in health care delivery, noting specific examples where GPT-3 parrots extremist language from the internet [<xref ref-type="bibr" rid="ref16">16</xref>] and affirms suicidal ideation [<xref ref-type="bibr" rid="ref17">17</xref>].</p>
        <p>Aside from the inherent bias of GPT-3’s initial training, fine-tuning on medical data could also introduce the unintentional biases present in historic medical data. Practical biases, such as the undertesting of marginalized subpopulations, can influence underlying clinical data and introduce bias during the training of predictive models [<xref ref-type="bibr" rid="ref9">9</xref>]. Additionally, the implicit biases of health care professionals can influence diagnoses and treatments and are reflected in clinical notes [<xref ref-type="bibr" rid="ref18">18</xref>], which, if used to fine-tune GPT-3, would potentially affect the developed model.</p>
        <p>Given these biases, it would be unwise to deploy GPT-3 or any other sizable language model without active bias testing [<xref ref-type="bibr" rid="ref15">15</xref>]. Explicit procedures should be put in place to monitor, report, and react to potential biases produced by GPT-3 predictions. These mechanisms would ensure that GPT-3 can be used effectively without introducing harm to the patient. In our use case (<xref rid="figure1" ref-type="fig">Figure 1</xref>), we also added a human-in-the-loop mechanism, which can mandate the control, assessment, and training protocols and yield interpretable and manageable results.</p>
      </sec>
      <sec>
        <title>Evaluation Metrics</title>
        <p>Aside from physical implementation, there are methodological considerations for deploying GPT-3. As Watson et al [<xref ref-type="bibr" rid="ref10">10</xref>] notes in their investigation of model deployment in academic medical centers, clinical utility is a major concern for institutions. Understanding the best way to receive and interpret model results is imperative for a successful deployment, and ideally, model performance should be tracked and assessed by using evaluation methodologies and frameworks.</p>
        <p>The evaluation of text generation tasks, that is, those that GPT-3 is designed to address, is notoriously difficult. Standard metrics, such as prediction sensitivity and positive predictive value, do not cleanly reflect correctness in text generation, as ideas can be expressed in many ways in text. More specialized text generation metrics, such as BLEU (Bilingual Evaluation Understudy) [<xref ref-type="bibr" rid="ref19">19</xref>] and METEOR (Metric for Evaluation of Translation with Explicit Ordering) [<xref ref-type="bibr" rid="ref20">20</xref>], try to account for text variation but still only examine text at a word level without capturing the fundamental meaning. Methods that do try to incorporate the meaning of text in text evaluation rely on other black-box deep learning models to produce a value [<xref ref-type="bibr" rid="ref21">21</xref>]. Relying on a black-box evaluation method to evaluate a black-box model does not increase interpretability. Such a method would only result in lower trust overall and thus decrease the likelihood of the model being deployed.</p>
        <p>Health care–specific evaluation methods and frameworks for text generation tasks are therefore needed. The development of more robust methodologies for evaluating text generation tasks in the health care domain is required before the significant adoption of GPT-3 technology can be achieved. It is imperative that data scientists, informaticists, developers, clinicians, and health care practitioners collaborate in the development of evaluation measures to ensure a successful implementation of GPT-3.</p>
      </sec>
    </sec>
    <sec>
      <title>Operational Considerations: Compliance, Trust, and Access</title>
      <p>In addition to implementation, there are 3 major operational factors driving the adoption of GPT-3 in health care, as follows: (1) GPT-3 needs to work in compliance with the Health Insurance Portability and Accountability Act (HIPAA), (2) technology providers need to earn trust from health care providers, and (3) technology providers should improve access to the tool (<xref rid="figure2" ref-type="fig">Figure 2</xref>).</p>
      <p>Similar to GPT-3, there was huge enthusiasm to use the Amazon Alexa (Amazon.com Inc) voice assistant in health care delivery when it was released in 2014. However, at the time, Alexa was not yet legally able to store or transmit private health information. It took Amazon 5 years to become HIPAA compliant and to be able to sign business associate agreements with health care providers [<xref ref-type="bibr" rid="ref22">22</xref>]. A limited number of Alexa skills was released, and there is still a long list of other Alexa skills waiting to become HIPAA compliant. This example shows the slow progress of legislation changes and regulation updates for including new technologies in health care, suggesting that efforts should be put forward as early as possible for GPT-3. Without HIPAA compliance, the adoption of GPT-3 in health care can be a false start [<xref ref-type="bibr" rid="ref23">23</xref>]. However, although HIPAA compliance may not be immediate, it may be gradually progressing. GPT-3 is a black-box model, which complicates the HIPAA compliance process because unlike with other types of programmatic solutions, it is harder to decipher how data are processed internally by the model itself. However, assuming that GPT-3 will be deployable in the future, operations will start with implementing the limited capabilities of GPT-3 (ie, storing and transmitting data, running behind the firewalls of specific hardware [security rules], and analyzing a specific data set or patient cohort [privacy rules]). In parallel, further practices are needed to optimize the payment models for accommodating GPT-3 and seek opportunities for satisfying the US Food and Drug Administration’s requirements for software as a medical device [<xref ref-type="bibr" rid="ref24">24</xref>] with regard to using AI in clinical applications.</p>
      <p>In addition to legal requirements, trust must be established among patients, health care providers, and technology companies to adopt GPT-3 [<xref ref-type="bibr" rid="ref25">25</xref>]. It is common for technology companies to claim the right that they can use their customers’ data to further improve their services or achieve additional commercial value. Additionally, the culture of skepticism toward AI among clinicians can place a heavy burden on model interpretability and result in lower trust in clinical care than in other industries [<xref ref-type="bibr" rid="ref10">10</xref>]. Unlike commercial implementations, GPT-3 needs to be explicitly discussed in terms of what it will and will not do with a patient’s data. Health care providers’ data governance committees need to be aware and comfortable when they sign the service agreement with GPT-3. Given the black-box nature of GPT-3, an operational strategic approach will be necessary for interpreting the evaluation reports and outcomes that are generated through the human-in-the-loop model.</p>
      <fig id="figure2" position="float">
        <label>Figure 2</label>
        <caption>
          <p>A model of operational and implementation considerations for generative pretrained transformer 3. IS: information systems.</p>
        </caption>
        <graphic xlink:href="medinform_v10i2e32875_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
      </fig>
      <p>Access also needs to be ensured. Training large language models like GPT-3 can cost tens of millions of dollars. As such, GPT-3 is innovating the business model of access. Currently, GPT-3 is privately controlled by OpenAI, and health care providers can remotely run the program and pay for usage per token (1000 tokens are approximately equivalent to 750 words) [<xref ref-type="bibr" rid="ref26">26</xref>]. In September 2020, Microsoft bought an exclusive license to GPT-3, with plans to integrate it into its existing products. Similarly, a number of companies are already integrating GPT-3 model predictions into their products. However, this business model also limits open-access research and development and will eventually limit improvements, such as advancements in translation mechanisms and all-inclusive, equity-driven approaches in conversational agent development. In these early stages, open-source alternatives, such as GPT-J [<xref ref-type="bibr" rid="ref27">27</xref>], may help health care developers and institutions assess operational viability. In future iterations, once the value of using GPT-3 in the health care setting is assured, the responsibility of accessibility could be delegated to health care and government agencies. Such agencies may distribute the “GPT-3-as-a-Service” platform through secure cloud platforms and establish a federated learning mechanism to run decentralized training services while collaboratively contributing to the GPT-3 model [<xref ref-type="bibr" rid="ref28">28</xref>]. This would also reduce the burden on individual health systems when it comes to building, training, and deploying their own GPT-3 platforms and reduce costs. These advantages are especially beneficial for hospitals in low-resource settings.</p>
    </sec>
    <sec>
      <title>Conclusion</title>
      <p>In this viewpoint, we briefly introduce GPT-3 and its capabilities and outline considerations for its implementation and operationalization in clinical practice through a use case. Building on top of Korngiebel and Mooney’s [<xref ref-type="bibr" rid="ref12">12</xref>] remarks toward unrealistic, realistic, feasible, and realistic but challenging use cases, we provide consideration points for implementing and operationalizing GPT-3 in clinical practice. We believe that our work can inform health care practitioners, developers, clinicians, and decision makers toward understanding the use of the powerful AI tools integrated into hospital systems and health care.</p>
    </sec>
  </body>
  <back>
    <app-group/>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">AI</term>
          <def>
            <p>artificial intelligence</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">BLEU</term>
          <def>
            <p>Bilingual Evaluation Understudy</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">EHR</term>
          <def>
            <p>electronic health record</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">GPT-3</term>
          <def>
            <p>generative pretrained transformer 3</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">HIPAA</term>
          <def>
            <p>Health Insurance Portability and Accountability Act</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">METEOR</term>
          <def>
            <p>Metric for Evaluation of Translation With Explicit Ordering</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb7">PCORI</term>
          <def>
            <p>Patient-Centered Outcomes Research Institute</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <ack>
      <p>This study is partially supported through a Patient-Centered Outcomes Research Institute (PCORI) award (award number: ME-2017C1–6413) under the name “Unlocking Clinical Text in EMR by Query Refinement Using Both Knowledge Bases and Word Embedding.” All statements in this report, including its findings and conclusions, are solely those of the authors and do not necessarily represent the views of the PCORI, its Board of Governors, or its Methodology Committee.</p>
    </ack>
    <fn-group>
      <fn fn-type="con">
        <p>ES, JS, and SLL conceived the presented ideas. ES and JS drafted the manuscript. SL supervised and critically reviewed the manuscript. All authors approved the final version of the manuscript.</p>
      </fn>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Brown</surname>
              <given-names>TB</given-names>
            </name>
            <name name-style="western">
              <surname>Mann</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Ryder</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Subbiah</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Kaplan</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Dhariwal</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Neelakantan</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Shyam</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Sastry</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Askell</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Agarwal</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Herbert-Voss</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Krueger</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Henighan</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Child</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Ramesh</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Ziegler</surname>
              <given-names>DM</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Winter</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Hesse</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Sigler</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Litwin</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Gray</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Chess</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Clark</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Berner</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>McCandlish</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Radford</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Sutskever</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Amodei</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Language models are few-shot learners</article-title>
          <source>arXiv. Preprint posted online on July 22, 2020</source>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/pdf/2005.14165.pdf"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Vaswani</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Shazeer</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Parmar</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Uszkoreit</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Jones</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Gomez</surname>
              <given-names>AN</given-names>
            </name>
            <name name-style="western">
              <surname>Kaiser</surname>
              <given-names>Ł</given-names>
            </name>
            <name name-style="western">
              <surname>Polosukhin</surname>
              <given-names>I</given-names>
            </name>
          </person-group>
          <article-title>Attention is all you need</article-title>
          <year>2017</year>
          <conf-name>31st Conference on Neural Information Processing Systems (NIPS 2017)</conf-name>
          <conf-date>December 4-9, 2017</conf-date>
          <conf-loc>Long Beach, California, USA</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://papers.nips.cc/paper/7181-attention-is-all-you-%0Aneed.pdf"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hutson</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Robo-writers: the rise and risks of language-generating AI</article-title>
          <source>Nature</source>
          <year>2021</year>
          <month>03</month>
          <volume>591</volume>
          <issue>7848</issue>
          <fpage>22</fpage>
          <lpage>25</lpage>
          <pub-id pub-id-type="doi">10.1038/d41586-021-00530-0</pub-id>
          <pub-id pub-id-type="medline">33658699</pub-id>
          <pub-id pub-id-type="pii">10.1038/d41586-021-00530-0</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Shen</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Dolan</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Carin</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>W</given-names>
            </name>
          </person-group>
          <article-title>What makes good in-context examples for GPT-3?</article-title>
          <source>arXiv. Preprint posted online on January 17, 2021</source>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/pdf/2101.06804.pdf"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="web">
          <article-title>Emerson</article-title>
          <source>GPT-3 Demo</source>
          <access-date>2021-12-14</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://gpt3demo.com/apps/quickchat-emerson">https://gpt3demo.com/apps/quickchat-emerson</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Langston</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>From conversation to code: Microsoft introduces its first product features powered by GPT-3</article-title>
          <source>The AI Blog</source>
          <year>2021</year>
          <month>05</month>
          <day>25</day>
          <access-date>2021-12-14</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://blogs.microsoft.com/ai/from-conversation-to-code-microsoft-introduces-its-first-product-features-powered-by-gpt-3/">https://blogs.microsoft.com/ai/from-conversation-to-code-microsoft-introduces-its-first-product-features-powered-by-gpt-3/</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="web">
          <article-title>A robot wrote this entire article. Are you scared yet, human? GPT-3</article-title>
          <source>The Guardian</source>
          <year>2020</year>
          <month>09</month>
          <day>08</day>
          <access-date>2022-02-01</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.theguardian.com/commentisfree/2020/sep/08/robot-wrote-this-article-gpt-3">https://www.theguardian.com/commentisfree/2020/sep/08/robot-wrote-this-article-gpt-3</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="web">
          <article-title>Examples - OpenAI API</article-title>
          <source>OpenAI</source>
          <access-date>2021-12-14</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://beta.openai.com/examples/">https://beta.openai.com/examples/</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wiens</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Price</surname>
              <given-names>WN 2nd</given-names>
            </name>
            <name name-style="western">
              <surname>Sjoding</surname>
              <given-names>MW</given-names>
            </name>
          </person-group>
          <article-title>Diagnosing bias in data-driven algorithms for healthcare</article-title>
          <source>Nat Med</source>
          <year>2020</year>
          <month>01</month>
          <volume>26</volume>
          <issue>1</issue>
          <fpage>25</fpage>
          <lpage>26</lpage>
          <pub-id pub-id-type="doi">10.1038/s41591-019-0726-6</pub-id>
          <pub-id pub-id-type="medline">31932798</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41591-019-0726-6</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Watson</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Hutyra</surname>
              <given-names>CA</given-names>
            </name>
            <name name-style="western">
              <surname>Clancy</surname>
              <given-names>SM</given-names>
            </name>
            <name name-style="western">
              <surname>Chandiramani</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Bedoya</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Ilangovan</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Nderitu</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Poon</surname>
              <given-names>EG</given-names>
            </name>
          </person-group>
          <article-title>Overcoming barriers to the adoption and implementation of predictive modeling and machine learning in clinical care: what can we learn from US academic medical centers?</article-title>
          <source>JAMIA Open</source>
          <year>2020</year>
          <month>07</month>
          <day>10</day>
          <volume>3</volume>
          <issue>2</issue>
          <fpage>167</fpage>
          <lpage>172</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/32734155"/>
          </comment>
          <pub-id pub-id-type="doi">10.1093/jamiaopen/ooz046</pub-id>
          <pub-id pub-id-type="medline">32734155</pub-id>
          <pub-id pub-id-type="pii">ooz046</pub-id>
          <pub-id pub-id-type="pmcid">PMC7382631</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Waring</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Lindvall</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Umeton</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Automated machine learning: Review of the state-of-the-art and opportunities for healthcare</article-title>
          <source>Artif Intell Med</source>
          <year>2020</year>
          <month>04</month>
          <volume>104</volume>
          <fpage>101822</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S0933-3657(19)31043-7"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.artmed.2020.101822</pub-id>
          <pub-id pub-id-type="medline">32499001</pub-id>
          <pub-id pub-id-type="pii">S0933-3657(19)31043-7</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Korngiebel</surname>
              <given-names>DM</given-names>
            </name>
            <name name-style="western">
              <surname>Mooney</surname>
              <given-names>SD</given-names>
            </name>
          </person-group>
          <article-title>Considering the possibilities and pitfalls of Generative Pre-trained Transformer 3 (GPT-3) in healthcare delivery</article-title>
          <source>NPJ Digit Med</source>
          <year>2021</year>
          <month>06</month>
          <day>03</day>
          <volume>4</volume>
          <issue>1</issue>
          <fpage>93</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41746-021-00464-x"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41746-021-00464-x</pub-id>
          <pub-id pub-id-type="medline">34083689</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41746-021-00464-x</pub-id>
          <pub-id pub-id-type="pmcid">PMC8175735</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Miller</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>When algorithmic fairness fixes fail: The case for keeping humans in the loop</article-title>
          <source>Stanford University</source>
          <year>2020</year>
          <month>11</month>
          <day>02</day>
          <access-date>2021-12-14</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://hai.stanford.edu/news/when-algorithmic-fairness-fixes-fail-case-keeping-humans-loop">https://hai.stanford.edu/news/when-algorithmic-fairness-fixes-fail-case-keeping-humans-loop</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lucy</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Bamman</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Gender and representation bias in GPT-3 generated stories</article-title>
          <source>Proceedings of the Third Workshop on Narrative Understanding</source>
          <year>2021</year>
          <month>06</month>
          <conf-name>Third Workshop on Narrative Understanding</conf-name>
          <conf-date>June 11, 2021</conf-date>
          <conf-loc>Virtual</conf-loc>
          <fpage>48</fpage>
          <lpage>55</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://aclanthology.org/2021.nuse-1.5.pdf"/>
          </comment>
          <pub-id pub-id-type="doi">10.18653/v1/2021.nuse-1.5</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bender</surname>
              <given-names>EM</given-names>
            </name>
            <name name-style="western">
              <surname>Gebru</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>McMillan-Major</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Shmitchell</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>On the dangers of stochastic parrots: Can language models be too big?</article-title>
          <year>2021</year>
          <month>03</month>
          <conf-name>FAccT '21: 2021 ACM Conference on Fairness, Accountability, and Transparency</conf-name>
          <conf-date>March 3-10, 2021</conf-date>
          <conf-loc>Canada</conf-loc>
          <pub-id pub-id-type="doi">10.1145/3442188.3445922</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>McGuffie</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Newhouse</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>The radicalization risks of GPT-3 and advanced neural language models</article-title>
          <source>arXiv. Preprint posted online on September 15, 2020</source>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/pdf/2009.06807.pdf"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Daws</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Medical chatbot using OpenAI’s GPT-3 told a fake patient to kill themselves</article-title>
          <source>AI News</source>
          <year>2020</year>
          <month>10</month>
          <day>28</day>
          <access-date>2022-02-01</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://artificialintelligence-news.com/2020/10/28/medical-chatbot-openai-gpt3-patient-kill-themselves/">https://artificialintelligence-news.com/2020/10/28/medical-chatbot-openai-gpt3-patient-kill-themselves/</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>FitzGerald</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Hurst</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Implicit bias in healthcare professionals: a systematic review</article-title>
          <source>BMC Med Ethics</source>
          <year>2017</year>
          <month>03</month>
          <day>01</day>
          <volume>18</volume>
          <issue>1</issue>
          <fpage>19</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://bmcmedethics.biomedcentral.com/articles/10.1186/s12910-017-0179-8"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/s12910-017-0179-8</pub-id>
          <pub-id pub-id-type="medline">28249596</pub-id>
          <pub-id pub-id-type="pii">10.1186/s12910-017-0179-8</pub-id>
          <pub-id pub-id-type="pmcid">PMC5333436</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Papineni</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Roukos</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Ward</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>WJ</given-names>
            </name>
          </person-group>
          <article-title>Bleu: a method for automatic evaluation of machine translation</article-title>
          <year>2002</year>
          <month>07</month>
          <conf-name>40th Annual Meeting of the Association for Computational Linguistics</conf-name>
          <conf-date>July 7-12, 2002</conf-date>
          <conf-loc>Philadelphia, Pennsylvania, USA</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.aclweb.org/anthology/P02-1040.pdf"/>
          </comment>
          <pub-id pub-id-type="doi">10.3115/1073083.1073135</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Banerjee</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Lavie</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>METEOR: An automatic metric for MT evaluation with improved correlation with human judgments</article-title>
          <year>2005</year>
          <month>06</month>
          <conf-name>ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization</conf-name>
          <conf-date>June 29, 2005</conf-date>
          <conf-loc>Ann Arbor, Michigan</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.aclweb.org/anthology/W05-0909.pdf"/>
          </comment>
          <pub-id pub-id-type="doi">10.3115/1626355.1626389</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Celikyilmaz</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Clark</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Gao</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Evaluation of text generation: A survey</article-title>
          <source>arXiv. Preprint posted online on May 18, 2021</source>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/pdf/2006.14799.pdf"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jiang</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Introducing new Alexa healthcare skills</article-title>
          <source>Amazon Alexa</source>
          <year>2019</year>
          <month>04</month>
          <day>04</day>
          <access-date>2022-02-01</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://developer.amazon.com/blogs/alexa/post/ff33dbc7-6cf5-4db8-b203-99144a251a21/introducing-new-alexa-healthcare-skills">https://developer.amazon.com/blogs/alexa/post/ff33dbc7-6cf5-4db8-b203-99144a251a21/introducing-new-alexa-healthcare-skills</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>McGraw</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Mandl</surname>
              <given-names>KD</given-names>
            </name>
          </person-group>
          <article-title>Privacy protections to encourage use of health-relevant digital data in a learning health system</article-title>
          <source>NPJ Digit Med</source>
          <year>2021</year>
          <month>01</month>
          <day>04</day>
          <volume>4</volume>
          <issue>1</issue>
          <fpage>2</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41746-020-00362-8"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41746-020-00362-8</pub-id>
          <pub-id pub-id-type="medline">33398052</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41746-020-00362-8</pub-id>
          <pub-id pub-id-type="pmcid">PMC7782585</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="web">
          <article-title>Artificial intelligence and machine learning in Software as a Medical Device</article-title>
          <source>U.S. Food &#38; Drug Administration</source>
          <access-date>2021-12-14</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.fda.gov/medical-devices/software-medical-device-samd/artificial-intelligence-and-machine-learning-software-medical-device">https://www.fda.gov/medical-devices/software-medical-device-samd/artificial-intelligence-and-machine-learning-software-medical-device</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="journal">
          <article-title>Patient trust must come at the top of researchers' priority list</article-title>
          <source>Nat Med</source>
          <year>2020</year>
          <month>03</month>
          <volume>26</volume>
          <issue>3</issue>
          <fpage>301</fpage>
          <pub-id pub-id-type="doi">10.1038/s41591-020-0813-8</pub-id>
          <pub-id pub-id-type="medline">32161404</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41591-020-0813-8</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Brockman</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Murati</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Welinder</surname>
              <given-names>P</given-names>
            </name>
            <collab>OpenAI</collab>
          </person-group>
          <article-title>OpenAI API</article-title>
          <source>OpenAI</source>
          <year>2020</year>
          <month>06</month>
          <day>11</day>
          <access-date>2021-12-14</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://openai.com/blog/openai-api/">https://openai.com/blog/openai-api/</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref27">
        <label>27</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Romero</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Can’t access GPT-3? Here’s GPT-J — its open-source cousin</article-title>
          <source>Towards Data Science</source>
          <year>2021</year>
          <month>06</month>
          <day>24</day>
          <access-date>2021-12-14</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://towardsdatascience.com/cant-access-gpt-3-here-s-gpt-j-its-open-source-cousin-8af86a638b11">https://towardsdatascience.com/cant-access-gpt-3-here-s-gpt-j-its-open-source-cousin-8af86a638b11</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref28">
        <label>28</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sadilek</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Nguyen</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Kamruzzaman</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Serghiou</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Rader</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Ingerman</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Mellem</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Kairouz</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Nsoesie</surname>
              <given-names>EO</given-names>
            </name>
            <name name-style="western">
              <surname>MacFarlane</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Vullikanti</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Marathe</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Eastham</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Brownstein</surname>
              <given-names>JS</given-names>
            </name>
            <name name-style="western">
              <surname>Arcas</surname>
              <given-names>BAY</given-names>
            </name>
            <name name-style="western">
              <surname>Howell</surname>
              <given-names>MD</given-names>
            </name>
            <name name-style="western">
              <surname>Hernandez</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Privacy-first health research with federated learning</article-title>
          <source>NPJ Digit Med</source>
          <year>2021</year>
          <month>09</month>
          <day>07</day>
          <volume>4</volume>
          <issue>1</issue>
          <fpage>132</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41746-021-00489-2"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41746-021-00489-2</pub-id>
          <pub-id pub-id-type="medline">34493770</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41746-021-00489-2</pub-id>
          <pub-id pub-id-type="pmcid">PMC8423792</pub-id>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
