<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.0" xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMI</journal-id>
      <journal-id journal-id-type="nlm-ta">JMIR Med Inform</journal-id>
      <journal-title>JMIR Medical Informatics</journal-title>
      <issn pub-type="epub">2291-9694</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v14i1e83745</article-id>
      <article-id pub-id-type="pmid">41915420</article-id>
      <article-id pub-id-type="doi">10.2196/83745</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>Analysis Model for Infant Incubator Adverse Events Using Retrieval-Augmented Generation Combined With Dual-Adapter Fine-Tuning: Development and Evaluation Study</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Coristine</surname>
            <given-names>Andrew</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Kim</surname>
            <given-names>Eunhee</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Wang</surname>
            <given-names>Chuang</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Yusuff</surname>
            <given-names>Taofeek</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author">
          <name name-style="western">
            <surname>Xia</surname>
            <given-names>Wenke</given-names>
          </name>
          <degrees>MEng</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0009-0009-6286-0595</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author">
          <name name-style="western">
            <surname>Zhu</surname>
            <given-names>Wanting</given-names>
          </name>
          <degrees>MEng</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0009-0004-2696-3873</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author">
          <name name-style="western">
            <surname>Li</surname>
            <given-names>Tianchun</given-names>
          </name>
          <degrees>MEng</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0009-0002-0354-2756</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author">
          <name name-style="western">
            <surname>Wang</surname>
            <given-names>Li</given-names>
          </name>
          <degrees>MSc</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0009-0002-6588-6917</ext-link>
        </contrib>
        <contrib id="contrib5" contrib-type="author">
          <name name-style="western">
            <surname>Li</surname>
            <given-names>Weiqi</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0009-0005-2537-8784</ext-link>
        </contrib>
        <contrib id="contrib6" contrib-type="author" corresp="yes">
          <name name-style="western">
            <surname>Zhang</surname>
            <given-names>Peiming</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <address>
            <institution>School of Health Science and Engineering</institution>
            <institution>University of Shanghai for Science and Technology</institution>
            <addr-line>Number 516, Jungong Road</addr-line>
            <addr-line>Yangpu District</addr-line>
            <addr-line>Shanghai, 200093</addr-line>
            <country>China</country>
            <phone>86 13524693176</phone>
            <email>zpmking@163.com</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0009-0003-2409-5323</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>School of Health Science and Engineering</institution>
        <institution>University of Shanghai for Science and Technology</institution>
        <addr-line>Shanghai</addr-line>
        <country>China</country>
      </aff>
      <aff id="aff2">
        <label>2</label>
        <institution>Center for Drug Reevaluation of Henan</institution>
        <addr-line>Zhengzhou</addr-line>
        <country>China</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Peiming Zhang <email>zpmking@163.com</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <year>2026</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>31</day>
        <month>3</month>
        <year>2026</year>
      </pub-date>
      <volume>14</volume>
      <elocation-id>e83745</elocation-id>
      <history>
        <date date-type="received">
          <day>8</day>
          <month>9</month>
          <year>2025</year>
        </date>
        <date date-type="rev-request">
          <day>27</day>
          <month>12</month>
          <year>2025</year>
        </date>
        <date date-type="accepted">
          <day>10</day>
          <month>3</month>
          <year>2026</year>
        </date>
      </history>
      <copyright-statement>©Wenke Xia, Wanting Zhu, Tianchun Li, Li Wang, Weiqi Li, Peiming Zhang. Originally published in JMIR Medical Informatics (https://medinform.jmir.org), 31.03.2026.</copyright-statement>
      <copyright-year>2026</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on https://medinform.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://medinform.jmir.org/2026/1/e83745" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>Infant incubator adverse events refer to various harmful incidents that occur during the normal use of marketed infant incubators and result in, or may result in, bodily harm. In recent years, however, the number of reported adverse events has continued to rise. This trend has made the monitoring of infant incubator adverse events time-consuming and labor-intensive when relying solely on manual processing by medical device adverse-event monitoring personnel. Meanwhile, general-purpose large language models (LLMs) still face domain knowledge gaps and hallucination issues in specialized fields. Through fine-tuning, LLMs can be adapted to specific application scenarios, while retrieval-augmented generation (RAG) enhances their ability to handle knowledge-intensive tasks. Therefore, LLMs that integrate these 2 technologies hold significant potential for addressing monitoring challenges.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>This study aims to enhance the intelligent monitoring of adverse events related to medical devices by integrating RAG with dual-adapter fine-tuning to construct an adverse-event analysis model for infant incubators. The model enables an integrated workflow centered on extraction and analysis, encompassing structured extraction of adverse events, narrative analysis, and regulatory question answering.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>This study leveraged adverse-event data from Chinese infant incubators to construct a high-quality dataset through prompt engineering. Technologically, it combined 2 parameter-efficient fine-tuning methods—low-rank adaptation and infused adapter by inhibiting and amplifying inner activations—to achieve efficient adaptation on the Qwen2-7B base model. Simultaneously, it introduced the FINBGE embedding model with supervised contrastive semantic optimization to build a knowledge retrieval system that mitigates hallucination.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>This study comprises 1565 pediatric disease question-and-answer entries from PediaBench, 2530 specific corpora on adverse events in infant incubators, and 1488 regulatory corpora. Extensive experiments demonstrate the superiority of this analytical model across various metrics. Under the same experimental conditions, the element recall rate reaches 0.815, the accuracy of infant incubator adverse event analysis is 0.898, and the accuracy in regulatory clause question-answering tasks attains 0.938.</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>The analytical model proposed in this study demonstrates significant advantages in analyzing adverse events related to infant incubators, while also achieving substantial improvements in text generation metrics. When combined with RAG, it not only effectively mitigates hallucination but also enhances knowledge timeliness. This study employs the Qwen model as its foundational framework, leveraging large-model fine-tuning and RAG to achieve intelligent analysis of high-risk medical device adverse-event monitoring data. It demonstrates the feasibility of implementing intelligent medical device regulation within the Qwen ecosystem.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>adverse events in medical devices</kwd>
        <kwd>infant incubators</kwd>
        <kwd>large language models</kwd>
        <kwd>fine-tuning techniques</kwd>
        <kwd>retrieval-augmented generation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <sec>
        <title>Background</title>
        <p>Collecting and analyzing adverse event monitoring data for high-risk medical devices is a crucial means of ensuring their safety and effectiveness [<xref ref-type="bibr" rid="ref1">1</xref>]. Currently, medical device regulatory authorities have widely adopted electronic monitoring systems, which not only enhance data collection efficiency and convenience but also lead to a sustained increase in the volume of adverse event reports submitted [<xref ref-type="bibr" rid="ref2">2</xref>]. Badnjević et al [<xref ref-type="bibr" rid="ref3">3</xref>] found that despite clear regulatory mandates for postmarket surveillance obligations across countries, issues persist in adverse event reporting, including inconsistent standards, noncomparable data, and fragmented oversight. In China, infant incubators are classified as class III high-risk medical devices. Adverse event monitoring for these products involves interdisciplinary knowledge integration across multiple fields. Challenges include large volumes of adverse event data, predominantly in unstructured and nonstandardized formats [<xref ref-type="bibr" rid="ref4">4</xref>], as well as significant variation in data quality reported by health care institutions [<xref ref-type="bibr" rid="ref5">5</xref>]. These factors substantially increase the workload and complexity of analyzing and processing reports. Recent innovations in natural language processing technology offer new possibilities for balancing efficiency and professionalism. Ruseva et al [<xref ref-type="bibr" rid="ref6">6</xref>] proposed a deep learning model based on the TensorFlow framework, demonstrating strong performance in assessing adverse drug reaction risks and enhancing postmarketing drug surveillance capabilities. Hauben [<xref ref-type="bibr" rid="ref7">7</xref>] developed artificial intelligence (AI)–based methods to enhance drug-drug interaction detection and prediction. Borjali et al [<xref ref-type="bibr" rid="ref8">8</xref>] integrated deep learning to accurately and promptly detect adverse medical events from free-text medical narratives.</p>
        <p>While traditional deep learning models can perform text classification, they lack long-range contextual associations [<xref ref-type="bibr" rid="ref9">9</xref>] and causal reasoning capabilities [<xref ref-type="bibr" rid="ref10">10</xref>], exhibit weak generalization, and demonstrate poor transferability between events [<xref ref-type="bibr" rid="ref11">11</xref>]. Lu et al [<xref ref-type="bibr" rid="ref12">12</xref>] found that Transformer models consistently outperformed traditional convolutional neural network models in nearly all medical text classification scenarios, with convolutional neural networks demonstrating acceptable performance only when categories were balanced. Naik et al [<xref ref-type="bibr" rid="ref13">13</xref>] highlighted that adapting event extractors to new medical domains suffers from covariate shift issues, rendering direct transfer ineffective and necessitating specialized domain adaptation techniques.</p>
        <p>Although large language models (LLMs) can provide solutions to core challenges in medical device adverse event monitoring, pretraining foundational models from scratch for vertical domains incurs prohibitively high costs. For instance, first, training a model with tens of billions of parameters requires multiple graphics processing units and months of computational time [<xref ref-type="bibr" rid="ref14">14</xref>], demanding substantial computational resources and time investment. Second, pretraining often requires datasets in the gigabyte or even terabyte range [<xref ref-type="bibr" rid="ref15">15</xref>], while vertical domain data are typically strictly confidential [<xref ref-type="bibr" rid="ref16">16</xref>]. Publicly available data may constitute less than 0.1% of pretraining requirements, and the processes of data collection, cleaning, annotation, and storage involve substantial human and material resources. Finally, pretraining imposes steep technical barriers, demanding both deep algorithmic engineering expertise and the development of large-scale distributed training frameworks [<xref ref-type="bibr" rid="ref17">17</xref>]—a burden beyond the reach of most small-to-medium teams.</p>
        <p>Therefore, this study constructs an infant incubator adverse event analysis model by fine-tuning LLMs. Methods such as low-rank adaptation (LoRA) and infused adapter by inhibiting and amplifying inner activations, that is, (IA)<sup>3</sup>, require only minor weight updates, preserving general language prior knowledge while incorporating domain-specific expertise [<xref ref-type="bibr" rid="ref18">18</xref>]. However, for tasks involving the analysis of adverse events related to infant incubators—which are complex, densely populated with specialized terminology, and characterized by imbalanced samples—a single method has inherent limitations. LoRA excels at capturing long-range contextual dependencies and temporal event correlations through low-rank updates, but its global low-rank matrix may lack sufficient feature resolution when handling rare fault descriptions or subtle terminological variations. Meanwhile, (IA)<sup>3</sup> enables fine-grained control over the strength of specific features through dynamic scaling of activation values, but it struggles to model long-range semantic relationships across sentences or paragraphs. Neither approach alone can comprehensively address these challenges.</p>
        <p>To this end, this study proposes a “dual-adapter joint fine-tuning” architecture. Its novelty lies not in simple superposition but in a hierarchical division of labor and collaboration: the LoRA component focuses on modeling logical chains and long-range dependencies within adverse events, while the (IA)<sup>3</sup> component performs local feature enhancement and normalization on terminology and key entities to suppress noise. Coupled through a dynamic weight transfer mechanism, this approach enables the model to achieve deep adaptation to vertical domain knowledge while preserving general medical semantic understanding. Although vertical domains often lack extensive data, Wang et al [<xref ref-type="bibr" rid="ref19">19</xref>] demonstrated the feasibility of studying small-sample adverse event data in medical devices when investigating knowledge transfer and calibration for small-sample medical signals (epileptic electroencephalography).</p>
        <p>High-quality and task-relevant data form the cornerstone for ensuring that fine-tuned models achieve the expected performance, accuracy, and specific behavioral patterns. Data quality directly determines the nuances learned by the model, the biases it may inherit or mitigate, and its generalization capabilities [<xref ref-type="bibr" rid="ref20">20</xref>]. The data source used in this study comprises nonpublic adverse event data from infant incubators within China’s Medical Device Adverse Event Monitoring System (the details of adverse events are shown in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>). This dataset authentically captures adverse events generated by infant incubator use in medical institutions in recent years, encompassing complex and diverse semantic expressions within the domain. Real-world noise reflects the challenges of practical application scenarios; mild noise can actually compel the model to learn deeper semantic features rather than relying on spurious patterns in the data, thereby enhancing model robustness [<xref ref-type="bibr" rid="ref21">21</xref>].</p>
        <p>It is important to note that even after model fine-tuning, hallucination persists. From a computational theory perspective, all AI models—including current mainstream LLMs—are fundamentally probabilistic mathematical function approximators [<xref ref-type="bibr" rid="ref22">22</xref>]. These models map inputs to output spaces through complex parameter matrix operations, generating outputs that are essentially responses sampled from high-dimensional probability distributions. Retrieval-augmented generation (RAG) significantly mitigates this issue. It is an AI technique that combines information retrieval technology with language generation models [<xref ref-type="bibr" rid="ref23">23</xref>]. This approach enhances the model’s capability to handle knowledge-intensive tasks by retrieving relevant information from external knowledge bases and feeding it as prompts to LLMs.</p>
      </sec>
      <sec>
        <title>Related Work</title>
        <p>In recent years, the rapid advancement of AI technology has increasingly highlighted the application value of large-scale pretrained models across multiple professional domains. Advanced models such as GPT and Bidirectional Encoder Representations from Transformers (BERT), with their exceptional language comprehension and generation capabilities, have demonstrated significant application potential in the health care sector. Research data indicate that models based on the Transformer architecture possess unique advantages in medical text processing: Kim et al [<xref ref-type="bibr" rid="ref24">24</xref>] confirmed that domain-adapted BERT variants perform well in predicting appropriate medical departments based on patient symptom descriptions and in answering specialized medical questions. Zhang et al [<xref ref-type="bibr" rid="ref25">25</xref>] utilized LLMs for medical record generation tasks, saving physicians approximately 0.5-1 hour of work time daily.</p>
        <p>Notably, the application scope of such models has expanded beyond text processing into medical devices and regulatory domains. Tang et al [<xref ref-type="bibr" rid="ref26">26</xref>] highlighted that the global AI medical device industry is experiencing robust growth, while the urgent need for regulatory oversight is becoming increasingly apparent—a trend expected to intensify. For instance, Kim et al [<xref ref-type="bibr" rid="ref27">27</xref>] developed a deep learning predictive model to rapidly and accurately detect mechanical ventilation requirements in neonatal intensive care units using electronic health records; Matheny et al [<xref ref-type="bibr" rid="ref28">28</xref>] systematically explored how applying LLMs to the US Food and Drug Administration’s (FDA) medical product safety surveillance system enhances the identification and assessment of adverse drug events.</p>
        <p>Regarding fine-tuning medical LLMs, multiple studies have explored application methods. For instance, Jia et al [<xref ref-type="bibr" rid="ref29">29</xref>] employed LoRA for instruction fine-tuning of pretrained LLMs, ultimately enhancing the automated generation of children’s health science resource labels. Abbasian et al [<xref ref-type="bibr" rid="ref30">30</xref>] utilized electronic health records from patients with diabetes for P-tuning and LoRA fine-tuning of large models, opening new avenues for diabetes management and improving therapeutic efficacy. Integrating large models with RAG has also become a research hotspot. Hu et al [<xref ref-type="bibr" rid="ref31">31</xref>] introduced a pioneering personalized dementia care approach by integrating GPT-4, fine-tuned and RAG-enabled, with a vector database. Kulshreshtha et al [<xref ref-type="bibr" rid="ref32">32</xref>] enhanced a medical chatbot based on the Llama source model using RAG methods to generate reliable, contextually relevant responses, thereby improving clinical judgment and reducing health care workers’ workload.</p>
      </sec>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <sec>
        <title>Overview</title>
        <p>The adverse event analysis model AE2IN proposed in this study for infant incubators comprises 2 core modules: a data preprocessing layer and a knowledge-enhanced analysis layer. The specific framework is illustrated in <xref rid="figure1" ref-type="fig">Figure 1</xref>.</p>
        <fig id="figure1" position="float">
          <label>Figure 1</label>
          <caption>
            <p>Framework diagram of the AE2IN adverse event analysis model for infant incubators. The preprocessing layer cleans the data, normalizes terminology, and enhances adverse event reports. The knowledge enhancement layer enables collaborative analysis through external knowledge injection and large-model semantic reasoning, using a Qwen model fine-tuned with dual adapters alongside a fine-tuned embedding model. FFN: feedforward network; (IA)3: Infused Adapter by Inhibiting and Amplifying Inner Activations; LLM: large language model; LoRA: low-rank adaptation.</p>
          </caption>
          <graphic xlink:href="medinform_v14i1e83745_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <p>At the data preprocessing layer, the system first standardizes input adverse event reports for infant incubators through data filtering, cleaning, and augmentation. The data standardization, cleaning, and augmentation processes for this study’s training dataset leveraged GPT-4 as a teacher model for auxiliary processing. Through knowledge distillation in a teacher-student paradigm, it provided high-quality training samples for subsequent domain-specific fine-tuning of the Qwen2_7B model. Standardization during conventional inference fully utilizes specialized medical dictionaries and regular expression rules to ensure accurate recognition of diverse terminology. The preprocessed structured data are subsequently fed into the knowledge-enhanced analysis layer.</p>
        <p>The knowledge-enhanced analysis layer serves as the system’s core, employing a dual-path parallel processing mechanism. On the one hand, supervised comparison-based semantic optimization fine-tunes the embedded model FINBGE to vectorize queries and retrieve the most relevant clause content from the medical device regulatory knowledge base. On the other hand, the Qwen2_7B model, jointly fine-tuned via dual adapters, performs deep semantic understanding of the input text. Notably, this model innovatively combines 2 parameter-efficient fine-tuning techniques: LoRA and (IA)<sup>3</sup>.</p>
      </sec>
      <sec>
        <title>Fine-Tuning</title>
        <sec>
          <title>Dual-Adapter Fine-Tuning</title>
          <p>This study designed a dual-adapter joint fine-tuning method that combines LoRA and (IA)<sup>3</sup>, 2 parameter-efficient fine-tuning techniques. LoRA focuses on optimizing the Q (query)/V (value) projection matrices within the attention mechanism through low-rank matrix updates (<italic>r</italic>=32), effectively capturing contextual dependencies in descriptions of adverse events in infant incubators. Meanwhile, (IA)<sup>3</sup> refines the feature representations of the feedforward network by learning feature scaling coefficients without introducing additional parameters, thereby enhancing the representation of abnormal features and the generation of handling suggestions. Mathematically, this can be expressed as a composite transformation:</p>
          <graphic xlink:href="medinform_v14i1e83745_fig16.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          <p>where <italic>W</italic> represents the original pretrained weights, Δ<italic>W</italic> is the low-rank update matrix introduced by LoRA, composed of the product of 2 small matrices A and B, and <italic>l</italic> is the feature scaling vector learned by (IA)<sup>3</sup>. This architecture retains LoRA’s advantage in pattern discovery—learning domain-specific attention patterns through low-rank space—while integrating the precise feature selection control of (IA)<sup>3</sup>. The overall framework is illustrated in <xref rid="figure2" ref-type="fig">Figure 2</xref>.</p>
          <fig id="figure2" position="float">
            <label>Figure 2</label>
            <caption>
              <p>Dual-adapter fine-tuning framework diagram. The architecture is built on the Qwen2_7B foundation model and integrates 2 parameter-efficient methods: low-rank adaptation (LoRA) and Infused Adapter by Inhibiting and Amplifying Inner Activations (IA)3. The diagram illustrates the core design, with LoRA applied to the query/value projection matrices in the Transformer attention mechanism and (IA)3 applied to the feedforward network (FFN). It also depicts the fine-tuning strategy, where pretrained weights are frozen and only adapter parameters are updated, highlighting key processes such as attention computation, weight composition, and feature scaling.</p>
            </caption>
            <graphic xlink:href="medinform_v14i1e83745_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
          <p>The domain-adaptive training strategy employs a dynamic weighting mechanism to balance domain-specific data with general-purpose corpora. Its objective function is defined as follows:</p>
          <graphic xlink:href="medinform_v14i1e83745_fig17.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          <graphic xlink:href="medinform_v14i1e83745_fig18.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          <p>where <italic>λ<sub>d</sub></italic> is the domain data weight coefficient; <italic>λ<sub>g</sub></italic> is the general data weight coefficient, with a ratio of 3:1; <italic>D<sub>d</sub></italic> and <italic>D<sub>g</sub></italic> represent the domain dataset and general dataset, respectively; <inline-graphic xlink:href="medinform_v14i1e83745_fig19.png" xlink:type="simple" mimetype="image"/> denotes the loss function for a single sample; <italic>y<sub>t</sub></italic> is the <italic>t</italic>th token in the target sequence; <italic>y</italic><sub>&lt;</sub><italic><sub>t</sub></italic> indicates the historical output sequence; and <italic>T</italic> is the target sequence length.</p>
          <p>By employing a 3:1 weighted sampling mechanism, domain data receive high-frequency exposure to accelerate the model’s mastery of infant incubator adverse event norms, while continuous infusion of general data maintains foundational medical semantic understanding.</p>
          <p>This method can be divided into 2 stages. The first stage focuses on optimizing the attention mechanism. By utilizing low-rank matrix updates, it establishes cross-layer semantic associations for factors in the Q/V projection matrix that contribute to adverse events in infant incubators. This step notably enhances the model’s ability to understand temporal patterns in infant incubator adverse events. It can be conceptualized as the model’s capability to accurately capture causal chains such as “temperature alarm → skin temperature probe failure → component inspection.” The second stage specifically adjusts the feature representations in the feedforward network. By learning scaling coefficient vectors, it finely controls the generation intensity of infant incubator adverse event terminology and the standardization of subsequent handling recommendations. For example, operations such as “Immediately replace the incubator, monitor the infant’s temperature and pulse oxygen saturation, and report to the equipment department for repair” fall under the follow-up actions category, while information such as “Equipment model YP-90B, root cause analysis of the fault” is accurately categorized under the additional information category. This ensures that the model’s output maintains both professional precision and compliance with medical documentation standards.</p>
          <p>Knowledge inheritance between stages is achieved through dynamic weight transfer—using attention patterns obtained from LoRA training as initialization priors for the (IA)<sup>3</sup> stage. This strategy converges faster than parallel training while preserving model generalization capabilities and notably enhancing domain-specific task performance. It effectively overcomes the limitations of either approach: fine-grained feature variations that are difficult for LoRA to capture are compensated by (IA)<sup>3</sup>, while long-range dependency modeling—a weakness of (IA)<sup>3</sup>—is supplemented by LoRA.</p>
        </sec>
        <sec>
          <title>Supervised Contrastive Semantic Optimization Fine-Tuning</title>
          <p>In specialized domains, general-purpose embedding models often struggle to accurately capture subtle distinctions between professional terminology, leading to similar yet unrelated concepts being overly close in vector space [<xref ref-type="bibr" rid="ref33">33</xref>]. Therefore, this paper proposes a supervised contrastive semantic optimization method to fine-tune embedding models. The fine-tuning workflow framework is illustrated in <xref rid="figure3" ref-type="fig">Figure 3</xref>.</p>
          <fig id="figure3" position="float">
            <label>Figure 3</label>
            <caption>
              <p>Framework diagram of the supervised contrastive semantic optimization fine-tuning process. The model performs comparative learning and parameter updates using both the training and validation datasets.</p>
            </caption>
            <graphic xlink:href="medinform_v14i1e83745_fig3.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
          <p>By constructing an end-to-end fine-tuning pipeline, we achieve efficient domain adaptation for pretrained embedding models. Utilizing the Sentence Transformers Finetune Engine framework, we combine contrastive learning mechanisms with domain knowledge injection. This enables the model to accurately capture domain-specific semantic association patterns while preserving its original semantic understanding capabilities. The core training objective is mathematically expressed as follows:</p>
          <graphic xlink:href="medinform_v14i1e83745_fig20.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          <p>where <italic>q<sub>i</sub></italic> is the embedding vector for query <italic>i</italic>; <italic>p<sub>i</sub></italic><sup>+</sup> and <italic>p<sub>i</sub></italic><sup>–</sup> are the positive and negative samples associated with <italic>q<sub>i</sub></italic>, respectively; <italic>τ</italic> is the temperature coefficient controlling the sharpness of the distribution; <italic>N</italic> denotes the number of queries in the batch; and <italic>K</italic> represents the number of negative samples corresponding to each query. Cosine similarity establishes the relevance metric between queries and documents. The temperature parameter <italic>τ</italic> controls the sensitivity of similarity scores. The numerator term <italic>e</italic>[sim(<italic>q<sub>i</sub></italic>, <italic>p<sub>i</sub></italic><sup>+</sup>)/<italic>τ</italic>] generates centripetal force, drawing relevant documents closer to the query point in vector space. The denominator term <inline-graphic xlink:href="medinform_v14i1e83745_fig21.png" xlink:type="simple" mimetype="image"/> generates a repulsive field, pushing irrelevant documents away from the query point’s semantic neighborhood.</p>
          <p>Beyond reconstructing the semantic space, feature space calibration requires an embedding projection transformation, formulated as:</p>
          <disp-formula><italic>h</italic> = LayerNorm[<italic>W</italic><sub>2</sub>·GELU(W<sub>1</sub><italic>x</italic> + <italic>b</italic><sub>1</sub>) + <italic>b</italic><sub>2</sub>]</disp-formula>
          <p><italic>W</italic><sub>1</sub> and <italic>W</italic><sub>2</sub> form a parameterized transformation matrix that projects the original embedding onto a subspace more suitable for the current task. The GELU function introduces a smooth threshold, preserving important features while filtering out noise. LayerNorm maintains numerical stability, ensuring comparability between embeddings from different domains.</p>
        </sec>
      </sec>
      <sec>
        <title>Infant Incubator Adverse Event Evaluation Indicators</title>
        <p>When analyzing adverse events involving infant incubators, core elements are of paramount importance. Yan and Huifang [<xref ref-type="bibr" rid="ref34">34</xref>] proposed analyzing both the manifestations and root causes within adverse event data for infant incubators. Kaur et al [<xref ref-type="bibr" rid="ref35">35</xref>] suggested describing the implemented control measures. Additionally, according to the writing requirements of the guidelines issued by the National Medical Products Administration, adverse event reports typically require information such as location, time, source, injury manifestation, device malfunction, cause analysis, and specific control measures. Accordingly, we summarize the core and secondary elements of infant incubator adverse events. The core elements are the “occurrence date” of the adverse event, the “abnormal condition” of the infant incubator, and the “follow-up measures” for risk control. Secondary elements include the “cause of failure” and the “risk triggered” by the infant incubator. Beyond these core and secondary elements, any additional information captured from the original text is categorized as supplementary notes.</p>
        <p>Therefore, during testing experiments, in addition to employing general evaluation metrics, we have developed specialized evaluation indicators for adverse events in infant incubators: the element recall rate and the information density index of adverse events with correction factor (IDIAE). The formulas are as follows:</p>
        <graphic xlink:href="medinform_v14i1e83745_fig22.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        <p>where <italic>K</italic> denotes the core feature set; <italic>E<sub>k</sub></italic><sup>ref</sup> represents feature <italic>k</italic> in the reference text; <italic>E<sub>k</sub></italic><sup>pred</sup> denotes feature <italic>k</italic> in the predicted text; and <italic>I</italic> is the indicator function. During computation, text segments <italic>E<sub>k</sub></italic><sup>ref</sup> and <italic>E<sub>k</sub></italic><sup>pred</sup> corresponding to the 3 core elements are extracted from the reference text and prediction text, respectively. Subsequently, a hierarchical matching assessment is performed: an exact match is determined by checking whether the normalized reference text is fully contained within the prediction text. If the exact match fails, the similarity ratio based on edit distance between the 2 texts is calculated. If this ratio exceeds the threshold of 0.8, the segment is deemed a match.</p>
        <graphic xlink:href="medinform_v14i1e83745_fig23.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        <disp-formula><italic>β =</italic> 1 + <italic>α</italic>·(<italic>N</italic><sub>red</sub>/10) + γ·<italic>M</italic><sub>error</sub></disp-formula>
        <p>where <italic>C<sub>j</sub></italic> represents the coverage status of the <italic>j</italic>th standard report field; <italic>m</italic> denotes the total number of required fields; <italic>w<sub>i</sub></italic> is the weight coefficient for the <italic>i</italic>th category of information (core element weight is 1.0, secondary requirement weight is 0.8); <italic>β</italic> is the text redundancy coefficient, and the redundancy coefficient incorporates a redundancy word penalty weight <italic>α</italic> to assess whether the model’s response is excessively verbose; <italic>N</italic><sub>red</sub> indicates the number of irrelevant characters; <italic>γ</italic> is the terminology error penalty weight used to evaluate whether the model captures professional terminology correctly; and <italic>M</italic><sub>error</sub> represents the number of errors.</p>
        <p>To determine the globally optimal combination of <italic>α</italic> and <italic>γ</italic> and to validate the reliability of the IDIAE metric, this study randomly selected 80 adverse event reports for infant incubators. These were stratified into high-, medium-, and low-quality grades, with 26, 28, and 26 reports, respectively. Experts were invited to rescore the reports under blinded conditions. As shown in <xref rid="figure4" ref-type="fig">Figure 4</xref>, varying only the <italic>α</italic> and <italic>γ</italic> values revealed that the intraclass correlation coefficient peaked at 0.78 (95% CI 0.72-0.83) when <italic>α</italic>=.1 and <italic>γ</italic>=0.2. This indicates a strong positive correlation between IDIAE scores and expert ratings under this weighting configuration (correlation coefficient <italic>r</italic>=0.82, <italic>P</italic>&lt;.001), representing the globally optimal combination. To further validate IDIAE reliability, after fixing the weights (<italic>α</italic>=.1, <italic>γ</italic>=0.2), analyses across high-, medium-, and low-quality grades showed that the mean IDIAE scores for all 3 groups in <xref rid="figure5" ref-type="fig">Figures 5</xref> and <xref rid="figure6" ref-type="fig">6</xref> consistently aligned with expert-rated mean trends for both completeness and accuracy. The classification consistency <italic>κ</italic> coefficient reached 0.71. These results demonstrate that the IDIAE metric effectively reflects expert criteria for assessing report completeness, exhibiting good reliability and clinical applicability.</p>
        <fig id="figure4" position="float">
          <label>Figure 4</label>
          <caption>
            <p>Sensitivity heatmap of Information Density Index of Adverse Events (IDIAE) metric weights (intraclass correlation coefficient [ICC] values). The experiment used 80 randomly selected adverse event reports from infant incubators, with expert blinded scoring as the gold standard. The x- and y-axes represent the redundancy penalty weight (α) and the terminology error penalty weight (γ), respectively. Color blocks indicate the ICC values for each weight combination, enabling identification of the globally optimal parameter setting.</p>
          </caption>
          <graphic xlink:href="medinform_v14i1e83745_fig4.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure5" position="float">
          <label>Figure 5</label>
          <caption>
            <p>Comparison of Information Density Index of Adverse Events (IDIAE) and manual scoring integrity. The figure compares completeness scores and their mean values between the IDIAE metric and human expert ratings across 3 data groups with varying quality levels.</p>
          </caption>
          <graphic xlink:href="medinform_v14i1e83745_fig5.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure6" position="float">
          <label>Figure 6</label>
          <caption>
            <p>Accuracy comparison between Information Density Index of Adverse Events (IDIAE) and human scoring. The figure compares accuracy scores and their mean values between the IDIAE metric and human expert ratings across 3 data groups with varying quality levels.</p>
          </caption>
          <graphic xlink:href="medinform_v14i1e83745_fig6.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Prompt Engineering</title>
        <p>To efficiently align the model with the desired training outcomes, this study designs 3 prompts for different scenarios, as shown in <xref rid="figure7" ref-type="fig">Figures 7</xref>-<xref rid="figure9" ref-type="fig">9</xref>. <xref rid="figure7" ref-type="fig">Figure 7</xref> is used to assist in generating a suitable fine-tuning dataset. <xref rid="figure8" ref-type="fig">Figure 8</xref> presents the system prompt embedded in the inference stage, aiming to standardize the model output. The retrieval prompt integration employs a structured constraint template, as shown in <xref rid="figure9" ref-type="fig">Figure 9</xref>. System instructions enforce that the model must base its responses solely on retrieved content, annotating each sentence with a source identifier [Ref-N] after interpretation. Each retrieved chunk is organized according to a fixed format. The prompt template constrains model behavior through 3 core rules:</p>
        <list list-type="bullet">
          <list-item>
            <p>It explicitly prohibits vague references such as “Article X,” mandating explicit linkage to specific regulatory document names and clause numbers to prevent cross-document clause confusion.</p>
          </list-item>
          <list-item>
            <p>It requires declarations of “insufficient basis” for uncertain content, thereby eliminating model speculation or the fabrication of nonretrieved clauses.</p>
          </list-item>
          <list-item>
            <p>Regulatory requirement sections must alternate between verbatim citations and interpretations, with each interpretive sentence immediately followed by a [Ref-N] annotation.</p>
          </list-item>
        </list>
        <fig id="figure7" position="float">
          <label>Figure 7</label>
          <caption>
            <p>Prompt for generating fine-tuning datasets for embedding models.</p>
          </caption>
          <graphic xlink:href="medinform_v14i1e83745_fig7.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure8" position="float">
          <label>Figure 8</label>
          <caption>
            <p>System prompt for the inference phase. The template defines the model’s role, output specifications, and core constraints for adverse event analysis and regulatory questions and answers in infant incubators. It governs output format, professional compliance, and content boundaries to ensure responses meet specialized regulatory requirements for medical device adverse events.</p>
          </caption>
          <graphic xlink:href="medinform_v14i1e83745_fig8.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure9" position="float">
          <label>Figure 9</label>
          <caption>
            <p>Retrieval-augmented generation retrieval prompt template. Designed for a regulatory expert role aligned with the National Medical Products Administration, it defines 4 mandatory rules and standardized output requirements: conclusions must be based solely on retrieved regulatory provisions; each interpretation must include source citations; fabrication of provisions is prohibited; and any uncertainty must be labeled as “insufficient basis.” The template takes user queries and retrieved regulatory texts as input and produces standardized regulatory question-and-answer outputs.</p>
          </caption>
          <graphic xlink:href="medinform_v14i1e83745_fig9.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Optimal Parameter Configuration</title>
        <p>In typical small-sample medical data analysis tasks, <italic>r</italic>=32 has been validated as an effective LoRA range [<xref ref-type="bibr" rid="ref36">36</xref>]. Before experimentation, we initially favored selecting <italic>r</italic>=32. However, to rigorously validate the impact of rank-learning rate combinations on the performance of the infant incubator adverse event question-answering task, we conducted a 4 × 4 grid scan while fixing all other hyperparameters. The resulting <italic>F</italic><sub>1</sub>-scores are visualized as a heatmap, as shown in <xref rid="figure10" ref-type="fig">Figure 10</xref>. The heatmap shows that the <italic>F</italic><sub>1</sub>-score peaks at 0.823 at rank=32 and lr=1 × 10<sup>–4</sup>, then decreases toward the surrounding edges, exhibiting an overall bell-shaped “medium-high-medium” distribution. This phenomenon occurs because when the rank is too low, the low-rank matrix lacks expressive power and cannot fully capture the long-range dependencies and medical terminology details in event descriptions; when the rank is too high, the number of parameters increases, which can lead to overfitting on domain data and weaken generalization. Regarding the learning rate, 5 × 10<sup>–5</sup> converges too slowly, whereas 5 × 10<sup>–4</sup> has an excessively large step size, making it difficult for the model to stabilize at the optimal weights. By contrast, 1 × 10<sup>–4</sup> strikes a balance between convergence speed and stability. Therefore, the experiment ultimately adopted the combination of rank=32 and lr=1 × 10<sup>–4</sup>.</p>
        <fig id="figure10" position="float">
          <label>Figure 10</label>
          <caption>
            <p>Hyperparameter sensitivity heatmap. The figure shows variation in F1-scores on the test set for the infant incubator adverse event question-answering task as a function of low-rank adaptation rank and learning rate, evaluated via a 4 × 4 grid search.</p>
          </caption>
          <graphic xlink:href="medinform_v14i1e83745_fig10.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <p>To investigate the impact of batch size on training stability and performance, we conducted experiments while fixing the other hyperparameters. As shown in <xref rid="figure11" ref-type="fig">Figure 11</xref>A, when batch size is 1, the loss curve oscillates notably and converges slowly due to high gradient update noise; a batch size of 2 (the configuration selected for this study) achieves an optimal balance between training stability and convergence speed, with a smoothly declining loss curve and the lowest final convergence value; increasing the batch size to 4 or 8 yields a more stable training process but slightly slower convergence; a batch size of 16 fails to iterate sufficiently due to graphics processing unit memory constraints. <xref rid="figure11" ref-type="fig">Figure 11</xref>B further quantifies the final performance metrics across different batch sizes: a batch size of 2 achieves peak values for <italic>F</italic><sub>1</sub>-score, accuracy, and IDIAE, showing average improvements of 4.0% and 3.4%, compared with batch sizes of 1 and 8, respectively. Therefore, the final experimental configuration adopts a batch size of 2.</p>
        <fig id="figure11" position="float">
          <label>Figure 11</label>
          <caption>
            <p>Batch size impact analysis. The figure examines how different training batch sizes affect the AE2IN model’s training stability and final performance. (A) Training loss curves across epochs for various batch sizes, with epochs on the x-axis and loss on the y-axis. (B) The final F1-score, accuracy, and Information Density Index of Adverse Events (IDIAE) metric for each batch size, with batch size on the x-axis and metric values on the y-axis.</p>
          </caption>
          <graphic xlink:href="medinform_v14i1e83745_fig11.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <p>In supervised contrastive semantic optimization, the temperature coefficient <italic>τ</italic> controls the sharpness of the similarity distribution. To achieve optimal results, we fixed the other parameters and conducted experiments on the temperature coefficient. As shown in <xref rid="figure12" ref-type="fig">Figure 12</xref>, when <italic>τ</italic>=0.05, the similarity distribution becomes excessively sharp, causing the loss curve to oscillate and converge slowly. At <italic>τ</italic>=0.15 (the value selected for this study), the loss curve declines smoothly and rapidly to a stable region. When <italic>τ</italic> increases beyond 0.3, the similarity distribution becomes overly smooth, notably slowing loss convergence. Therefore, <italic>τ</italic>=0.15 is determined to be the optimal configuration for medical device regulatory retrieval tasks.</p>
        <fig id="figure12" position="float">
          <label>Figure 12</label>
          <caption>
            <p>Effect of the temperature coefficient (τ) on supervised contrastive learning. The figure shows how different τ values influence convergence of the training loss during supervised contrastive semantic optimization fine-tuning of the embedding model. The x-axis represents training steps, and the y-axis shows the contrastive learning loss.</p>
          </caption>
          <graphic xlink:href="medinform_v14i1e83745_fig12.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Ethical Considerations</title>
        <p>Our data originate from China’s medical device regulatory authorities. The data consist of medical device adverse event reports (similar to the US FDA’s Manufacturer and User Facility Device Experience [MAUDE] database). The content primarily includes failure information recorded during device use. These data contain no patient information; therefore, formal ethics approval is not required. This project was jointly completed with regulatory agencies under government-funded research initiatives. All data were authorized, and the research complies with relevant regulations.</p>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>Experimental Data</title>
        <p>This large-model experimental dataset comprises general-domain data and domain-specific data. The general-domain data originate from the PediaBench dataset publicly released by the State Key Laboratory of Public Big Data at the School of Computer Science and Technology, Guizhou University, China. It contains 1565 question-and-answer entries covering common pediatric diseases such as neonatal disorders, pediatric infectious diseases, pediatric respiratory diseases, and pediatric primary care and developmental behavioral disorders. Domain-specific data originate from private adverse event data on infant incubators within China’s Key Monitoring Information System for Medical Device Adverse Events. After manual and GPT-4 screening, data cleaning, and augmentation, combined with prompt engineering, the data were processed into a specialized question-and-answer dataset on infant incubator adverse events, totaling 2530 entries (<xref ref-type="table" rid="table1">Table 1</xref>).</p>
        <table-wrap position="float" id="table1">
          <label>Table 1</label>
          <caption>
            <p>Fine-tuning the Qwen model dataset sources and scale.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="280"/>
            <col width="570"/>
            <col width="150"/>
            <thead>
              <tr valign="top">
                <td>Dataset</td>
                <td>Data source</td>
                <td>Data volume</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>PediaBench</td>
                <td>Public dataset from the State Key Laboratory of Public Big Data, School of Computer Science and Technology, Guizhou University, China</td>
                <td>1565</td>
              </tr>
              <tr valign="top">
                <td>Infant Incubator Adverse Event Questions and Answers</td>
                <td>Adverse event data on private infant incubators in China’s Key Monitoring Information System for Medical Device Adverse Events</td>
                <td>2530</td>
              </tr>
            </tbody>
          </table>
        </table-wrap>
        <p>We split the data at the adverse-event report level: all question-and-answer pairs derived from the same report were assigned entirely to either the training, validation, or test set, ensuring that no report appears in more than one set. This experimental dataset allocates 60% (2457/4095) to the training set, 30% (1229/4095) to the test set, and 10% (409/4095) to the validation set.</p>
        <p>The knowledge base data originate from regulatory documents issued by the China National Medical Products Administration. We selected administrative regulations and departmental rules to form the regulatory provisions knowledge base, comprising a total of 29 documents. The fine-tuning dataset for the embedding model is derived from these regulatory documents. We utilized documents from this knowledge base and, combined with prompt engineering, leveraged GPT-4 to generate fine-tuning datasets. This dataset comprises 1488 entries, with 60% (893/1488) allocated to the training set, 30% (446/1488) to the test set, and 10% (149/1488) to the validation set (<xref ref-type="table" rid="table2">Table 2</xref>).</p>
        <table-wrap position="float" id="table2">
          <label>Table 2</label>
          <caption>
            <p>Knowledge base and fine-tuning the embedding model dataset.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="330"/>
            <col width="520"/>
            <col width="150"/>
            <thead>
              <tr valign="top">
                <td>Dataset</td>
                <td>Data source</td>
                <td>Data volume</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Regulatory Knowledge Base</td>
                <td>Regulatory documents issued by the China National Medical Products Administration</td>
                <td>29 documents</td>
              </tr>
              <tr valign="top">
                <td>Embedding Model Fine-Tuning Dataset</td>
                <td>Generated using regulatory documents issued by the China National Medical Products Administration</td>
                <td>1488</td>
              </tr>
            </tbody>
          </table>
        </table-wrap>
        <p>To mitigate the “student-teacher” loop bias introduced by large-model rewriting, we collaborated with the Henan Drug Administration to implement a manual quality control process for all data generated by large models. First, we established a 3-tier annotation standard (perfect equivalence/minor drift/significant drift) and trained annotators (<italic>κ</italic>≥0.78). Next, data were grouped into batches of 100 items for double-blind annotation by pairs of annotators, with a third party serving as the arbitrator. Only when the “significant drift” rate was controlled at ≤5% were the data released. Finally, senior experts centrally reviewed and refined the “significant drift” data.</p>
      </sec>
      <sec>
        <title>Experimental Environment</title>
        <p>This work selected Qwen2_7B as the source model and employed a dual-adapter joint fine-tuning approach for training, with specific configurations detailed in <xref ref-type="table" rid="table3">Tables 3</xref>-<xref ref-type="table" rid="table5">5</xref>.</p>
        <table-wrap position="float" id="table3">
          <label>Table 3</label>
          <caption>
            <p>Environmental configuration.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="500"/>
            <col width="500"/>
            <thead>
              <tr valign="top">
                <td>Category</td>
                <td>Configuration information</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Central processing unit</td>
                <td>Xeon(R) Platinum 8352V</td>
              </tr>
              <tr valign="top">
                <td>Graphics processing unit</td>
                <td>RTX 4090 × 2</td>
              </tr>
              <tr valign="top">
                <td>Graphics memory</td>
                <td>24 GB</td>
              </tr>
              <tr valign="top">
                <td>Distributed training strategy</td>
                <td>Data parallelism</td>
              </tr>
              <tr valign="top">
                <td>Memory</td>
                <td>90 GB</td>
              </tr>
              <tr valign="top">
                <td>Compute Unified Device Architecture</td>
                <td>12.2</td>
              </tr>
            </tbody>
          </table>
        </table-wrap>
        <table-wrap position="float" id="table4">
          <label>Table 4</label>
          <caption>
            <p>Hyperparameter configuration.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="500"/>
            <col width="500"/>
            <thead>
              <tr valign="top">
                <td>Hyperparameter</td>
                <td>Value</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Train epochs</td>
                <td>3</td>
              </tr>
              <tr valign="top">
                <td>Learning rate</td>
                <td>1.00 × 10<sup>–4</sup></td>
              </tr>
              <tr valign="top">
                <td>Batch size</td>
                <td>2</td>
              </tr>
              <tr valign="top">
                <td>Lora_rank</td>
                <td>32</td>
              </tr>
              <tr valign="top">
                <td>Logging_steps</td>
                <td>3</td>
              </tr>
              <tr valign="top">
                <td>Gradient accumulation</td>
                <td>4</td>
              </tr>
            </tbody>
          </table>
        </table-wrap>
        <table-wrap position="float" id="table5">
          <label>Table 5</label>
          <caption>
            <p>Retrieval-augmented generation parameter configuration.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="500"/>
            <col width="500"/>
            <thead>
              <tr valign="top">
                <td>Parameters</td>
                <td>Value</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Chunk size</td>
                <td>512</td>
              </tr>
              <tr valign="top">
                <td>Overlap</td>
                <td>80</td>
              </tr>
              <tr valign="top">
                <td>Top-<italic>k</italic></td>
                <td>5</td>
              </tr>
              <tr valign="top">
                <td>Cosine</td>
                <td>0.72</td>
              </tr>
              <tr valign="top">
                <td>Nlist</td>
                <td>128</td>
              </tr>
            </tbody>
          </table>
        </table-wrap>
      </sec>
      <sec>
        <title>Base Model Testing</title>
        <p>This study selected BGE-Large-ZH-V1.5 as the embedding source model. Here, BGE-Large-ZH-V1.5 was compared with M3E-Large, Text2Vec-Large, and Luotuo on the Massive Text Embedding Benchmark (MTEB) and the Chinese Massive Text Embedding Benchmark (C-MTEB). The embedding dimension was unified to 1024, and the evaluation metrics selected were retrieval, semantic textual similarity, pairwise classification, and classification.</p>
        <p><xref ref-type="table" rid="table6">Table 6</xref> compares BGE-Large-ZH-V1.5 with mainstream models, demonstrating that it achieves leading performance on both MTEB and C-MTEB. Its high recall score of 70.46 indicates that it is particularly well-suited as a foundation for RAG systems, demonstrating a superior ability to accurately match specialized terminology within knowledge bases compared with other models.</p>
        <p>For large models, this paper compares base-model performance by evaluating Qwen2_7B, Llama3_8B, and GLM4_9B across 8 metrics. The results are shown in <xref ref-type="table" rid="table7">Table 7</xref>.</p>
        <table-wrap position="float" id="table6">
          <label>Table 6</label>
          <caption>
            <p>Embedding base model test results. This table compares the performance of the BGE-Large-ZH-V1.5 embedding model selected for this work against 3 mainstream Chinese embedding models across multiple dimensions on the MTEBa and C-MTEBb standard evaluation datasets.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="180"/>
            <col width="170"/>
            <col width="180"/>
            <col width="130"/>
            <col width="190"/>
            <col width="150"/>
            <thead>
              <tr valign="top">
                <td>Model</td>
                <td>Embedded dimension</td>
                <td>Retrospective retrieval</td>
                <td>Semantic textual similarity</td>
                <td>Pairwise classification</td>
                <td>Classification</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>BGE-Large-ZH-V1.5</td>
                <td>1024</td>
                <td>70.46</td>
                <td>56.25</td>
                <td>81.6</td>
                <td>69.13</td>
              </tr>
              <tr valign="top">
                <td>M3e-Large</td>
                <td>1024</td>
                <td>54.75</td>
                <td>50.42</td>
                <td>64.3</td>
                <td>68.2</td>
              </tr>
              <tr valign="top">
                <td>Text2Vec-Large</td>
                <td>1024</td>
                <td>41.94</td>
                <td>44.97</td>
                <td>70.86</td>
                <td>60.66</td>
              </tr>
              <tr valign="top">
                <td>luotuo</td>
                <td>1024</td>
                <td>44.4</td>
                <td>42.78</td>
                <td>66.62</td>
                <td>61</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table6fn1">
              <p><sup>a</sup>MTEB: Massive Text Embedding Benchmark.</p>
            </fn>
            <fn id="table6fn2">
              <p><sup>b</sup>C-MTEB: Chinese Massive Text Embedding Benchmark.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <table-wrap position="float" id="table7">
          <label>Table 7</label>
          <caption>
            <p>Base model test results. This table presents the comparison results of general-purpose models, specialized medical models (HuatuoGPT-7B and DISC-MedLLM), and AE2IN on benchmark metric tests. This evaluation verifies AE2IN’s ability to adapt to specialized domains while preserving foundational general capabilities and assesses its mitigation of catastrophic forgetting.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="200"/>
            <col width="110"/>
            <col width="100"/>
            <col width="80"/>
            <col width="100"/>
            <col width="100"/>
            <col width="110"/>
            <col width="80"/>
            <col width="120"/>
            <thead>
              <tr valign="bottom">
                <td>Indicator/model</td>
                <td>AlignBench</td>
                <td>MT-Bench</td>
                <td>MMLU<sup>a</sup></td>
                <td>GSM8K</td>
                <td>MATH</td>
                <td>HumanEval</td>
                <td>C-Eval</td>
                <td>PediaBench</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Qwen2_7B</td>
                <td>7.21</td>
                <td>8.41</td>
                <td>70.5</td>
                <td>82.3</td>
                <td>49.6</td>
                <td>79.9</td>
                <td>77.2</td>
                <td>79.7</td>
              </tr>
              <tr valign="top">
                <td>Llama3_8B</td>
                <td>6.20</td>
                <td>8.05</td>
                <td>68.4</td>
                <td>79.6</td>
                <td>30.0</td>
                <td>62.2</td>
                <td>45.9</td>
                <td>59.4</td>
              </tr>
              <tr valign="top">
                <td>GLM4_9B</td>
                <td>7.01</td>
                <td>8.35</td>
                <td>72.4</td>
                <td>79.6</td>
                <td>50.6</td>
                <td>71.8</td>
                <td>75.6</td>
                <td>71.5</td>
              </tr>
              <tr valign="top">
                <td>HuatuoGPT-7B</td>
                <td>6.09</td>
                <td>8.01</td>
                <td>65.1</td>
                <td>77.5</td>
                <td>28.3</td>
                <td>59.3</td>
                <td>80.7</td>
                <td>88.7</td>
              </tr>
              <tr valign="top">
                <td>DISC-MedLLM</td>
                <td>7.30</td>
                <td>8.22</td>
                <td>67.3</td>
                <td>76.1</td>
                <td>45.4</td>
                <td>62.1</td>
                <td>82.4</td>
                <td>90.6</td>
              </tr>
              <tr valign="top">
                <td>AE2IN</td>
                <td>7.32</td>
                <td>8.44</td>
                <td>67.8</td>
                <td>79.2</td>
                <td>46.3</td>
                <td>64.8</td>
                <td>75.1</td>
                <td>80.1</td>
              </tr>
              <tr valign="top">
                <td>Only fine-tuning for specific domains</td>
                <td>7.08</td>
                <td>8.21</td>
                <td>62.5</td>
                <td>75.3</td>
                <td>39.8</td>
                <td>59.4</td>
                <td>71.8</td>
                <td>73.5</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table7fn1">
              <p><sup>a</sup>MMLU: Massive Multitask Language Understanding.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <p><xref ref-type="table" rid="table7">Table 7</xref> shows that the Qwen2 model has stronger overall performance than the other 2 models. Qwen2 performs well on the Massive Multitask Language Understanding (MMLU) medical subset and C-Eval medical questions, indicating that the model is stable in diagnostic reasoning and clinical terminology comprehension, but exhibits slight errors in complex calculation questions. Meanwhile, the fine-tuned AE2IN model shows score decreases ranging from 2.1 to 2.7 on the MMLU medical subset and C-Eval medical tasks. Because of the incorporation of PediaBench data during fine-tuning, its scores slightly surpass those of the base model by 0.4 points. By contrast, models fine-tuned solely on domain-specific datasets experienced score declines ranging from 5.4 to 8.0 on the MMLU medical subset and C-Eval medical tasks, along with a 6.2-point drop on PediaBench. These comparative results validate that AE2IN effectively mitigates catastrophic forgetting.</p>
      </sec>
      <sec>
        <title>Comparative Experiment</title>
        <p>We selected 5 models for comparative testing with the fine-tuned model. The results are shown in <xref ref-type="table" rid="table8">Table 8</xref>.</p>
        <table-wrap position="float" id="table8">
          <label>Table 8</label>
          <caption>
            <p>Comparative evaluation of the fine-tuned model AE2IN against general models and specialized medical models (HuatuoGPT, DISC-MedLLM) in the infant incubator adverse event domain. This table compares the overall performance of the base large model, the base model combined with the same RAGa framework model, mainstream medical-specific large models, and the AE2IN model proposed in this study on the infant incubator adverse event analysis task. All experiments were replicated 3 times independently. Results are presented as mean (SD), supplemented with 2-tailed (unpaired and independent) samples t test <italic>P</italic> values and statistical significance annotations for each model versus AE2IN.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="120"/>
            <col width="100"/>
            <col width="70"/>
            <col width="90"/>
            <col width="80"/>
            <col width="90"/>
            <col width="80"/>
            <col width="90"/>
            <col width="70"/>
            <col width="130"/>
            <col width="80"/>
            <thead>
              <tr valign="bottom">
                <td>Model</td>
                <td>BLEU<sup>b</sup>-4</td>
                <td><italic>P</italic> value</td>
                <td>ROUGE-L<sup>c</sup></td>
                <td><italic>P</italic> value</td>
                <td>Accuracy</td>
                <td><italic>P</italic> value</td>
                <td>Element recall rate</td>
                <td><italic>P</italic> value</td>
                <td>Average IDIAE<sup>d</sup></td>
                <td><italic>P</italic> value</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Qwen2_7B</td>
                <td>0.137 (0.031)</td>
                <td>&lt;.001</td>
                <td>0.208 (0.025)</td>
                <td>&lt;.001</td>
                <td>0.785 (0.028)</td>
                <td>&lt;.001</td>
                <td>0.691 (0.023)</td>
                <td>&lt;.001</td>
                <td>0.61 (0.018)</td>
                <td>&lt;.001</td>
              </tr>
              <tr valign="top">
                <td>Qwen2_7B+RAG</td>
                <td>0.162 (0.025)</td>
                <td>.003</td>
                <td>0.315 (0.022)</td>
                <td>.001</td>
                <td>0.800 (0.024)</td>
                <td>.002</td>
                <td>0.715 (0.019)</td>
                <td>.002</td>
                <td>0.75 (0.015)</td>
                <td>.002</td>
              </tr>
              <tr valign="top">
                <td>Llama3_8B</td>
                <td>0.108 (0.026)</td>
                <td>&lt;.001</td>
                <td>0.237 (0.033)</td>
                <td>&lt;.001</td>
                <td>0.722 (0.030)</td>
                <td>&lt;.001</td>
                <td>0.651 (0.028)</td>
                <td>&lt;.001</td>
                <td>0.49 (0.025)</td>
                <td>&lt;.001</td>
              </tr>
              <tr valign="top">
                <td>Llama3_8B+RAG</td>
                <td>0.131 (0.022)</td>
                <td>.003</td>
                <td>0.302 (0.028)</td>
                <td>.002</td>
                <td>0.756 (0.026)</td>
                <td>.003</td>
                <td>0.688 (0.024)</td>
                <td>.002</td>
                <td>0.62 (0.019)</td>
                <td>.003</td>
              </tr>
              <tr valign="top">
                <td>GLM4_9B</td>
                <td>0.105 (0.025)</td>
                <td>&lt;.001</td>
                <td>0.282 (0.033)</td>
                <td>&lt;.001</td>
                <td>0.763 (0.031)</td>
                <td>&lt;.001</td>
                <td>0.636 (0.025)</td>
                <td>&lt;.001</td>
                <td>0.55 (0.020)</td>
                <td>&lt;.001</td>
              </tr>
              <tr valign="top">
                <td>GLM4_9B+RAG</td>
                <td>0.158 (0.021)</td>
                <td>.002</td>
                <td>0.356 (0.025)</td>
                <td>.001</td>
                <td>0.801 (0.023)</td>
                <td>.002</td>
                <td>0.702 (0.021)</td>
                <td>.001</td>
                <td>0.71 (0.017)</td>
                <td>.002</td>
              </tr>
              <tr valign="top">
                <td>HuatuoGPT-7B</td>
                <td>0.155 (0.021)</td>
                <td>.03</td>
                <td>0.381 (0.022)</td>
                <td>.03</td>
                <td>0.803 (0.025)</td>
                <td>.03</td>
                <td>0.722 (0.031)</td>
                <td>.03</td>
                <td>0.80 (0.022)</td>
                <td>.03</td>
              </tr>
              <tr valign="top">
                <td>DISC-MedLLM</td>
                <td>0.174 (0.020)</td>
                <td>.02</td>
                <td>0.411 (0.024)</td>
                <td>.02</td>
                <td>0.816 (0.022)</td>
                <td>.02</td>
                <td>0.740 (0.020)</td>
                <td>.02</td>
                <td>0.82 (0.019)</td>
                <td>.02</td>
              </tr>
              <tr valign="top">
                <td>AE2IN</td>
                <td>0.215 (0.018)</td>
                <td>N/A<sup>e</sup></td>
                <td>0.484 (0.021)</td>
                <td>N/A</td>
                <td>0.898 (0.022)</td>
                <td>N/A</td>
                <td>0.815 (0.016)</td>
                <td>N/A</td>
                <td>0.91 (0.013)</td>
                <td> N/A</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table8fn1">
              <p><sup>a</sup>RAG: retrieval-augmented generation.</p>
            </fn>
            <fn id="table8fn2">
              <p><sup>b</sup>BLEU: Bilingual Evaluation Understudy.</p>
            </fn>
            <fn id="table8fn3">
              <p><sup>c</sup>ROUGE-L: Recall-Oriented Understudy for Gisting Evaluation - Longest Common Subsequence.</p>
            </fn>
            <fn id="table8fn4">
              <p><sup>d</sup>IDIAE: Information Density Index of Adverse Events.</p>
            </fn>
            <fn id="table8fn5">
              <p><sup>e</sup>N/A: not applicable.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <p>Through systematic quantitative evaluation, the AE2IN model demonstrates notable advantages across all key performance indicators in the domain of infant incubator adverse events. In terms of text generation quality, AE2IN attained a BLEU-4 score of 0.215, reflecting a 56.9% improvement relative to the baseline model and a 32.7% improvement over the baseline equipped with the same RAG framework. Its Recall-Oriented Understudy for Gisting Evaluation - Longest Common Subsequence (ROUGE-L) score reached 0.484, corresponding to a 53.7% increase. For core element analysis of infant incubator adverse events, the element recall rate reached 0.815, representing a 17.9% improvement over the baseline model and a 14.0% improvement over the baseline integrated with the same RAG framework. In the experiments, the dual-adapter architecture exhibited strong causal reasoning capacity when processing complex alarm system failure events.</p>
        <p>To conduct a more comprehensive analysis of the AE2IN model, we also selected cases in which accuracy, element recall, and IDIAE were all below 0.5. We found that misjudgment of core elements directly affected completeness conclusions and notably impacted both accuracy and IDIAE values. In failed cases, we observed that for rare fault descriptions—such as the onomatopoeic “pushing clack-clack sound”—low frequency and susceptibility to semantic integrity disruption caused the inferred Query-Key similarity to be drowned out by the attention distribution of primary terms after softmax. This prevented the activation of corresponding neurons during value vector weighting. Furthermore, the scaling coefficient update for (IA)<sup>3</sup> relies on batch statistics, leading to inconsistent weighting across cases. When encountering such descriptions, the update process may become stuck at a saddle point, failing to amplify this feature.</p>
        <p>We continuously record training and validation loss, along with key evaluation metrics, to determine in real time whether the model is in a state of normal convergence, overfitting, or underfitting. As shown in <xref rid="figure13" ref-type="fig">Figure 13</xref>, validation loss and training loss decrease smoothly and synchronously over the first 600 steps, maintaining a small difference of less than 0.15, indicating that the learning rate of 1 × 10<sup>–4</sup> effectively balances model capacity and regularization. BLEU-4 increases monotonically from 0.07 to 0.22, only 0.005 away from the final test set score of 0.215, confirming that the validation and test sets are identically distributed and that the model has fully converged. Feature recall also increases from 0.60 to 0.815 over the first 600 steps, rising in parallel with BLEU-4 without intersecting, indicating that the model improves fluency without sacrificing its ability to extract key information. All 3 metrics plateau simultaneously, demonstrating stable convergence without the need for early stopping or secondary parameter tuning. This also indicates that the final test score is not a random peak but the result of stable convergence.</p>
        <fig id="figure13" position="float">
          <label>Figure 13</label>
          <caption>
            <p>Convergence of dual-adapter fine-tuning (AE2IN). The figure shows the joint evolution of training loss, validation loss, Bilingual Evaluation Understudy-4 score, and ERR metric over training steps. Core hyperparameters were fixed (low-rank adaptation rank=32, learning rate=1 × 10–4, batch size=2). The x-axis represents training steps; the left y-axis shows loss values, and the right y-axis shows text generation and key element extraction scores.</p>
          </caption>
          <graphic xlink:href="medinform_v14i1e83745_fig13.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <p>As shown in <xref rid="figure14" ref-type="fig">Figure 14</xref>, when determining the completeness of adverse events related to infant incubators, AE2IN directly extracts core elements—incident date, nature of the abnormality, and subsequent corrective actions—while supplementing them with details such as operator information, risk warnings, and injury status. This approach concisely highlights essential information.</p>
        <fig id="figure14" position="float">
          <label>Figure 14</label>
          <caption>
            <p>AE2IN response examples. The figure presents sample outputs from AE2IN, including assessments of adverse event report completeness for infant incubators and responses to medical device adverse event definitions.</p>
          </caption>
          <graphic xlink:href="medinform_v14i1e83745_fig14.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <p>However, compared with other models, AE2IN’s responses tend to be more concise. To investigate whether this brevity affects safety, a supplementary risk recall assessment experiment was conducted to quantify the quality of model outputs from a safety perspective. First, a core safety item list was established through expert consultation and alignment with the “Measures for the Administration of Medical Device Adverse Event Monitoring and Re-evaluation” and the national standard “GB 9706.219-2021.” This list encompasses adverse event phenomenon identification, safety risk identification, immediate patient safety measures, equipment disposal measures, adverse event reporting measures, and risk cause localization. Next, 100 high-quality data entries, each containing core safety items, were extracted and annotated, and the process was repeated 5 times. The results of the risk recall experiment are shown in <xref ref-type="table" rid="table9">Table 9</xref>.</p>
        <table-wrap position="float" id="table9">
          <label>Table 9</label>
          <caption>
            <p>Risk recall rate comparison. This table displays the risk recall rates and the average number of core safety items identified by AE2IN across various models.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="210"/>
            <col width="290"/>
            <col width="500"/>
            <thead>
              <tr valign="top">
                <td>Model</td>
                <td>Risk recall, mean (SD)</td>
                <td>Average number of core safety items identified</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>AE2IN</td>
                <td>0.92 (0.012)</td>
                <td>5.52</td>
              </tr>
              <tr valign="top">
                <td>Qwen2_7B</td>
                <td>0.77 (0.024)</td>
                <td>4.62</td>
              </tr>
              <tr valign="top">
                <td>Llama3_8B</td>
                <td>0.75 (0.022)</td>
                <td>4.50</td>
              </tr>
              <tr valign="top">
                <td>GLM4_9B</td>
                <td>0.80 (0.019)</td>
                <td>4.80</td>
              </tr>
              <tr valign="top">
                <td>HuatuoGPT-7B</td>
                <td>0.82 (0.016)</td>
                <td>4.92</td>
              </tr>
              <tr valign="top">
                <td>DISC-MedLLM</td>
                <td>0.84 (0.017)</td>
                <td>5.04</td>
              </tr>
            </tbody>
          </table>
        </table-wrap>
        <p>AE2IN demonstrates the highest risk recall rate, achieving an improvement of approximately 15% over other models. Although some data may be underreported due to rare fault descriptions, its overall risk recall rate confirms that AE2IN effectively maximizes risk extraction. By outputting structured information that highlights core details, it helps ensure safety.</p>
        <p>We used the AE2IN model with the same RAG framework and different embedding models for regulatory question answering. The experiments used a regulatory clause validation set to compare the performance of the fine-tuned embedding models. The comparative results are shown in <xref ref-type="table" rid="table1">Table 10</xref>.</p>
        <table-wrap position="float" id="table10">
          <label>Table 10</label>
          <caption>
            <p>Comparative evaluation of regulatory question-answering performance between the AE2IN model with different embedding models and general-purpose models.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="200"/>
            <col width="100"/>
            <col width="100"/>
            <col width="100"/>
            <col width="100"/>
            <col width="100"/>
            <col width="0"/>
            <col width="100"/>
            <col width="100"/>
            <col width="100"/>
            <thead>
              <tr valign="top">
                <td rowspan="2">Configuration/indicators</td>
                <td colspan="6">AE2IN</td>
                <td colspan="3">General model</td>
              </tr>
              <tr valign="top">
                <td>FINBGE<sup>a</sup> (ours)</td>
                <td>BGE-Large-ZH-V1.5</td>
                <td>M3e-Large</td>
                <td>Text2Vec-Large</td>
                <td>Luotuo</td>
                <td colspan="2">Qwen2_7B</td>
                <td>Llama3_8B</td>
                <td>GLM4_9B</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Accuracy, mean (SD)</td>
                <td>0.938 (0.011)</td>
                <td>0.863 (0.015)</td>
                <td>0.804 (0.015)</td>
                <td>0.780 (0.016)</td>
                <td>0.761 (0.017)</td>
                <td colspan="2">0.715 (0.038)</td>
                <td>0.652 (0.040)</td>
                <td>0.727  0.042)</td>
              </tr>
              <tr valign="top">
                <td>Recall, mean (SD)</td>
                <td>0.894 (0.013)</td>
                <td>0.807 (0.016)</td>
                <td>0.763 (0.017)</td>
                <td>0.759 (0.017)</td>
                <td>0.685 (0.019)</td>
                <td colspan="2">0.652 (0.042)</td>
                <td>0.594 (0.044)</td>
                <td>0.671 (0.050)</td>
              </tr>
              <tr valign="top">
                <td><italic>F</italic><sub>1</sub>-score, mean (SD)</td>
                <td>0.915 (0.013)</td>
                <td>0.834 (0.016)</td>
                <td>0.783 (0.016)</td>
                <td>0.769 (0.017)</td>
                <td>0.721 (0.018)</td>
                <td colspan="2">0.682 (0.041)</td>
                <td>0.622 (0.038)</td>
                <td>0.698 (0.04)</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table10fn1">
              <p><sup>a</sup>FINBGE is a BGE model fine-tuned using supervised contrastive learning.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <p>The experimental results in <xref ref-type="table" rid="table1">Table 11</xref> show that the AE2IN model, when equipped with different embedding models, exhibits a clear performance gradient in the regulatory question-answering task. The FINBGE model leads by a notable margin, achieving the highest levels across the 3 core metrics—accuracy, recall, and <italic>F</italic><sub>1</sub>-score—compared with the source model BGE-Large-ZH-V1.5. This result validates the effectiveness of the supervised contrastive semantic optimization fine-tuning strategy, which successfully narrows the vector distance between the question and the correct answer. Furthermore, comparison with common models reveals that Llama3_8B achieves the lowest scores, likely because the test dataset consists entirely of Chinese regulatory question-answer pairs. The best-performing general model, GLM4_9B, achieves scores of 0.727, 0.671, and 0.698 in accuracy, recall, and <italic>F</italic><sub>1</sub>-score, respectively, but still lags behind AE2IN. This demonstrates the necessity of domain adaptation and the RAG architecture, which can mitigate model hallucinations, address knowledge timeliness, and improve accuracy.</p>
        <table-wrap position="float" id="table11">
          <label>Table 11</label>
          <caption>
            <p>AE2IN model performance ablation study.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="210"/>
            <col width="130"/>
            <col width="160"/>
            <col width="190"/>
            <col width="150"/>
            <col width="160"/>
            <thead>
              <tr valign="bottom">
                <td>Indicators/configuration</td>
                <td>BLEU<sup>a</sup>-4, mean (SD)</td>
                <td>ROUGE-L<sup>b</sup>, mean (SD)</td>
                <td>Accuracy, mean (SD)</td>
                <td>Element recall rate, mean (SD)</td>
                <td>IDIAE<sup>c</sup>, mean (SD)</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>AE2IN</td>
                <td>0.215 (0.018)</td>
                <td>0.484 (0.021)</td>
                <td>0.898 (0.022)</td>
                <td>0.815 (0.016)</td>
                <td>0.91 (0.01)</td>
              </tr>
              <tr valign="top">
                <td>Only LoRA<sup>d</sup></td>
                <td>0.187 (0.015)</td>
                <td>0.412 (0.019)</td>
                <td>0.824 (0.020)</td>
                <td>0.763 (0.015)</td>
                <td>0.83 (0.01)</td>
              </tr>
              <tr valign="top">
                <td>Only (IA)<sup>3e</sup></td>
                <td>0.176 (0.015)</td>
                <td>0.398 (0.017)</td>
                <td>0.795 (0.018)</td>
                <td>0.741 (0.014)</td>
                <td>0.79 (0.01)</td>
              </tr>
              <tr valign="top">
                <td>Full fine-tuning</td>
                <td>0.168 (0.022)</td>
                <td>0.388 (0.024)</td>
                <td>0.802 (0.028)</td>
                <td>0.735 (0.021)</td>
                <td>0.78 (0.02)</td>
              </tr>
              <tr valign="top">
                <td>None</td>
                <td>0.137 (0.031)</td>
                <td>0.208 (0.025)</td>
                <td>0.785 (0.028)</td>
                <td>0.691 (0.023)</td>
                <td>0.61 (0.02)</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table11fn1">
              <p><sup>a</sup>BLEU: Bilingual Evaluation Understudy.</p>
            </fn>
            <fn id="table11fn2">
              <p><sup>b</sup>ROUGE-L: Recall-Oriented Understudy for Gisting Evaluation - Longest Common Subsequence.</p>
            </fn>
            <fn id="table11fn3">
              <p><sup>c</sup>IDIAE: Information Density Index of Adverse Events.</p>
            </fn>
            <fn id="table11fn4">
              <p><sup>d</sup>LoRA: low-rank adaptation.</p>
            </fn>
            <fn id="table11fn5">
              <p><sup>e</sup>(IA)<sup>3</sup>: Infused Adapter by Inhibiting and Amplifying Inner Activations.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <p>Similarly, to conduct a more comprehensive analysis of the system model, we selected cases in which regulatory question-answering accuracy was below 0.5 for further study. We found that hallucinations still occurred—for instance, when the model should have referenced article 23 of the “Administrative Measures for the Monitoring and Re-evaluation of Adverse Events of Medical Devices,” it instead cited article 24. This occurs because clauses with high similarity but different article numbers exhibit only minor differences in cosine similarity, leading to the retrieval of incorrect clauses and subsequently triggering hallucinations.</p>
      </sec>
      <sec>
        <title>Ablation Experiments</title>
        <p>To validate the contributions of each component in the dual-adapter joint fine-tuning architecture, this study designed a systematic ablation experiment, maintaining the same testing environment and hyperparameter settings as previously described. The experiment employed a controlled variable approach, sequentially removing key components and observing changes in model performance.</p>
        <p><xref ref-type="table" rid="table1">Table 11</xref> shows that the complete model improves feature recall by an average of 9.25% compared with the single-adapter solution, validating the complementarity of low-rank updates and feature scaling in equation (1).</p>
        <p>When the LoRA component is removed, BLEU-4 and ROUGE-L decrease notably. This is mainly because LoRA optimizes the long-range dependency modeling capability of the Transformer layer through low-rank decomposition. Its absence makes it difficult for the model to effectively capture the logic in infant incubator failure events. At the same time, the feature recall rate decreases by 10.0%, further validating LoRA’s key role in structured information extraction—its trainable parameters are concentrated in the Δ<italic>W</italic> matrix of the attention layer, enabling more accurate localization of core elements in fault descriptions.</p>
        <p>When the (IA)<sup>3</sup> component is removed, the IDIAE index decreases notably, due to the weakening of the dynamic feature-scaling function achieved by (IA)<sup>3</sup> through the gating mechanism. Specifically, the key-value intervention vector in (IA)<sup>3</sup> suppresses the activation of irrelevant features while enhancing the response strength of neurons associated with medical terms.</p>
        <p>Following full fine-tuning, metrics exhibited varying degrees of decline compared with AE2IN or single-method fine-tuning. This phenomenon stems from the inevitable overfitting of the 7-billion-parameter model on finite, small samples, validating the necessity of structured parameter fine-tuning for efficiency in small-sample scenarios.</p>
        <p>Observation of <xref ref-type="table" rid="table1">Table 12</xref> data reveals that, in long-text scenarios, the recall rate of the LoRA configuration alone exhibits a lower decay rate than (IA)<sup>3</sup>, demonstrating that LoRA’s low-rank updates effectively capture long-range semantic associations across clauses. <xref rid="figure15" ref-type="fig">Figure 15</xref> further shows that the attention weights in the LoRA branch decay by 16% for long-distance word pairs, representing a 14-percentage-point reduction compared with the baseline. This indicates that LoRA effectively enhances the interaction strength of distant semantic units. (IA)<sup>3</sup> can boost weights over short distances, but long-range dependencies are easily lost, leading to reduced recall rates. This also explains why, under dual-adapter collaboration, the final performance in the infant incubator adverse-event domain improved by approximately 14%, further demonstrating that the 2 adapters can complement each other across different distance scales.</p>
        <table-wrap position="float" id="table12">
          <label>Table 12</label>
          <caption>
            <p>AE2IN element recall rate ablation experiment results at different input lengths. This table compares the element recall rate of different model architectures across short and long text adverse event reports for infant incubators.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="240"/>
            <col width="260"/>
            <col width="300"/>
            <col width="200"/>
            <thead>
              <tr valign="top">
                <td>Configuration</td>
                <td>Short text (≤64 tokens), mean (SD)</td>
                <td>Long text (&gt;128 tokens), mean (SD)</td>
                <td>Attenuation level (%)</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>AE2IN</td>
                <td>0.847 (0.012)</td>
                <td>0.810 (0.019)</td>
                <td>4.4 (1.1)</td>
              </tr>
              <tr valign="top">
                <td>Only LoRA<sup>a</sup></td>
                <td>0.834 (0.015)</td>
                <td>0.766 (0.024)</td>
                <td>8.2 (1.8)</td>
              </tr>
              <tr valign="top">
                <td>Only (IA)<sup>3b</sup></td>
                <td>0.829 (0.018)</td>
                <td>0.739 (0.029)</td>
                <td>10.9 (2.2)</td>
              </tr>
              <tr valign="top">
                <td>Full fine-tuning</td>
                <td>0.781 (0.021)</td>
                <td>0.704 (0.032)</td>
                <td>9.9 (2.6)</td>
              </tr>
              <tr valign="top">
                <td>None</td>
                <td>0.733 (0.028)</td>
                <td>0.661 (0.035)</td>
                <td>9.7 (3.1)</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table12fn1">
              <p><sup>a</sup>LoRA: low-rank adaptation.</p>
            </fn>
            <fn id="table12fn2">
              <p><sup>b</sup>(IA)<sup>3</sup>: Infused Adapter by Inhibiting and Amplifying Inner Activations.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <fig id="figure15" position="float">
          <label>Figure 15</label>
          <caption>
            <p>Dual-adapter semantic modeling capability. The figure demonstrates the model’s ability to capture semantics using a long-text test dataset of infant incubator adverse events. It shows the decay of attention weights with increasing token distance: the x-axis represents token distance, and the y-axis shows normalized attention weights, comparing long- and short-range semantic capture across four architectures. (IA)3: Infused Adapter by Inhibiting and Amplifying Inner Activations; LoRA: low-rank adaptation.</p>
          </caption>
          <graphic xlink:href="medinform_v14i1e83745_fig15.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <sec>
        <title>Principal Findings</title>
        <p>This study proposes AE2IN, a large-model framework that integrates RAG with dual-adapter joint fine-tuning for the intelligent analysis of adverse events in infant incubators. Experimental results demonstrate that this framework significantly outperforms baseline models across key metrics, including text generation quality, element recall rate, and regulatory question-and-answer accuracy. Under the same experimental conditions, the element recall rate reaches 0.815, the accuracy of infant incubator adverse event analysis is 0.898, and the accuracy in regulatory clause question-answering tasks attains 0.938. The dual-adapter architecture, that is, LoRA + (IA)<sup>3</sup>, effectively infuses domain knowledge while preserving model generalization, thereby mitigating catastrophic forgetting. The RAG module, combined with the FINBGE embedding model optimized through supervised contrastive semantic learning, significantly enhances knowledge retrieval accuracy and reduces hallucination.</p>
      </sec>
      <sec>
        <title>Comparison With Prior Work</title>
        <p>Compared with traditional deep learning models, AE2IN demonstrates superior contextual modeling and causal reasoning capabilities, making it particularly well-suited for complex tasks involving multiple intertwined factors and specialized terminology, such as adverse events in infant incubators. Compared with single fine-tuning approaches, the dual-adapter architecture proposed in this paper achieves a better balance between parameter efficiency and task adaptability. Compared with existing RAG medical applications, AE2IN further introduces a domain-specific embedding model fine-tuning mechanism, enhancing the matching accuracy of Chinese regulatory clauses. When benchmarked against general models such as Qwen and Llama, AE2IN demonstrates higher professional consistency and stronger element-extraction capabilities in specialized tasks, particularly excelling in structured output and terminology standardization.</p>
      </sec>
      <sec>
        <title>Limitations</title>
        <p>First, the dataset size is limited. The infant incubator adverse-event dataset used in this study comprises 2530 entries. Despite undergoing joint cleaning and enhancement by human experts and GPT-4, the dataset remains insufficient for rare fault types or complex scenarios. To mitigate this issue, we introduced PediaBench general pediatric data for mixed training, employing a 3:1 domain-to-general data sampling ratio to enhance generalization. Future efforts could expand multicenter, multiyear data through collaborations with additional medical institutions to further improve model robustness. Second, the regulatory knowledge base updates lag. The regulatory knowledge base used in this study cannot guarantee real-time coverage of the latest policy revisions. We partially mitigated this knowledge lag by improving the embedding model’s semantic understanding of clauses through supervised contrastive fine-tuning. Future work should consider establishing a dynamic regulatory update mechanism, such as periodically crawling the National Medical Products Administration website or integrating official application programming interfaces to enable automatic knowledge base synchronization. Third, prospective clinical validation was not conducted. This retrospective study did not verify the model’s usability and safety within actual clinical workflows. We collaborated with the Henan Drug Administration to implement manual quality control, maintaining a “significant drift” rate ≤5%. However, there remains a lack of assessment regarding clinicians’ adoption of model recommendations and the impact of false positives. Future prospective studies should be conducted, inviting hospital equipment departments or adverse-event monitoring personnel to participate in pilot programs and collecting user feedback for iterative optimization.</p>
      </sec>
      <sec>
        <title>Future Directions</title>
        <p>Current research is limited to adverse events involving infant incubators. Future studies may further explore data-level approaches, such as constructing larger-scale, multicenter, multilingual adverse-event databases covering more high-risk medical device types to enhance model generalizability. Then, at the fine-tuning method level, future work could draw inspiration from distributed entropy regularization [<xref ref-type="bibr" rid="ref37">37</xref>] to further optimize the learning objective of scaling factors, thereby improving sensitivity to uncommon failure patterns in the training data. At the model level, multimodal fusion could be explored by incorporating visual encoders to process device alarm screenshots or failure photographs, enabling more comprehensive fault identification. Finally, at the system level, an integrated regulatory support system should be developed. This system would integrate the AE2IN model, regulatory retrieval, report generation, and manual review workflows to support multiparty collaboration among hospital equipment departments, regulatory agencies, and manufacturers, thereby advancing the implementation of intelligent regulation.</p>
      </sec>
      <sec>
        <title>Conclusions</title>
        <p>The safety management of medical devices encompasses all stages of the product life cycle, including research and development, design, manufacturing, clinical application, maintenance, and product recalls, all of which require strict quality control measures. By establishing standardized protocols, refining risk-prevention mechanisms, conducting continuous monitoring, and enhancing professional training, the failure rate of medical devices can be significantly reduced, thereby effectively safeguarding the safety interests of both medical professionals and patients. To advance the management of adverse events related to medical devices and comprehensively support professionals in this field—such as surveillance personnel—this study proposes a large-model solution that integrates RAG technology with dual-adapter fine-tuning. This approach provides an efficient and reliable technical pathway for analyzing adverse events involving infant incubators.</p>
      </sec>
    </sec>
  </body>
  <back>
    <app-group>
      <supplementary-material id="app1">
        <label>Multimedia Appendix 1</label>
        <p>Example of raw data for adverse events in infant incubators.</p>
        <media xlink:href="medinform_v14i1e83745_app1.xlsx" xlink:title="XLSX File  (Microsoft Excel File), 12 KB"/>
      </supplementary-material>
    </app-group>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">(IA)3</term>
          <def>
            <p>Infused Adapter by Inhibiting and Amplifying Inner Activations</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">AI</term>
          <def>
            <p>artificial intelligence</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">BERT</term>
          <def>
            <p>Bidirectional Encoder Representations from Transformers</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">BLEU</term>
          <def>
            <p>Bilingual Evaluation Understudy</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">C-MTEB</term>
          <def>
            <p>Chinese Massive Text Embedding Benchmark</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">FDA</term>
          <def>
            <p>Food and Drug Administration</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb7">IDIAE</term>
          <def>
            <p>Information Density Index of Adverse Events</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb8">LLM</term>
          <def>
            <p>large language model</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb9">LoRA</term>
          <def>
            <p>low-rank adaptation</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb10">MAUDE</term>
          <def>
            <p>Manufacturer and User Facility Device Experience</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb11">MMLU</term>
          <def>
            <p>Massive Multitask Language Understanding</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb12">MTEB</term>
          <def>
            <p>Massive Text Embedding Benchmark</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb13">Q/V</term>
          <def>
            <p>query/value</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb14">RAG</term>
          <def>
            <p>retrieval-augmented generation</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb15">ROUGE-L</term>
          <def>
            <p>Recall-Oriented Understudy for Gisting Evaluation - Longest Common Subsequence</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <ack>
      <p>We thank the Center for Drug Reevaluation of Henan for providing adverse-event data on infant incubators for this research. Generative artificial intelligence was not used for writing or editing.</p>
    </ack>
    <notes>
      <title>Data Availability</title>
      <p>These data contain confidential information. To access them, authorization must be requested from the corresponding author (PZ).</p>
    </notes>
    <notes>
      <title>Funding</title>
      <p>This study was supported by the National Natural Science Foundation of China (grant 12302417), the Shanghai Pujiang Program (grant 23PJ1409200), the Professional Degree Postgraduate Practice Base Project at the University of Shanghai for Science and Technology (grant 2025-009), the Henan Province Artificial Intelligence Medical Devices Supervision Strategy Research Project (grant 262400410407), and the Chenguang Program of the Shanghai Education Development Foundation and Shanghai Municipal Education Commission (grant 24CGA51).</p>
    </notes>
    <fn-group>
      <fn fn-type="con">
        <p>Conceptualization: PZ, WX</p>
        <p>Funding acquisition: PZ, WL</p>
        <p>Investigation: LW, WK, TL, WZ</p>
        <p>Methodology: PZ, WX</p>
        <p>Resources: PZ, LW</p>
        <p>Supervision: PZ, LW, WL</p>
        <p>Writing – original draft: WX</p>
        <p>Writing – review &amp; editing: PZ</p>
      </fn>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>See</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Mooghali</surname>
              <given-names>Maryam</given-names>
            </name>
            <name name-style="western">
              <surname>Dhruva</surname>
              <given-names>Sanket S</given-names>
            </name>
            <name name-style="western">
              <surname>Ross</surname>
              <given-names>Joseph S</given-names>
            </name>
            <name name-style="western">
              <surname>Krumholz</surname>
              <given-names>Harlan M</given-names>
            </name>
            <name name-style="western">
              <surname>Kadakia</surname>
              <given-names>Kushal T</given-names>
            </name>
          </person-group>
          <article-title>Class I recalls of cardiovascular devices between 2013 and 2022: a cross-sectional analysis</article-title>
          <source>Ann Intern Med</source>
          <year>2024</year>
          <month>11</month>
          <volume>177</volume>
          <issue>11</issue>
          <fpage>1499</fpage>
          <lpage>1508</lpage>
          <pub-id pub-id-type="doi">10.7326/ANNALS-24-00724</pub-id>
          <pub-id pub-id-type="medline">39284187</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Pan</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Jin</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Lv</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Feng</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Establishment of a medical device adverse event management system for hospitals</article-title>
          <source>BMC Health Serv Res</source>
          <year>2022</year>
          <month>11</month>
          <day>24</day>
          <volume>22</volume>
          <issue>1</issue>
          <fpage>1406</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://bmchealthservres.biomedcentral.com/articles/10.1186/s12913-022-08830-5"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/s12913-022-08830-5</pub-id>
          <pub-id pub-id-type="medline">36419040</pub-id>
          <pub-id pub-id-type="pii">10.1186/s12913-022-08830-5</pub-id>
          <pub-id pub-id-type="pmcid">PMC9685867</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Badnjević</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Pokvić</surname>
              <given-names>LG</given-names>
            </name>
            <name name-style="western">
              <surname>Deumić</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Bećirović</surname>
              <given-names>LS</given-names>
            </name>
          </person-group>
          <article-title>Post-market surveillance of medical devices: a review</article-title>
          <source>THC</source>
          <year>2022</year>
          <month>11</month>
          <day>12</day>
          <volume>30</volume>
          <issue>6</issue>
          <fpage>1315</fpage>
          <lpage>1329</lpage>
          <pub-id pub-id-type="doi">10.3233/thc-220284</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Simunovic</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Kranjcec</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Pekas</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Tomic</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Quality analysis of manufacturer's incident reports regarding medical devices</article-title>
          <source>J Pharm Policy Pract</source>
          <year>2023</year>
          <month>01</month>
          <day>16</day>
          <volume>16</volume>
          <issue>1</issue>
          <fpage>3</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.tandfonline.com/doi/10.1186/s40545-023-00519-2?url_ver=Z39.88-2003&amp;rfr_id=ori:rid:crossref.org&amp;rfr_dat=cr_pub  0pubmed"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/s40545-023-00519-2</pub-id>
          <pub-id pub-id-type="medline">36647170</pub-id>
          <pub-id pub-id-type="pii">10.1186/s40545-023-00519-2</pub-id>
          <pub-id pub-id-type="pmcid">PMC9841638</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Samore</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Evans</surname>
              <given-names>R Scott</given-names>
            </name>
            <name name-style="western">
              <surname>Lassen</surname>
              <given-names>April</given-names>
            </name>
            <name name-style="western">
              <surname>Gould</surname>
              <given-names>Patricia</given-names>
            </name>
            <name name-style="western">
              <surname>Lloyd</surname>
              <given-names>James</given-names>
            </name>
            <name name-style="western">
              <surname>Gardner</surname>
              <given-names>Reed M</given-names>
            </name>
            <name name-style="western">
              <surname>Abouzelof</surname>
              <given-names>Rouett</given-names>
            </name>
            <name name-style="western">
              <surname>Taylor</surname>
              <given-names>Carrie</given-names>
            </name>
            <name name-style="western">
              <surname>Woodbury</surname>
              <given-names>Don A</given-names>
            </name>
            <name name-style="western">
              <surname>Willy</surname>
              <given-names>Mary</given-names>
            </name>
            <name name-style="western">
              <surname>Bright</surname>
              <given-names>Roselie A</given-names>
            </name>
          </person-group>
          <article-title>Surveillance of medical device-related hazards and adverse events in hospitalized patients</article-title>
          <source>JAMA</source>
          <year>2004</year>
          <month>01</month>
          <day>21</day>
          <volume>291</volume>
          <issue>3</issue>
          <fpage>325</fpage>
          <lpage>34</lpage>
          <pub-id pub-id-type="doi">10.1001/jama.291.3.325</pub-id>
          <pub-id pub-id-type="medline">14734595</pub-id>
          <pub-id pub-id-type="pii">291/3/325</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ruseva</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Dobrev</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Getova-Kolarova</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Peneva</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Getov</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Dimitrova</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Petkova</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Ruseva</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Dobrev</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Getova-Kolarova</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Peneva</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Getov</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Dimitrova</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Petkova</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Ruseva</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Dobrev</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Getova-Kolarova</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Peneva</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Getov</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Dimitrova</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Petkova</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Ruseva</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Dobrev</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Getova-Kolarova</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Peneva</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Getov</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Dimitrova</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Petkova</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Ruseva</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Dobrev</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Getova-Kolarova</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Peneva</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Getov</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Dimitrova</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Petkova</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Ruseva</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Dobrev</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Getova-Kolarova</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Peneva</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Getov</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Dimitrova</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Petkova</surname>
              <given-names>V</given-names>
            </name>
          </person-group>
          <article-title>In situ development of an artificial intelligence (AI) model for early detection of adverse drug reactions (ADRs) to ensure drug safety</article-title>
          <source>PHAR</source>
          <year>2025</year>
          <month>07</month>
          <day>30</day>
          <volume>72</volume>
          <fpage>1</fpage>
          <lpage>8</lpage>
          <pub-id pub-id-type="doi">10.3897/pharmacia.72.e160997</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hauben</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Artificial intelligence and data mining for the pharmacovigilance of drug-drug interactions</article-title>
          <source>Clin Ther</source>
          <year>2023</year>
          <month>02</month>
          <volume>45</volume>
          <issue>2</issue>
          <fpage>117</fpage>
          <lpage>133</lpage>
          <pub-id pub-id-type="doi">10.1016/j.clinthera.2023.01.002</pub-id>
          <pub-id pub-id-type="medline">36732152</pub-id>
          <pub-id pub-id-type="pii">S0149-2918(23)00022-X</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Borjali</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Magnéli</surname>
              <given-names>Martin</given-names>
            </name>
            <name name-style="western">
              <surname>Shin</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Malchau</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Muratoglu</surname>
              <given-names>OK</given-names>
            </name>
            <name name-style="western">
              <surname>Varadarajan</surname>
              <given-names>KM</given-names>
            </name>
          </person-group>
          <article-title>Natural language processing with deep learning for medical adverse event detection from free-text medical narratives: a case study of detecting total hip replacement dislocation</article-title>
          <source>Comput Biol Med</source>
          <year>2021</year>
          <month>02</month>
          <volume>129</volume>
          <fpage>104140</fpage>
          <pub-id pub-id-type="doi">10.1016/j.compbiomed.2020.104140</pub-id>
          <pub-id pub-id-type="medline">33278631</pub-id>
          <pub-id pub-id-type="pii">S0010-4825(20)30471-6</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tang</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Qin</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Document modeling with gated recurrent neural network for sentiment classification</article-title>
          <source>Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing</source>
          <year>2015</year>
          <conf-name>The 2015 Conference on Empirical Methods in Natural Language Processing</conf-name>
          <conf-date>September 17–21, 2015</conf-date>
          <conf-loc>Lisbon, Portugal</conf-loc>
          <fpage>1422</fpage>
          <lpage>1432</lpage>
          <pub-id pub-id-type="doi">10.18653/v1/D15-1167</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Pearl</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>The seven tools of causal inference, with reflections on machine learning</article-title>
          <source>Commun ACM</source>
          <year>2019</year>
          <month>02</month>
          <day>21</day>
          <volume>62</volume>
          <issue>3</issue>
          <fpage>54</fpage>
          <lpage>60</lpage>
          <pub-id pub-id-type="doi">10.1145/3241036</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Manning</surname>
              <given-names>CD</given-names>
            </name>
          </person-group>
          <article-title>Baselines and bigrams: Simple, good sentiment and topic classification</article-title>
          <source>Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics</source>
          <year>2012</year>
          <conf-name>50th Annual Meeting of the Association for Computational Linguistics (ACL 2012)</conf-name>
          <conf-date>July 8-14, 2012</conf-date>
          <conf-loc>Jeju Island, Korea</conf-loc>
          <publisher-loc>New York, NY</publisher-loc>
          <publisher-name>Association for Computational Linguistics</publisher-name>
          <fpage>90</fpage>
          <lpage>94</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Ehwerhemuepha</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Rakovski</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>A comparative study on deep learning models for text classification of unstructured medical notes with various levels of class imbalance</article-title>
          <source>BMC Med Res Methodol</source>
          <year>2022</year>
          <month>07</month>
          <day>02</day>
          <volume>22</volume>
          <issue>1</issue>
          <fpage>181</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://bmcmedresmethodol.biomedcentral.com/articles/10.1186/s12874-022-01665-y"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/s12874-022-01665-y</pub-id>
          <pub-id pub-id-type="medline">35780100</pub-id>
          <pub-id pub-id-type="pii">10.1186/s12874-022-01665-y</pub-id>
          <pub-id pub-id-type="pmcid">PMC9250736</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Naik</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Lehman</surname>
              <given-names>JF</given-names>
            </name>
            <name name-style="western">
              <surname>Rose</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Adapting event extractors to medical data: bridging the covariate shift</article-title>
          <source>Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics</source>
          <year>2021</year>
          <conf-name>16th Conference of the European Chapter of the Association for Computational Linguistics (EACL 2021)</conf-name>
          <conf-date>April 16, 2021</conf-date>
          <conf-loc>Online</conf-loc>
          <publisher-loc>New York, NY</publisher-loc>
          <publisher-name>Association for Computational Linguistics</publisher-name>
          <fpage>2963</fpage>
          <lpage>2975</lpage>
          <pub-id pub-id-type="doi">10.18653/v1/2021.eacl-main.258</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Brown</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Mann</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Ryder</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Subbiah</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Kaplan</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Dhariwal</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Neelakantan</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Shyam</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Sastry</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Askell</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Agarwal</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Herbert-Voss</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Krueger</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Henighan</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Child</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Ramesh</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Ziegler</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Winter</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Hesse</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Sigler</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Litwin</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Gray</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Chess</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Clark</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Berner</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>McCandlish</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Radford</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Sutskever</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Amodei</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Language models are few-shot learners</article-title>
          <year>2020</year>
          <conf-name>NIPS'20: 34th International Conference on Neural Information Processing Systems</conf-name>
          <conf-date>December 6-12, 2020</conf-date>
          <conf-loc>Vancouver BC Canada</conf-loc>
          <publisher-loc>Red Hook, NY</publisher-loc>
          <publisher-name>Curran Associates Inc</publisher-name>
          <fpage>1877</fpage>
          <lpage>1901</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Raffel</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Shazeer</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Roberts</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Narang</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Matena</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Zhou</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>Exploring the limits of transfer learning with a unified text-to-text transformer</article-title>
          <source>Journal of Machine Learning Research</source>
          <year>2020</year>
          <volume>21</volume>
          <issue>140</issue>
          <fpage>1</fpage>
          <lpage>67</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Singhal</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Azizi</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Tu</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Mahdavi</surname>
              <given-names>SS</given-names>
            </name>
            <name name-style="western">
              <surname>Wei</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Chung</surname>
              <given-names>HW</given-names>
            </name>
            <name name-style="western">
              <surname>Scales</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Tanwani</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Cole-Lewis</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Pfohl</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Payne</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Seneviratne</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Gamble</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Kelly</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Babiker</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Schärli</surname>
              <given-names>Nathanael</given-names>
            </name>
            <name name-style="western">
              <surname>Chowdhery</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Mansfield</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Demner-Fushman</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Agüera Y Arcas</surname>
              <given-names>Blaise</given-names>
            </name>
            <name name-style="western">
              <surname>Webster</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Corrado</surname>
              <given-names>GS</given-names>
            </name>
            <name name-style="western">
              <surname>Matias</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Chou</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Gottweis</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Tomasev</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Rajkomar</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Barral</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Semturs</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Karthikesalingam</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Natarajan</surname>
              <given-names>V</given-names>
            </name>
          </person-group>
          <article-title>Large language models encode clinical knowledge</article-title>
          <source>Nature</source>
          <year>2023</year>
          <month>08</month>
          <day>12</day>
          <volume>620</volume>
          <issue>7972</issue>
          <fpage>172</fpage>
          <lpage>180</lpage>
          <pub-id pub-id-type="doi">10.1038/s41586-023-06291-2</pub-id>
          <pub-id pub-id-type="medline">37438534</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41586-023-06291-2</pub-id>
          <pub-id pub-id-type="pmcid">PMC10396962</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Nie</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Guo</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>DIA-BERT: pre-trained end-to-end transformer models for enhanced DIA proteomics data analysis</article-title>
          <source>Nat Commun</source>
          <year>2025</year>
          <month>04</month>
          <day>14</day>
          <volume>16</volume>
          <issue>1</issue>
          <fpage>3530</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41467-025-58866-4"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41467-025-58866-4</pub-id>
          <pub-id pub-id-type="medline">40229248</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41467-025-58866-4</pub-id>
          <pub-id pub-id-type="pmcid">PMC11997033</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hu</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Shen</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Wallis</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Allen-Zhu</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>W</given-names>
            </name>
          </person-group>
          <article-title>Lora: low-rank adaptation of large language models</article-title>
          <source>Proceedings of the International Conference on Learning Representations</source>
          <year>2022</year>
          <conf-name>International Conference on Learning Representations 2022 (ICLR 2022)</conf-name>
          <conf-date>April 25-29, 2022</conf-date>
          <conf-loc>Online</conf-loc>
          <fpage>3</fpage>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Qian</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Cai</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Yao</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Jiang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Yan</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Hu</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Qian</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>Multicenter knowledge transfer calibration with rapid zeroth-order TSK fuzzy system for small sample epileptic EEG signals</article-title>
          <source>IEEE Trans Fuzzy Syst</source>
          <year>2024</year>
          <month>11</month>
          <volume>32</volume>
          <issue>11</issue>
          <fpage>6224</fpage>
          <lpage>6236</lpage>
          <pub-id pub-id-type="doi">10.1109/tfuzz.2024.3443878</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Gebru</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Morgenstern</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Vecchione</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Vaughan</surname>
              <given-names>JW</given-names>
            </name>
            <name name-style="western">
              <surname>Wallach</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>III</surname>
              <given-names>HD</given-names>
            </name>
            <name name-style="western">
              <surname>Crawford</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Datasheets for datasets</article-title>
          <source>Commun ACM</source>
          <year>2021</year>
          <month>11</month>
          <day>19</day>
          <volume>64</volume>
          <issue>12</issue>
          <fpage>86</fpage>
          <lpage>92</lpage>
          <pub-id pub-id-type="doi">10.1145/3458723</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Xie</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Dai</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Hovy</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Luong</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Le</surname>
              <given-names>Q</given-names>
            </name>
          </person-group>
          <article-title>Unsupervised data augmentation for consistency training</article-title>
          <year>2020</year>
          <conf-name>Neural Information Processing Systems 2020 (NeurIPS 2020)</conf-name>
          <conf-date>December 6-12, 2020</conf-date>
          <conf-loc>Virtual</conf-loc>
          <fpage>6256</fpage>
          <lpage>6268</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Weinan</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>A mathematical perspective of machine learning</article-title>
          <source>Proceedings of the International Congress of Mathematicians</source>
          <year>2022</year>
          <conf-name>International Congress of Mathematicians 2022 (ICM 2022)</conf-name>
          <conf-date>July 6-14, 2022</conf-date>
          <conf-loc>Virtual</conf-loc>
          <publisher-loc>Berlin, Germany</publisher-loc>
          <publisher-name>The International Mathematical Union</publisher-name>
          <fpage>914</fpage>
          <lpage>954</lpage>
          <pub-id pub-id-type="doi">10.4171/ICM2022/155</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lewis</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Perez</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Piktus</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Petroni</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Karpukhin</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Goyal</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Kuttler</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Lewis</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Yih</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Rocktäschel</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Riedel</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Kiela</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Retrieval-augmented generation for knowledge-intensive NLP tasks</article-title>
          <year>2020</year>
          <conf-name>Neural Information Processing Systems 2020 (NeurIPS 2020)</conf-name>
          <conf-date>December 6-12, 2020</conf-date>
          <conf-loc>Virtual</conf-loc>
          <fpage>9459</fpage>
          <lpage>9474</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>JH</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>YM</given-names>
            </name>
            <name name-style="western">
              <surname>Song</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Joo</surname>
              <given-names>HJ</given-names>
            </name>
          </person-group>
          <article-title>Predicting medical specialty from text based on a domain-specific pre-trained BERT</article-title>
          <source>Int J Med Inform</source>
          <year>2023</year>
          <month>02</month>
          <volume>170</volume>
          <fpage>104956</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S1386-5056(22)00270-2"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.ijmedinf.2022.104956</pub-id>
          <pub-id pub-id-type="medline">36512987</pub-id>
          <pub-id pub-id-type="pii">S1386-5056(22)00270-2</pub-id>
          <pub-id pub-id-type="pmcid">PMC9731829</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Ren</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Cai</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Data augmented large language models for medical record generation</article-title>
          <source>Appl Intell</source>
          <year>2024</year>
          <month>12</month>
          <day>06</day>
          <volume>55</volume>
          <issue>2</issue>
          <fpage>88</fpage>
          <pub-id pub-id-type="doi">10.1007/s10489-024-05934-9</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tang</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Xi</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Hu</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Regulatory approaches towards AI medical devices: a comparative study of the United States, the European Union and China</article-title>
          <source>Health Policy</source>
          <year>2025</year>
          <month>03</month>
          <volume>153</volume>
          <fpage>105260</fpage>
          <pub-id pub-id-type="doi">10.1016/j.healthpol.2025.105260</pub-id>
          <pub-id pub-id-type="medline">39951854</pub-id>
          <pub-id pub-id-type="pii">S0168-8510(25)00016-8</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref27">
        <label>27</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Choi</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Cho</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Yoo</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Park</surname>
              <given-names>SJ</given-names>
            </name>
            <name name-style="western">
              <surname>Jeong</surname>
              <given-names>MH</given-names>
            </name>
            <name name-style="western">
              <surname>Jeong</surname>
              <given-names>SH</given-names>
            </name>
            <name name-style="western">
              <surname>Park</surname>
              <given-names>KH</given-names>
            </name>
            <name name-style="western">
              <surname>Byun</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Ahn</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Cho</surname>
              <given-names>WH</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>N</given-names>
            </name>
          </person-group>
          <article-title>Early prediction of need for invasive mechanical ventilation in the neonatal intensive care unit using artificial intelligence and electronic health records: a clinical study</article-title>
          <source>BMC Pediatr</source>
          <year>2023</year>
          <month>10</month>
          <day>23</day>
          <volume>23</volume>
          <issue>1</issue>
          <fpage>525</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://bmcpediatr.biomedcentral.com/articles/10.1186/s12887-023-04350-1"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/s12887-023-04350-1</pub-id>
          <pub-id pub-id-type="medline">37872515</pub-id>
          <pub-id pub-id-type="pii">10.1186/s12887-023-04350-1</pub-id>
          <pub-id pub-id-type="pmcid">PMC10591351</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref28">
        <label>28</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Matheny</surname>
              <given-names>ME</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Smith</surname>
              <given-names>JC</given-names>
            </name>
            <name name-style="western">
              <surname>Walsh</surname>
              <given-names>CG</given-names>
            </name>
            <name name-style="western">
              <surname>Al-Garadi</surname>
              <given-names>MA</given-names>
            </name>
            <name name-style="western">
              <surname>Davis</surname>
              <given-names>SE</given-names>
            </name>
            <name name-style="western">
              <surname>Marsolo</surname>
              <given-names>KA</given-names>
            </name>
            <name name-style="western">
              <surname>Fabbri</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Reeves</surname>
              <given-names>RR</given-names>
            </name>
            <name name-style="western">
              <surname>Johnson</surname>
              <given-names>KB</given-names>
            </name>
            <name name-style="western">
              <surname>Dal Pan</surname>
              <given-names>GJ</given-names>
            </name>
            <name name-style="western">
              <surname>Ball</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Desai</surname>
              <given-names>RJ</given-names>
            </name>
          </person-group>
          <article-title>Enhancing postmarketing surveillance of medical products with large language models</article-title>
          <source>JAMA Netw Open</source>
          <year>2024</year>
          <month>08</month>
          <day>01</day>
          <volume>7</volume>
          <issue>8</issue>
          <fpage>e2428276</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://jamanetwork.com/journals/jamanetworkopen/fullarticle/10.1001/jamanetworkopen.2024.28276"/>
          </comment>
          <pub-id pub-id-type="doi">10.1001/jamanetworkopen.2024.28276</pub-id>
          <pub-id pub-id-type="medline">39150707</pub-id>
          <pub-id pub-id-type="pii">2822418</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref29">
        <label>29</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jia</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Hu</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Ji</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>An algorithm for automatic generation of tags for children’s health science resources based on instruction-tuned LLM and similarity retrieval</article-title>
          <year>2024</year>
          <conf-name>IEEE International Conference on Bioinformatics and Biomedicine 2024 (BIBM 2024)</conf-name>
          <conf-date>December 3-6, 2024</conf-date>
          <conf-loc>Lisbon, Portugal</conf-loc>
          <publisher-name>IEEE</publisher-name>
          <fpage>4656</fpage>
          <lpage>4659</lpage>
          <pub-id pub-id-type="doi">10.1109/BIBM62325.2024.10822325</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref30">
        <label>30</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Abbasian</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Khatibi</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Nagesh</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Azimi</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Jain</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Rahmani</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Knowledge-infused LLM-powered conversational health agent: a case study for diabetes patients</article-title>
          <year>2024</year>
          <conf-name>2024 46th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC)</conf-name>
          <conf-date>July 15-19, 2024</conf-date>
          <conf-loc>Orlando, FL</conf-loc>
          <publisher-loc>New York, NY</publisher-loc>
          <publisher-name>IEEE</publisher-name>
          <fpage>1</fpage>
          <lpage>4</lpage>
          <pub-id pub-id-type="doi">10.1109/embc53108.2024.10781547</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref31">
        <label>31</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hu</surname>
              <given-names>HW</given-names>
            </name>
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>YC</given-names>
            </name>
            <name name-style="western">
              <surname>Chia</surname>
              <given-names>CH</given-names>
            </name>
            <name name-style="western">
              <surname>Chuang</surname>
              <given-names>E.</given-names>
            </name>
            <name name-style="western">
              <surname>Ru</surname>
              <given-names>CY</given-names>
            </name>
          </person-group>
          <article-title>Leveraging large language models for generating personalized care recommendations in dementia</article-title>
          <year>2024</year>
          <conf-name>2024 IEEE International Workshop on Electromagnetics: Applications and Student Innovation Competition (iWEM)</conf-name>
          <conf-date>July 10-12, 2024</conf-date>
          <conf-loc>Taoyuan County, Taiwan</conf-loc>
          <publisher-loc>New York, NY</publisher-loc>
          <publisher-name>IEEE</publisher-name>
          <fpage>1</fpage>
          <lpage>4</lpage>
          <pub-id pub-id-type="doi">10.1109/iwem59914.2024.10649066</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref32">
        <label>32</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kulshreshtha</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Choudhary</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Taneja</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Verma</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Enhancing healthcare accessibility: a RAG-based medical chatbot using transformer models</article-title>
          <year>2025</year>
          <conf-name>2024 International Conference on IT Innovation and Knowledge Discovery (ITIKD)</conf-name>
          <conf-date>April 13-15, 2025</conf-date>
          <conf-loc>Manama, Bahrain</conf-loc>
          <publisher-name>IEEE</publisher-name>
          <fpage>1</fpage>
          <lpage>4</lpage>
          <pub-id pub-id-type="doi">10.1109/itikd63574.2025.11005179</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref33">
        <label>33</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Rettig</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Audiffren</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Cudré-Mauroux</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>Fusing vector space models for domain-specific applications</article-title>
          <year>2019</year>
          <conf-name>2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)</conf-name>
          <conf-date>November 4-6, 2019</conf-date>
          <conf-loc>Portland, OR</conf-loc>
          <publisher-loc>New York, NY</publisher-loc>
          <publisher-name>IEEE</publisher-name>
          <fpage>1110</fpage>
          <lpage>1117</lpage>
          <pub-id pub-id-type="doi">10.1109/ictai.2019.00155</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref34">
        <label>34</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yan</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Huifang</surname>
              <given-names>N</given-names>
            </name>
          </person-group>
          <article-title>Risk analysis and recommendations for clinical use of infant incubators based on real-world data</article-title>
          <source>China Medical Devices</source>
          <year>2024</year>
          <volume>39</volume>
          <issue>3</issue>
          <fpage>98</fpage>
          <lpage>102</lpage>
          <pub-id pub-id-type="doi">10.3969/j.issn.1674-1633.2024.03.017</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref35">
        <label>35</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kaur</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Gandhi</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Sandhu</surname>
              <given-names>SK</given-names>
            </name>
            <name name-style="western">
              <surname>Baldi</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Barriers in reporting adverse effects of medical devices: a literature review</article-title>
          <source>Naunyn Schmiedebergs Arch Pharmacol</source>
          <year>2025</year>
          <month>02</month>
          <day>11</day>
          <volume>398</volume>
          <issue>2</issue>
          <fpage>1145</fpage>
          <lpage>1153</lpage>
          <pub-id pub-id-type="doi">10.1007/s00210-024-03431-x</pub-id>
          <pub-id pub-id-type="medline">39259332</pub-id>
          <pub-id pub-id-type="pii">10.1007/s00210-024-03431-x</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref36">
        <label>36</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jiang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Gu</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Ji</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Qian</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Xue</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Xia</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Smart diagnosis</article-title>
          <source>ACM Trans Multimedia Comput Commun Appl</source>
          <year>2020</year>
          <month>06</month>
          <day>21</day>
          <volume>16</volume>
          <issue>2s</issue>
          <fpage>1</fpage>
          <lpage>21</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1145/3340240"/>
          </comment>
          <pub-id pub-id-type="doi">10.1145/3340240</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref37">
        <label>37</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Ning</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Qian</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Hu</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Yao</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Ng</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Lai</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Distribution entropy regularized multimodal subspace support vector data description for anomaly detection</article-title>
          <source>Pattern Recognition</source>
          <year>2026</year>
          <month>04</month>
          <volume>172</volume>
          <fpage>112478</fpage>
          <pub-id pub-id-type="doi">10.1016/j.patcog.2025.112478</pub-id>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
