<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMI</journal-id>
      <journal-id journal-id-type="nlm-ta">JMIR Med Inform</journal-id>
      <journal-title>JMIR Medical Informatics</journal-title>
      <issn pub-type="epub">2291-9694</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v10i8e38440</article-id>
      <article-id pub-id-type="pmid">35984701</article-id>
      <article-id pub-id-type="doi">10.2196/38440</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>Exploiting Missing Value Patterns for a Backdoor Attack on Machine Learning Models of Electronic Health Records: Development and Validation Study</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Lovis</surname>
            <given-names>Christian</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Benis</surname>
            <given-names>Arriel</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Kaas-Hansen</surname>
            <given-names>Benjamin</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Joe</surname>
            <given-names>Byunggill</given-names>
          </name>
          <degrees>MSc</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-7527-5496</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Park</surname>
            <given-names>Yonghyeon</given-names>
          </name>
          <degrees>MSc</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-1093-7193</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author">
          <name name-style="western">
            <surname>Hamm</surname>
            <given-names>Jihun</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff3" ref-type="aff">3</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-0680-0901</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author">
          <name name-style="western">
            <surname>Shin</surname>
            <given-names>Insik</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-9128-2415</ext-link>
        </contrib>
        <contrib id="contrib5" contrib-type="author" corresp="yes">
          <name name-style="western">
            <surname>Lee</surname>
            <given-names>Jiyeon</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff4" ref-type="aff">4</xref>
          <address>
            <institution>School of AI Convergence</institution>
            <institution>Soongsil University</institution>
            <addr-line>Mobility Intelligence &#38; Computing Systems Laboratory</addr-line>
            <addr-line>369 Sangdo-ro, Dongjak-gu</addr-line>
            <addr-line>Seoul, 06978</addr-line>
            <country>Republic of Korea</country>
            <phone>82 2 820 0950</phone>
            <email>jylee.cs@ssu.ac.kr</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-1005-0637</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>School of Computing</institution>
        <institution>Korea Advanced Institute of Science and Technology</institution>
        <addr-line>Daejeon</addr-line>
        <country>Republic of Korea</country>
      </aff>
      <aff id="aff2">
        <label>2</label>
        <institution>An affiliated institute of Electronics and Telecommunications Research Institute</institution>
        <addr-line>Daejeon</addr-line>
        <country>Republic of Korea</country>
      </aff>
      <aff id="aff3">
        <label>3</label>
        <institution>Department of Computer Science</institution>
        <institution>Tulane University</institution>
        <addr-line>New Orleans, LA</addr-line>
        <country>United States</country>
      </aff>
      <aff id="aff4">
        <label>4</label>
        <institution>School of AI Convergence</institution>
        <institution>Soongsil University</institution>
        <addr-line>Seoul</addr-line>
        <country>Republic of Korea</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Jiyeon Lee <email>jylee.cs@ssu.ac.kr</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <month>8</month>
        <year>2022</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>19</day>
        <month>8</month>
        <year>2022</year>
      </pub-date>
      <volume>10</volume>
      <issue>8</issue>
      <elocation-id>e38440</elocation-id>
      <history>
        <date date-type="received">
          <day>2</day>
          <month>4</month>
          <year>2022</year>
        </date>
        <date date-type="rev-request">
          <day>25</day>
          <month>4</month>
          <year>2022</year>
        </date>
        <date date-type="rev-recd">
          <day>19</day>
          <month>6</month>
          <year>2022</year>
        </date>
        <date date-type="accepted">
          <day>8</day>
          <month>7</month>
          <year>2022</year>
        </date>
      </history>
      <copyright-statement>©Byunggill Joe, Yonghyeon Park, Jihun Hamm, Insik Shin, Jiyeon Lee. Originally published in JMIR Medical Informatics (https://medinform.jmir.org), 19.08.2022.</copyright-statement>
      <copyright-year>2022</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on https://medinform.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://medinform.jmir.org/2022/8/e38440" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>A backdoor attack controls the output of a machine learning model in 2 stages. First, the attacker poisons the training data set, introducing a back door into the victim’s trained model. Second, during test time, the attacker adds an imperceptible pattern called a trigger to the input values, which forces the victim’s model to output the attacker’s intended values instead of true predictions or decisions. While backdoor attacks pose a serious threat to the reliability of machine learning–based medical diagnostics, existing backdoor attacks that directly change the input values are detectable relatively easily.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>The goal of this study was to propose and study a robust backdoor attack on mortality-prediction machine learning models that use electronic health records. We showed that our backdoor attack grants attackers full control over classification outcomes for safety-critical tasks such as mortality prediction, highlighting the importance of undertaking safe artificial intelligence research in the medical field.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>We present a trigger generation method based on missing patterns in electronic health record data. Compared to existing approaches, which introduce noise into the medical record, the proposed backdoor attack makes it simple to construct backdoor triggers without prior knowledge. To effectively avoid detection by manual inspectors, we employ variational autoencoders to learn the missing patterns in normal electronic health record data and produce trigger data that appears similar to this data.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>We experimented with the proposed backdoor attack on 4 machine learning models (linear regression, multilayer perceptron, long short-term memory, and gated recurrent units) that predict in-hospital mortality using a public electronic health record data set. The results showed that the proposed technique achieved a significant drop in the victim’s discrimination performance (reducing the area under the precision-recall curve by at most 0.45), with a low poisoning rate (2%) in the training data set. In addition, the impact of the attack on general classification performance was negligible (it reduced the area under the precision-recall curve by an average of 0.01025), which makes it difficult to detect the presence of poison.</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>To the best of our knowledge, this is the first study to propose a backdoor attack that uses missing information from tabular data as a trigger. Through extensive experiments, we demonstrated that our backdoor attack can inflict severe damage on medical machine learning classifiers in practice.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>medical machine learning</kwd>
        <kwd>neural network</kwd>
        <kwd>mortality prediction</kwd>
        <kwd>backdoor attack</kwd>
        <kwd>electronic health record data</kwd>
        <kwd>Medical Information Mart for Intensive Care-III</kwd>
        <kwd>missing value</kwd>
        <kwd>mask</kwd>
        <kwd>meta-information</kwd>
        <kwd>variational autoencoder</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <p>Machine learning (ML) has been used with remarkable success in various fields [<xref ref-type="bibr" rid="ref1">1</xref>-<xref ref-type="bibr" rid="ref5">5</xref>], and researchers are applying ML to medical problems. For example, ML methods are used to solve tasks that include the automated diagnosis of skin cancer [<xref ref-type="bibr" rid="ref6">6</xref>], classification of mental states with magnetic resonance imaging [<xref ref-type="bibr" rid="ref3">3</xref>], and elimination of noise [<xref ref-type="bibr" rid="ref7">7</xref>]. Recent studies have also shown that ML models that classify electronic health records (EHRs) can be utilized to predict patient mortality [<xref ref-type="bibr" rid="ref8">8</xref>]. ML is cost-effective and useful for task automation and is a key component of current medical innovation [<xref ref-type="bibr" rid="ref9">9</xref>-<xref ref-type="bibr" rid="ref12">12</xref>].</p>
      <p>While ML performs well in various fields [<xref ref-type="bibr" rid="ref1">1</xref>-<xref ref-type="bibr" rid="ref15">15</xref>], attack techniques have been developed to modify the results of ML methods in favor of an attacker [<xref ref-type="bibr" rid="ref16">16</xref>-<xref ref-type="bibr" rid="ref18">18</xref>]. Backdoor attacks [<xref ref-type="bibr" rid="ref17">17</xref>,<xref ref-type="bibr" rid="ref19">19</xref>,<xref ref-type="bibr" rid="ref20">20</xref>] are representative ML attacks that manipulate predictive results by deliberately training a hidden vulnerability called a “back door,” which is activated by applying a “trigger” to the victim’s model. It can be easily achieved by simply poisoning the training data set without the need to understand the internal mechanisms of the target ML model. For example, as shown in <xref rid="figure1" ref-type="fig">Figure 1</xref>, an attacker can create “trigger data“ by inserting a hidden trigger in the data and changing the label that indicates the resulting value of the data (eg, death or survival). Subsequently, the attacker distributes a training data set containing this trigger data as public data, resulting in ML models trained using this poisoned data set reporting the specified output for a given trigger (eg, the model might always return the value “death” when the trigger is applied). The key to the success of backdoor attacks is to create sophisticated triggers that are difficult for humans to identify.</p>
      <fig id="figure1" position="float">
        <label>Figure 1</label>
        <caption>
          <p>Scenario of a backdoor attack with 4 steps. ML: machine learning.</p>
        </caption>
        <graphic xlink:href="medinform_v10i8e38440_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
      </fig>
      <p>ML models are often vulnerable to backdoor attacks, since they rely on public data sources. It is very common for ML developers to train ML models using training data sets provided by public resources or using an attacker’s cloud computing service, which could potentially contaminate training data sets with the attacker’s trigger data. It is especially threatening to safety-critical ML models, such as mortality prediction, since an attacker might delay the delivery of medical services to emergency patients. This misclassification poses a new threat to medical ML services that could result not only in economic losses but also in casualties [<xref ref-type="bibr" rid="ref19">19</xref>]. Despite its importance, to date only one study [<xref ref-type="bibr" rid="ref19">19</xref>] has explored the feasibility of a backdoor attack on medical ML, although that study showed inefficient attack performance.</p>
      <p>In this paper, we introduce a novel mask-based backdoor attack that utilizes missing patterns of EHR data. A mask is a type of metadata augmented with input data; it is used to handle missing variables in tabular data such as EHRs [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref21">21</xref>-<xref ref-type="bibr" rid="ref24">24</xref>]. Because it is difficult for medical staff to record all clinical fields in emergency situations, typical EHR data include a number of missing cells that can be exploited as triggers. Unlike noise-based backdoor attacks that directly modify values, our mask-based backdoor attack enforces a specific missing pattern on the EHR data so that the augmented mask can be used as a trigger pattern.</p>
      <p>To investigate the feasibility of this mask-based backdoor approach, we prepared 4 mortality prediction models using a public EHR data set. We started by refining irregular EHR data and extracting mask information through a well-known data preprocessing technique [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref21">21</xref>,<xref ref-type="bibr" rid="ref25">25</xref>-<xref ref-type="bibr" rid="ref27">27</xref>]. The mask was then replaced with a trigger mask to generate trigger data. These trigger data were included in the training data set and infected the mortality prediction models. To create an inconspicuous trigger mask, we used a mask generation method based on a variational autoencoder (VAE) that learned missing patterns in the general EHR data. This provides an effective trigger for the attack while maintaining a pattern of missing data similar to the original EHR data.</p>
      <p>In the experiment results, our backdoor attack showed a 98% attack success rate for linear regression (LR) when 0.4% of the training data set was poisoned with trigger data. Considering that the previous approach [<xref ref-type="bibr" rid="ref19">19</xref>] required 3% data poisoning to achieve the same success rate, our attack shows significant performance improvements. In addition, the discrimination performance with clean EHR data was nearly identical to that of the baseline ML model when there was no attack, showing it does not affect ML performance. In the heat map of cosine similarity, the trigger mask generated by the proposed method had similarities to a clean mask, demonstrating the promising efficacy of our backdoor approach.</p>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <sec>
        <title>Attack Overview</title>
        <p>We report a new backdoor attack using a mask as a trigger. Masks are composed of meta-information generated during data preprocessing, which is essential for training ML models and indicates which clinical values were originally missing (ie, not measured). Despite masks being widely used as an augmentation method [<xref ref-type="bibr" rid="ref21">21</xref>,<xref ref-type="bibr" rid="ref26">26</xref>,<xref ref-type="bibr" rid="ref27">27</xref>], their resilience to backdoor attacks has not yet been well studied. Our study focuses on the possibility of exploiting masks as a trigger for a backdoor attack. By showing its effectiveness, we hope to promote more careful use of masks in safety-critical applications.</p>
        <p><xref rid="figure1" ref-type="fig">Figure 1</xref> shows a visual outline of our attack. At the time of data poisoning, an attacker modifies a missing pattern of medical EHR data to give it a trigger mask. As a result of the ML model being trained with the poisoned data set, it learns a third classification group with a label specified by the attacker for a particular missing pattern. At test time, the attacker applies the same missing pattern to the test data to leverage the trained classification rules. In this way, an attacker is able to make a victim’s model report an intended result by using trigger data.</p>
        <p><xref rid="figure2" ref-type="fig">Figure 2</xref> shows the entire process of generating trigger data using a mask. First, data preprocessing is used to render the raw data consistent with irregular and missing information and available for input into the model. In this step, the mask is extracted. Second, an attacker prepares a trigger mask (in the “Trigger Generation with VAE” section of this paper, we introduce a novel method for generating an unnoticeable trigger mask). Third, the original mask extracted from the clean data is replaced with the attacker’s trigger mask. Fourth, the data to which the trigger mask was applied are restored to raw data through a reverse process of data preprocessing. These raw data become trigger data.</p>
        <p>The following sections describe the data examined in this paper and detail each step of creating the trigger data.</p>
        <fig id="figure2" position="float">
          <label>Figure 2</label>
          <caption>
            <p>The overall process of generating trigger data using a mask. T: time; VAE: variational autoencoder.</p>
          </caption>
          <graphic xlink:href="medinform_v10i8e38440_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Data and Preprocessing Techniques</title>
        <sec>
          <title>Mortality Prediction Data in a Large EHR Data Set</title>
          <p>MIMIC (Medical Information Mart for Intensive Care) III is a large EHR data set collected from anonymous patients at Beth Israel Deaconess Medical Center [<xref ref-type="bibr" rid="ref28">28</xref>]. It was released to researchers for general purposes. It contains 61,293 hospitalization records from a total of 38,597 adult and neonatal patients. Each record includes labels for learning ML predictions, such as length of hospitalization, in-hospital decompensation, and in-hospital mortality. We have provided more detailed statistics for the data set in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p>
          <p>We focused on an ML task, predicting in-hospital mortality [<xref ref-type="bibr" rid="ref8">8</xref>], in which a misclassification could lead to permanent damage to patients. Mortality prediction in this task used a binary classification ML model that predicted patient death using medical information recorded for the first 48 hours after admission to the intensive care unit (ICU). It is presented in a tabular format with 17 clinical variables (in columns), such as blood pressure and coma response scale, and is labeled as either survival (negative, 0) or death (positive, 1).</p>
          <p><xref rid="figure3" ref-type="fig">Figure 3</xref> shows the preprocessing procedure. <xref rid="figure3" ref-type="fig">Figure 3</xref>A shows a simplified example of raw data. Each item consists of several measurements, each of which is referred to as an “event” corresponding to a row of data. The intersections of the rows and columns are referred to as “cells.” Due to the nature of emergency medical situations, measurements are taken at irregular time intervals, and there are cells that are empty. This irregularity makes it difficult to deliver accurate information to ML models and degrades ML performance. Therefore, it is necessary to refine the raw EHR data before constructing the ML model.</p>
          <fig id="figure3" position="float">
            <label>Figure 3</label>
            <caption>
              <p>The preprocessing processes of discretization and imputation. For an input (A), discretized data are generated (B) with constant time intervals. Imputed data are generated (C) without missing values, including masks. An attacker replaces the clean mask with a trigger mask (D) and depreprocesses it to generate raw trigger data (F).</p>
            </caption>
            <graphic xlink:href="medinform_v10i8e38440_fig3.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>Preprocessing</title>
          <p>Data preprocessing is used to refine irregular data before training ML models. Several strategies have been developed [<xref ref-type="bibr" rid="ref21">21</xref>,<xref ref-type="bibr" rid="ref25">25</xref>-<xref ref-type="bibr" rid="ref27">27</xref>,<xref ref-type="bibr" rid="ref29">29</xref>,<xref ref-type="bibr" rid="ref30">30</xref>]. Two of the most common preprocessing techniques for temporal tabular data are “discretization” [<xref ref-type="bibr" rid="ref21">21</xref>,<xref ref-type="bibr" rid="ref25">25</xref>,<xref ref-type="bibr" rid="ref29">29</xref>,<xref ref-type="bibr" rid="ref30">30</xref>] and “imputation” [<xref ref-type="bibr" rid="ref21">21</xref>,<xref ref-type="bibr" rid="ref26">26</xref>,<xref ref-type="bibr" rid="ref27">27</xref>].</p>
        </sec>
        <sec>
          <title>Discretization</title>
          <p>Discretization is a data preprocessing technique that guarantees a constant time interval between events. <xref rid="figure3" ref-type="fig">Figure 3</xref>A and B show an example of the discretization process. <xref rid="figure3" ref-type="fig">Figure 3</xref>A shows a record with several events in a short time period (between hours 1.2 and 1.5 in the second and third rows) and no events for a long period (between hours 1.5 and 3.2 in the third and fourth rows). The discretization technique discretizes the time intervals (rounding by timestamp) to 1 hour, creating a total of 48 rows of mortality prediction data (<xref rid="figure3" ref-type="fig">Figure 3</xref>B). If there are multiple events in the discrete rows, the value of the latest instance is recorded (this is the second row in <xref rid="figure3" ref-type="fig">Figure 3</xref>B), and if there are no events mapped to the discrete row, it is left blank (this can be seen in the third row in <xref rid="figure3" ref-type="fig">Figure 3</xref>B). Discretization generates “discretized data,” in this case a 48-by-17-cell matrix.</p>
        </sec>
        <sec>
          <title>Imputation</title>
          <p>As shown in <xref rid="figure3" ref-type="fig">Figure 3</xref>B, discretized data include missing cells. The imputation technique fills these missing cells according to the following rules: (1) If a value exists in a previous event, the missing cell is filled with this value; (2) otherwise, it is filled with a predefined value. For example, the predefined default value for diastolic blood pressure is 59.0, so the cell for time 0 in <xref rid="figure3" ref-type="fig">Figure 3</xref>C is filled with this value. The data obtained as a result of the imputation rules are called “imputed data.”</p>
          <p>In addition to imputing the missing cells, imputation also creates a mask. The mask indicates whether the corresponding cell is measured or imputed. Since missing information is filled in after the imputation step, the mask supplies meta-information that improves the accuracy of the ML model [<xref ref-type="bibr" rid="ref21">21</xref>-<xref ref-type="bibr" rid="ref23">23</xref>]. The last 2 columns in <xref rid="figure3" ref-type="fig">Figure 3</xref>C show the mask. Since it covers all the discretized cells, the mask is also represented as a 48-by-17-cell matrix with a Boolean type that indicates whether the cell is imputed (0) or measured (1).</p>
          <p>The use of these rules for emergency patient data can be justified for the following reasons: (1) In general, clinical variables do not change dramatically over a short period of time, and (2) using representative values (ie, defaults) for missing values is a frequently used approach in first aid. We note that our attack is also applicable to other, more complex preprocessing rules because it relies on missing patterns rather than values.</p>
        </sec>
      </sec>
      <sec>
        <title>Trigger Generation</title>
        <sec>
          <title>Trigger Generation With Random Masks: Illustrative Example</title>
          <p><xref rid="figure3" ref-type="fig">Figure 3</xref> also shows an example of generating trigger data. An attacker creates a trigger mask with random discrete values (<xref rid="figure3" ref-type="fig">Figure 3</xref>D) and adjusts the imputed data according to the trigger mask (<xref rid="figure3" ref-type="fig">Figure 3</xref>E). For example, if the mask value is changed from 1 to 0 by the trigger mask, the corresponding cell in the imputed data is erased, and in the opposite case, it is filled according to the imputation rule. The discretized trigger data are then restored to their raw-data form according to the data’s original time information, thereby generating trigger data.</p>
          <p>The number of possible trigger masks in this example is 2<sup>48×17</sup>. Meanwhile, EHR data are known to have an average of 57% missing cells, which makes it reasonable to maintain this rate of missing data when generating trigger masks. Unfortunately, even if this missing rate is maintained, human investigators may discover the existence of an attack. This is because emergency patient data from ICUs have a typical missing pattern, as shown in <xref rid="figure4" ref-type="fig">Figure 4</xref>A, whereas random generation can produce a mask (<xref rid="figure4" ref-type="fig">Figure 4</xref>B) different from the typical mask. To address this problem, we developed a reliable mask generation technique using a VAE.</p>
          <fig id="figure4" position="float">
            <label>Figure 4</label>
            <caption>
              <p>Three types of masks. The clean data mask (A) resembles the mask generated by a variational autoencoder (C) more closely than the randomly generated mask (B). VAE: variational autoencoder.</p>
            </caption>
            <graphic xlink:href="medinform_v10i8e38440_fig4.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>Trigger Generation With a VAE</title>
          <p>This section introduces an automation technique for generating trigger masks that are difficult to detect using a VAE [<xref ref-type="bibr" rid="ref31">31</xref>]. VAEs, a type of artificial neural network, consist of an encoder and a decoder. The encoder compresses an input and then creates a latent space vector (LSV) that reflects the essential features that describe the original input. The decoder reconstructs the original input from the LSV.</p>
          <p><xref rid="figure5" ref-type="fig">Figure 5</xref>A shows the training phase of the VAE. An attacker provides a clean mask to the encoder. The encoder compresses it into an LSV and simultaneously tunes the LSV to follow a normal distribution. The decoder reconstructs the original masks from the LSV. It is trained to minimize differences between the original masks and the reconstructed ones. Since the LSV provided by the encoder follows a normal distribution, the trained decoder can reconstruct masks similar to the clean masks from any random normally distributed LSV (<xref rid="figure5" ref-type="fig">Figure 5</xref>B). <xref rid="figure4" ref-type="fig">Figure 4</xref>C shows an example of a mask created by a VAE (ie, a VAE mask). It has a missing pattern that is visually similar to the clean mask.</p>
          <fig id="figure5" position="float">
            <label>Figure 5</label>
            <caption>
              <p>Training and generating phase of a variational autoencoder. (A) The variational autoencoder is trained to reconstruct clean masks. (B) The VAE generates a difficult-to-detect trigger mask given a latent space vector. VAE: variational autoencoder.</p>
            </caption>
            <graphic xlink:href="medinform_v10i8e38440_fig5.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>Experiment Settings</title>
        <p>We evaluated the performance of our attack from two perspectives: (1) attack efficacy and (2) stealthiness. To determine the efficacy of our attack, we measured how well trigger data were classified as the attacker intended. In the “Attack Efficacy” section of this paper, we describe 2 experiments that investigated “random poisoning” and “target poisoning.” To assess the stealthiness of the attack, we experimented with the visual similarity between the trigger data and the clean data (described in the “Stealthiness” section) and the impact of an attack on general classification performance (“Impact on Classification Performance” section). We also compare performance with an existing technique [<xref ref-type="bibr" rid="ref19">19</xref>] in the “Comparative Performance” section.</p>
        <p>Each experiment went through the following steps in a single trial: (1) Trigger data were generated and the labels were negated. (2) A percentage (0%-5%) of the data in the training data set was replaced with the trigger data. (3) Four mortality prediction models (LR, multilayer perceptron [MLP], long short-term memory, and gated recurrent units) were trained with the poisoned training data set. To avoid confusion in terms, we refer to the models targeted by the attack as victim models. (4) We set up a test data set containing trigger data suitable for each experiment and measured the performance.</p>
        <p>A description of the data set used in the experiment is provided in <xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>. Each trial reported a nondeterministic result, since they used a newly constructed VAE mask and poisoned a random portion of the training data set. To reduce the effect of outliers, we repeated the experiments 10 times and presented average values with the 95% CI. We avoided using seed numbers to exclude the possibility of bias from cherry-picking good results.</p>
        <p>There are 2 ways in which an attacker can manipulate outcomes: “false alarms” and “missing detection.” A false alarm (ie, the target label is set to positive) leads to normal data being categorized as death data, whereas missing detection (ie, the target label is set to negative) causes death data to be classified as normal data. For each experiment, we tested both cases and plotted them on a graph. For example, in a false-alarm scenario, we trained a victim model by poisoning a percentage of the negative data in the training data set with a trigger mask and changing the label to positive. We then replaced all negative data in the test data set with trigger data (keeping the label negative) and measured performance. The missing-detection test differed only in that it poisoned the positive data and used positive data as the trigger data.</p>
      </sec>
      <sec>
        <title>Attack Efficacy</title>
        <p>We estimated the effectiveness of the proposed backdoor attack with the following method. Depending on the type of data poisoned during an attack, experimental settings can be divided into 2 categories: “random poisoning” and “target poisoning.” Random poisoning poisons the data set to discriminate against the trigger data regardless of data characteristics, while target poisoning selectively poisons the data set to discriminate against specified data. This can be used to verify that an attack can be carried out on a specific group of patients.</p>
        <sec>
          <title>Discrimination Performance in Random Poisoning</title>
          <p>In a random-poisoning setting, a victim model is trained with a percentage of trigger data randomly selected in the training data set. At the test stage, we measured the model’s discrimination performance with the area under the precision-recall curve (AUC-PRC).</p>
          <p>The AUC-PRC [<xref ref-type="bibr" rid="ref32">32</xref>] is a well-known metric used to evaluate binary classifiers that provides reliable scores, especially for imbalanced data sets (positive-data groups are small). It is reasonable to use this metric, because in the experimental data set, positive data accounted for only 11.5% of the test data set due to the nature of mortality prediction. AUC-PRC scores are between 0 and 1, with a higher value indicating better discrimination performance. Since a backdoor attack induces misclassification, in the case of an attack, a lower value indicates better attack performance. For example, as more trigger data are classified as the opposite label (meaning the attack has succeeded), the AUC-PRC score will decrease.</p>
          <p><xref rid="figure6" ref-type="fig">Figure 6</xref> shows the AUC-PRC of 4 victim models when the poisoning ratio of a training data set increased from 0% to 5%. <xref rid="figure6" ref-type="fig">Figure 6</xref>A shows the outcome of a false alarm, and <xref rid="figure6" ref-type="fig">Figure 6</xref>B shows the outcome of a missing detection with the 95% CI for 10 attempts. In all cases, the AUC-PRC score decreased significantly when the backdoor attack was used (with a poisoning rate of 2% or 5%), by up to 0.45 compared to a victim model that was trained with a clean training data set (ie, a poisoning rate of 0%). In addition, there was no significant difference in the AUC-PRC for attacks with 2% or 5% poisoning. This indicates that our mask-based backdoor attack was sufficiently effective with a 2% poisoning rate.</p>
          <p>The red horizontal line indicates the AUC-PRC score when a random classifier was trained with the same training data set containing the same quantity of negative and positive data. Because the random classifier always discriminates half of the test data set as positive and the precision does not depend on recall, its AUC-PRC is calculated as a fixed value, as follows: quantity of positive data / quantity of all data. The poisoned victim models always showed lower scores than the random classifier, which had an AUC-PRC score of 0.115, demonstrating that the attack was remarkably effective.</p>
          <fig id="figure6" position="float">
            <label>Figure 6</label>
            <caption>
              <p>The discrimination performance of 4 victim models with random poisoning for (A) false alarm and (B) missing detection scenarios. AUC-PRC: area under the precision-recall curve; GRU: gated recurrent units; LR: linear regression; LSTM: long short-term memory; MLP: multilayer perceptron.</p>
            </caption>
            <graphic xlink:href="medinform_v10i8e38440_fig6.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>Discrimination Performance in Target Poisoning</title>
          <p>Target poisoning determines the effectiveness of a mask-based backdoor attack on specific data. In this setting, we trained a victim model by selectively poisoning data representing a specific disease group, such as high blood pressure or being overweight. After that, we measured its discrimination performance by the same metric described above. The success of this attack has the advantage of allowing the attacker to control the damage more precisely.</p>
          <p>The overall attack process is as follows. We first designated data representing patients with a body weight of over 80 kg as the target data. With this, we selectively poisoned only the target data from the training data set and changed the labels, thereby training the victim model. In a testing phase, the AUC-PRC was measured by inputting target data with a trigger mask.</p>
          <p>It was possible that this poisoning process, however, might have not only triggered the target data but also triggered any data with a trigger mask. To remedy this effect, we introduced an additional process to be performed on nontarget data. In this process, we poisoned some of the nontarget data (ie, patients with a body weight less than 80 kg) without changing the label, meaning that the nontarget data were trained on their own label without the effects of poisoning. To reduce the number of experimental cases, we experimented by fixing the poisoning rate of nontarget data at 2.5%.</p>
          <p><xref rid="figure7" ref-type="fig">Figure 7</xref> shows the result. When a nontarget group was trained without a trigger mask (<xref rid="figure7" ref-type="fig">Figure 7</xref>A and B), both target and nontarget data were affected by the attack (reducing the AUC-PRC score). On the other hand, when the nontarget group was trained to have its original label on the trigger mask (<xref rid="figure7" ref-type="fig">Figure 7</xref>C and D), the target poisoning attack was more pronounced (as we intended). In the latter case, the AUC-PRC scores of all victim models for the target data were lower than those of the random classifier, except for LP and MLP (<xref rid="figure7" ref-type="fig">Figure 7</xref>D). Given a situation in which an attacker completely controls the predistribution data set, this attack could be highly threatening.</p>
          <fig id="figure7" position="float">
            <label>Figure 7</label>
            <caption>
              <p>The discrimination performance of 4 victim models when only target data was poisoned for (A) false alarm and (B) missing data scenarios, and when both target and nontarget data were poisoned for (C) false alarm and (D) missing data scenarios. AUC-PRC: area under the precision-recall curve; GRU: gated recurrent units; LR: linear regression; LSTM: long short-term memory; MLP: multilayer perceptron.</p>
            </caption>
            <graphic xlink:href="medinform_v10i8e38440_fig7.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
      </sec>
      <sec>
        <title>Stealthiness</title>
        <sec>
          <title>Mask Similarity</title>
          <p>In order to prevent an attack from being detected, it is important to make sure that the trigger data are visually similar to clean data. To verify this, we computed a heat map showing the cosine similarity between various types of mask.</p>
          <p>The cosine similarity is calculated by the cosine of the angle between the two vectors. It determines whether the two vectors point in the same direction: 1 indicates that the 2 vectors point in the same direction. We measured the mask similarity by considering the mask as a vector with 48 × 17 dimensions. For the experiment, we used 3 types of mask: clean, VAE, and random. For each type, we created 100 masks and represented them in a 300 × 300 heat map. The heat map was symmetrical, and the (<italic>i, j</italic>) elements of the heat map showed cosine similarity between the <italic>i</italic>th and <italic>j</italic>th masks.</p>
          <p><xref rid="figure8" ref-type="fig">Figure 8</xref> clearly shows that the VAE masks had a closer similarity to the clean masks than to the random masks. In particular, we calculated the threshold based on the top p percentile of the elements in the sub–heat map of the clean mask (shown by the red solid-line rectangle in <xref rid="figure8" ref-type="fig">Figure 8</xref>) and measured the ratio of elements above this threshold in the sub–heat map of the clean mask minus the VAE mask (shown by the red dashed-line rectangle in <xref rid="figure8" ref-type="fig">Figure 8</xref>). The result was 0.45 for the 50th percentile and 0.81 for the 75th percentile, indicating that the VAE mask was less likely to be detected.</p>
          <fig id="figure8" position="float">
            <label>Figure 8</label>
            <caption>
              <p>Cosine similarity heat map between 3 types of masks: clean, variational autoencoder, and random. VAE: variational autoencoder.</p>
            </caption>
            <graphic xlink:href="medinform_v10i8e38440_fig8.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>Impact on Classification Performance</title>
          <p>The backdoor should not affect classification performance. Otherwise, a user might detect the existence of an attack. Therefore, we measured the discrimination performance of victim models that used a clean test data set, and in addition to using the AUC-PRC, we evaluated the difference between the poisoned and clean models using a calibration curve [<xref ref-type="bibr" rid="ref33">33</xref>].</p>
          <p><xref rid="figure9" ref-type="fig">Figure 9</xref> shows the AUC-PRC for the 4 victim models when the training data set was poisoned at rates of 0%, 2%, and 5%. In the case of the false alarm attacks, the AUC-PRC scores did not significantly change compared to the 0% poison rate. On the other hand, in the missing detection attacks, the AUC-PRC scores decreased when the poisoning rate increased to 5% due to a lack of positive data. In the mortality prediction data set, positive data only accounted for 13.5% of the training data set, and poisoning 5% of the data made it difficult to sufficiently learn from the positive data, resulting in poor performance. Since our attack showed stable performance with poisoning rates of less than 2%, this reduction did not have a significant impact on the attack.</p>
          <p><xref rid="figure10" ref-type="fig">Figure 10</xref> shows the calibration curves [<xref ref-type="bibr" rid="ref33">33</xref>] that represent the reliability of the prediction probabilities of the input model. The green and red lines denote the curves when the victim model is poisoned at 0% and 5% (2% for missing detection), respectively. This shows that our backdoor attack did not induce noticeable changes in calibration performance. The maximum difference between the two curves is 0.04, when the x values are the same (attack: missing detection; model: LR; x: 0.48), which makes it difficult for victims to notice the difference.</p>
          <fig id="figure9" position="float">
            <label>Figure 9</label>
            <caption>
              <p>The discrimination performance of the 4 victim models on a clean test data set for (A) false alarm and (B) missing data scenarios. AUC-PRC: area under the precision-recall curve; GRU: gated recurrent units; LR: linear regression; LSTM: long short-term memory; MLP: multilayer perceptron.</p>
            </caption>
            <graphic xlink:href="medinform_v10i8e38440_fig9.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
          <fig id="figure10" position="float">
            <label>Figure 10</label>
            <caption>
              <p>Calibration curves before and after our backdoor attack. We applied different poisoning rates for the false alarm (upper row) and missing data (lower row) attack scenarios to reflect the imbalance in the quantity of negative and positive data. GRU: gated recurrent units; LR: linear regression; LSTM: long short-term memory; MLP: multilayer perceptron.</p>
            </caption>
            <graphic xlink:href="medinform_v10i8e38440_fig10.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
      </sec>
      <sec>
        <title>Comparative Performance</title>
        <p>We compared our approach with an existing noise-based backdoor approach (reported by Joe et al [<xref ref-type="bibr" rid="ref19">19</xref>]) that conducts a backdoor attack on EHR mortality classification models. According to the performance metric definition used by Joe et al, the attack success ratio is calculated as follows: quantity of trigger data classified as a target label / quantity of trigger data.</p>
        <p>The result is summarized in <xref rid="figure11" ref-type="fig">Figure 11</xref>. Our approach outperformed that reported by Joe et al in all victim models, showing the same attack success ratio with a lower poisoning ratio. For example, our attack required only a 0.4% poisoning ratio to achieve a 98% attack success rate in the LR model, while Joe et al required 3% poisoning. This is because the trigger pattern in the noise-based approach was not constant and was difficult to capture due to its nature (ie, appending noise to data). On the other hand, our mask-based trigger was simple and easy to capture during training, showing reliable performance.</p>
        <fig id="figure11" position="float">
          <label>Figure 11</label>
          <caption>
            <p>Attack success rates for a mask-based backdoor attack (ours) and a noise-based backdoor attack (Joe et al [<xref ref-type="bibr" rid="ref19">19</xref>]) on 4 machine learning models. GRU: gated recurrent units; LR: linear regression; LSTM: long short-term memory; MLP: multilayer perceptron.</p>
          </caption>
          <graphic xlink:href="medinform_v10i8e38440_fig11.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <sec>
        <title>Principal Findings</title>
        <p>To the best of our knowledge, this is the first study to introduce an ML backdoor attack based on meta-information. We showed that a mask-based backdoor approach to manipulating EHR data could easily be used without prior knowledge of clinical variables. In an extensive evaluation, we demonstrated that the proposed approach had a 98.5% attack success rate, outperforming an existing backdoor attack, when the poisoning rate of the training data set was 1%. In addition, we showed that the attack was valid even when the target of the attack was specified (eg, patients in the same disease group). Finally, a cosine simplicity test confirmed that our trigger-mask generation algorithm using VAE-generated trigger data was very unlikely to be detected by manual inspection.</p>
      </sec>
      <sec>
        <title>Comparison With Prior Work</title>
        <p>Early studies showed that backdoor attacks on image classifiers were feasible [<xref ref-type="bibr" rid="ref20">20</xref>,<xref ref-type="bibr" rid="ref34">34</xref>,<xref ref-type="bibr" rid="ref35">35</xref>]. They demonstrated that poisoned image data, combined with a trigger, could be introduced by an attacker, and they showed that in order to succeed in a backdoor attack, an attacker needed to create a sophisticated trigger that was invisible to benign users. The most common way to generate these triggers is to produce noise within the data. Many follow-up studies [<xref ref-type="bibr" rid="ref36">36</xref>-<xref ref-type="bibr" rid="ref38">38</xref>] revealed techniques to achieve high attack success rates with imperceptible noise that minimized detection.</p>
        <p>Unlike image data, it is difficult to apply existing noise-generation techniques to the tabular data used for EHRs. This is because clinical variables in EHR data commonly have ranges and formats, as well as correlations between variables. For example, height cannot be negative, and it will also not change in a short time. Joe et al [<xref ref-type="bibr" rid="ref19">19</xref>] addressed this difficulty by proposing a noise-based backdoor attack on a medical ML model that reflected the characteristics of EHR data. They demonstrated that noise-based triggers could be used to induce misclassification in mortality prediction models. However, this attack method requires prior knowledge of clinical variables to calculate noise and requires a higher poisoning rate for attack success, because noise can only be applied to measured cells.</p>
        <p>On the other hand, our mask-based approach can easily generate trigger data by simply eliminating or filling in values. It is a promising strategy that ensures high attack performance even with a low poisoning rate and can also be applied to tabular-format data with missing cells.</p>
      </sec>
      <sec>
        <title>Limitations</title>
        <p>Although our attack is effective, there are several limitations. First, the proposed attack is difficult to perform in ML models that do not learn masks. Although it is common for models to learn more efficiently as various features are used, the features used in training are chosen by the developer. Therefore, masks may not be learned in mortality prediction models. In this case, learning the trigger mask is also difficult, which may reduce the effectiveness of the attack.</p>
        <p>Second, our VAE-based mask generation algorithm requires more computational time in some cases to generate trigger data than the existing method [<xref ref-type="bibr" rid="ref19">19</xref>]. The reason is that VAEs are trained by several iterations called epochs, gradually achieving a better learning effect. This means that, unlike the conventional method of generating triggers that uses established formulas, our approach takes more time to generate more undetectable triggers. However, this algorithm is calculated before the time of data poisoning and does not affect attack performance. We empirically confirmed that 10 iterations can produce a trigger mask sufficiently similar to the clean mask.</p>
      </sec>
      <sec>
        <title>Conclusions</title>
        <p>In this paper, we present a new mask-based backdoor attack that manipulates missing patterns in EHR data. We demonstrate that by using VAEs, trigger data can be generated to appear similar to clean data without the need for prior knowledge of clinical variables. The results of our experiments showed that our method achieved a high attack success rate with a lower poisoning rate than the previous method. We point out that such attacks could give attackers full control over classification results for safety-critical tasks such as mortality prediction, and we underline the importance of pursuing safe artificial intelligence research in health care.</p>
      </sec>
    </sec>
  </body>
  <back>
    <app-group>
      <supplementary-material id="app1">
        <label>Multimedia Appendix 1</label>
        <p>MIMIC-III Data Statistics.</p>
        <media xlink:href="medinform_v10i8e38440_app1.docx" xlink:title="DOCX File , 26 KB"/>
      </supplementary-material>
      <supplementary-material id="app2">
        <label>Multimedia Appendix 2</label>
        <p>Experiment Settings.</p>
        <media xlink:href="medinform_v10i8e38440_app2.docx" xlink:title="DOCX File , 24 KB"/>
      </supplementary-material>
    </app-group>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">AUC-PRC</term>
          <def>
            <p>area under the precision-recall curve</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">DNN</term>
          <def>
            <p>deep neural network</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">EHR</term>
          <def>
            <p>electronic health record</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">ICU</term>
          <def>
            <p>intensive care unit</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">LR</term>
          <def>
            <p>linear regression</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">LSTM</term>
          <def>
            <p>long short-term memory</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb7">LSV</term>
          <def>
            <p>latent space vector</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb8">ML</term>
          <def>
            <p>machine learning</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb9">MLP</term>
          <def>
            <p>multilayer perceptron</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb10">VAE</term>
          <def>
            <p>variational autoencoder</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <ack>
      <p>This work was supported in part by the Institute for Information &#38; Communication Technology Planning &#38; Evaluation (2020-0-00209, 2019-0-01343: Regional Strategic Industry Convergence Security Core Talent Training Business) and funded by the Korean Ministry of Science and Information &#38; Communication Technology.</p>
    </ack>
    <fn-group>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Singh</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Handa</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Kumar</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Shukla</surname>
              <given-names>SK</given-names>
            </name>
          </person-group>
          <article-title>Malware Analysis Using Image Classification Techniques</article-title>
          <source>Cyber Security in India: IITK Directions, vol 4</source>
          <year>2020</year>
          <publisher-loc>Singapore</publisher-loc>
          <publisher-name>Springer</publisher-name>
          <fpage>33</fpage>
          <lpage>38</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Roy</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Banville</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Albuquerque</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Gramfort</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Falk</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Faubert</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Deep learning-based electroencephalography analysis: a systematic review</article-title>
          <source>J Neural Eng</source>
          <year>2019</year>
          <month>08</month>
          <day>14</day>
          <volume>16</volume>
          <issue>5</issue>
          <fpage>051001</fpage>
          <pub-id pub-id-type="doi">10.1088/1741-2552/ab260c</pub-id>
          <pub-id pub-id-type="medline">31151119</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bird</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Manso</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Ribeiro</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Ekárt</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Faria</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>A study on mental state classification using EEG-based brain-machine interface</article-title>
          <year>2018</year>
          <conf-name>International Conference on Intelligent Systems (IS)</conf-name>
          <conf-date>Sep 25-27, 2018</conf-date>
          <conf-loc>Funchal, Portugal</conf-loc>
          <fpage>795</fpage>
          <lpage>800</lpage>
          <pub-id pub-id-type="doi">10.1109/is.2018.8710576</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yisroel</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Tomer</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Yuval</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Asaf</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Kitsune: An Ensemble of Autoencoders for Online Network Intrusion Detection</article-title>
          <year>2018</year>
          <conf-name>Network and Distributed System Security (NDSS) Symposium</conf-name>
          <conf-date>Feb 18-21, 2018</conf-date>
          <conf-loc>San Diego, CA</conf-loc>
          <pub-id pub-id-type="doi">10.14722/ndss.2018.23204</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Raff</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Barker</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Sylvester</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Brandon</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Catanzaro</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Nicholas</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Malware detection by eating a whole exe</article-title>
          <year>2018</year>
          <conf-name>Workshops at the Thirty-Second AAAI Conference on Artificial Intelligence</conf-name>
          <conf-date>Feb 2-7, 2018</conf-date>
          <conf-loc>New Orleans, LA</conf-loc>
          <pub-id pub-id-type="doi">10.48550/arXiv.1710.09435</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kadampur</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Al Riyaee</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Skin cancer detection: Applying a deep learning based model driven architecture in the cloud for classifying dermal cell images</article-title>
          <source>Inform Med Unlocked</source>
          <year>2020</year>
          <volume>18</volume>
          <fpage>100282</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.sciencedirect.com/science/article/pii/S2352914819302047/pdfft?md5=c9f24f17c91fe37bd4dade828e928a40&#38;pid=1-s2.0-S2352914819302047-main.pdf"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.imu.2019.100282</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Muckley</surname>
              <given-names>MJ</given-names>
            </name>
            <name name-style="western">
              <surname>Ades-Aron</surname>
              <given-names>Benjamin</given-names>
            </name>
            <name name-style="western">
              <surname>Papaioannou</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Lemberskiy</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Solomon</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Lui</surname>
              <given-names>YW</given-names>
            </name>
            <name name-style="western">
              <surname>Sodickson</surname>
              <given-names>DK</given-names>
            </name>
            <name name-style="western">
              <surname>Fieremans</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Novikov</surname>
              <given-names>DS</given-names>
            </name>
            <name name-style="western">
              <surname>Knoll</surname>
              <given-names>F</given-names>
            </name>
          </person-group>
          <article-title>Training a neural network for Gibbs and noise removal in diffusion MRI</article-title>
          <source>Magn Reson Med</source>
          <year>2021</year>
          <month>01</month>
          <volume>85</volume>
          <issue>1</issue>
          <fpage>413</fpage>
          <lpage>428</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/32662910"/>
          </comment>
          <pub-id pub-id-type="doi">10.1002/mrm.28395</pub-id>
          <pub-id pub-id-type="medline">32662910</pub-id>
          <pub-id pub-id-type="pmcid">PMC7722184</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Harutyunyan</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Khachatrian</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Kale</surname>
              <given-names>DC</given-names>
            </name>
            <name name-style="western">
              <surname>Ver Steeg</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Galstyan</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Multitask learning and benchmarking with clinical time series data</article-title>
          <source>Sci Data</source>
          <year>2019</year>
          <month>06</month>
          <day>17</day>
          <volume>6</volume>
          <issue>1</issue>
          <fpage>96</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/31209213"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41597-019-0103-9</pub-id>
          <pub-id pub-id-type="medline">31209213</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41597-019-0103-9</pub-id>
          <pub-id pub-id-type="pmcid">PMC6572845</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ayala Solares</surname>
              <given-names>JR</given-names>
            </name>
            <name name-style="western">
              <surname>Diletta Raimondi</surname>
              <given-names>FE</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Rahimian</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Canoy</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Tran</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Pinho Gomes</surname>
              <given-names>AC</given-names>
            </name>
            <name name-style="western">
              <surname>Payberah</surname>
              <given-names>AH</given-names>
            </name>
            <name name-style="western">
              <surname>Zottoli</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Nazarzadeh</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Conrad</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Rahimi</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Salimi-Khorshidi</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>Deep learning for electronic health records: A comparative review of multiple deep neural architectures</article-title>
          <source>J Biomed Inform</source>
          <year>2020</year>
          <month>01</month>
          <volume>101</volume>
          <fpage>103337</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S1532-0464(19)30256-4"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.jbi.2019.103337</pub-id>
          <pub-id pub-id-type="medline">31916973</pub-id>
          <pub-id pub-id-type="pii">S1532-0464(19)30256-4</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kanevsky</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Corban</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Gaster</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Kanevsky</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Gilardino</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Big Data and Machine Learning in Plastic Surgery: A New Frontier in Surgical Innovation</article-title>
          <source>Plast Reconstr Surg</source>
          <year>2016</year>
          <month>05</month>
          <volume>137</volume>
          <issue>5</issue>
          <fpage>890e</fpage>
          <lpage>897e</lpage>
          <pub-id pub-id-type="doi">10.1097/PRS.0000000000002088</pub-id>
          <pub-id pub-id-type="medline">27119951</pub-id>
          <pub-id pub-id-type="pii">00006534-201605000-00045</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Choy</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Khalilzadeh</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Michalski</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Do</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Samir</surname>
              <given-names>AE</given-names>
            </name>
            <name name-style="western">
              <surname>Pianykh</surname>
              <given-names>OS</given-names>
            </name>
            <name name-style="western">
              <surname>Geis</surname>
              <given-names>JR</given-names>
            </name>
            <name name-style="western">
              <surname>Pandharipande</surname>
              <given-names>PV</given-names>
            </name>
            <name name-style="western">
              <surname>Brink</surname>
              <given-names>JA</given-names>
            </name>
            <name name-style="western">
              <surname>Dreyer</surname>
              <given-names>KJ</given-names>
            </name>
          </person-group>
          <article-title>Current Applications and Future Impact of Machine Learning in Radiology</article-title>
          <source>Radiology</source>
          <year>2018</year>
          <month>08</month>
          <volume>288</volume>
          <issue>2</issue>
          <fpage>318</fpage>
          <lpage>328</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/29944078"/>
          </comment>
          <pub-id pub-id-type="doi">10.1148/radiol.2018171820</pub-id>
          <pub-id pub-id-type="medline">29944078</pub-id>
          <pub-id pub-id-type="pmcid">PMC6542626</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tsang</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Kracov</surname>
              <given-names>DA</given-names>
            </name>
            <name name-style="western">
              <surname>Mulryne</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Strom</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Perkins</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Dickinson</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>The impact of artificial intelligence on medical innovation in the European Union and United States</article-title>
          <source>Intellect Prop Technol Law J</source>
          <year>2017</year>
          <volume>29</volume>
          <issue>8</issue>
          <fpage>3</fpage>
          <lpage>12</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.arnoldporter.com/-/media/files/perspectives/publications/2017/08/the-impact-of-artificial-inteelligence-on-medical-innovation.pdf?"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Guo</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Dong</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Gao</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Simple convolutional neural network on image classification</article-title>
          <year>2017</year>
          <conf-name>IEEE 2nd International Conference on Big Data Analysis (ICBDA)</conf-name>
          <conf-date>Mar 10-12, 2017</conf-date>
          <conf-loc>Beijing, China</conf-loc>
          <fpage>721</fpage>
          <lpage>724</lpage>
          <pub-id pub-id-type="doi">10.1109/icbda.2017.8078730</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Song</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Song</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Du</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>A neural network model for Chinese sentence generation with key word</article-title>
          <year>2019</year>
          <conf-name>IEEE 9th International Conference on Electronics Information and Emergency Communication (ICEIEC)</conf-name>
          <conf-date>Jul 12-14, 2019</conf-date>
          <conf-loc>Beijing, China</conf-loc>
          <fpage>334</fpage>
          <lpage>337</lpage>
          <pub-id pub-id-type="doi">10.1109/iceiec.2019.8784475</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Islam</surname>
              <given-names>MS</given-names>
            </name>
            <name name-style="western">
              <surname>Sharmin Mousumi</surname>
              <given-names>SS</given-names>
            </name>
            <name name-style="western">
              <surname>Abujar</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Hossain</surname>
              <given-names>SA</given-names>
            </name>
          </person-group>
          <article-title>Sequence-to-sequence Bangla Sentence Generation with LSTM Recurrent Neural Networks</article-title>
          <source>Procedia Comput Sci</source>
          <year>2019</year>
          <volume>152</volume>
          <fpage>51</fpage>
          <lpage>58</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://reader.elsevier.com/reader/sd/pii/S1877050919306775?token=854A7BA7A670D583608DC730B7F7077B2A35FD7E82874E702367459D16D1746D0B3D0BE4EED4B8DD66DC28F7B3888C29&#38;originRegion=us-east-1&#38;originCreation=20220719180535"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.procs.2019.05.026</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Biggio</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Nelson</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Laskov</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>Poisoning attacks against support vector machines</article-title>
          <source>Proceedings of the 29th International Conference on Machine Learning (ICML 12)</source>
          <year>2012</year>
          <conf-name>29th International Conference on Machine Learning (ICML 12)</conf-name>
          <conf-date>Jun 26–Jul 1, 2012</conf-date>
          <conf-loc>Edinburgh, Scotland</conf-loc>
          <fpage>9781450312851</fpage>
          <pub-id pub-id-type="doi">10.48550/arXiv.1206.6389</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Gu</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Dolan-Gavitt</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Garg</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>BadNets: Evaluating Backdooring Attacks on Deep Neural Networks</article-title>
          <source>IEEE Access</source>
          <year>2019</year>
          <volume>7</volume>
          <fpage>47230</fpage>
          <lpage>47244</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&#38;arnumber=8685687"/>
          </comment>
          <pub-id pub-id-type="doi">10.1109/access.2019.2909068</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Goodfellow</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Shlens</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Szegedy</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Explaining and harnessing adversarial examples</article-title>
          <year>2015</year>
          <conf-name>International Conference on Learning Representations (ICLR)</conf-name>
          <conf-date>May 7-9, 2015</conf-date>
          <conf-loc>San Diego, CA</conf-loc>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Joe</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Mehra</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Shin</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Hamm</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Machine Learning with Electronic Health Records is vulnerable to Backdoor Trigger Attacks</article-title>
          <year>2021</year>
          <conf-name>AAAI Workshop on Trustworthy AI for Healthcare</conf-name>
          <conf-date>Feb 9, 2021</conf-date>
          <conf-loc>Online</conf-loc>
          <pub-id pub-id-type="doi">10.48550/arXiv.2106.07925</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yao</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Zheng</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>Latent backdoor attacks on deep neural networks</article-title>
          <source>Proceedings of the ACM SIGSAC Conference on Computer and Communications Security (CCS)</source>
          <year>2019</year>
          <conf-name>ACM SIGSAC Conference on Computer and Communications Security (CCS)</conf-name>
          <conf-date>Nov 11-15, 2019</conf-date>
          <conf-loc>London, UK</conf-loc>
          <fpage>2041</fpage>
          <lpage>2055</lpage>
          <pub-id pub-id-type="doi">10.1145/3319535.3354209</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lipton</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Kale</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Wetzel</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Modeling missing data in clinical time series with RNNs</article-title>
          <year>2016</year>
          <conf-name>Machine Learning for Healthcare</conf-name>
          <conf-date>Aug 19-20, 2016</conf-date>
          <conf-loc>Los Angeles, CA</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://proceedings.mlr.press/v56/Lipton16.pdf"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Josse</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Prost</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Scornet</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Varoquaux</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>On the consistency of supervised learning with missing values</article-title>
          <source>ArXiv. Preprint posted online on Feb 19, 2019</source>
          <year>2019</year>
          <pub-id pub-id-type="doi">10.48550/arXiv.1902.06931</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Groenwold</surname>
              <given-names>RHH</given-names>
            </name>
            <name name-style="western">
              <surname>White</surname>
              <given-names>IR</given-names>
            </name>
            <name name-style="western">
              <surname>Donders</surname>
              <given-names>ART</given-names>
            </name>
            <name name-style="western">
              <surname>Carpenter</surname>
              <given-names>JR</given-names>
            </name>
            <name name-style="western">
              <surname>Altman</surname>
              <given-names>DG</given-names>
            </name>
            <name name-style="western">
              <surname>Moons</surname>
              <given-names>KGM</given-names>
            </name>
          </person-group>
          <article-title>Missing covariate data in clinical research: when and when not to use the missing-indicator method for analysis</article-title>
          <source>CMAJ</source>
          <year>2012</year>
          <month>08</month>
          <day>07</day>
          <volume>184</volume>
          <issue>11</issue>
          <fpage>1265</fpage>
          <lpage>9</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://www.cmaj.ca/cgi/pmidlookup?view=long&#38;pmid=22371511"/>
          </comment>
          <pub-id pub-id-type="doi">10.1503/cmaj.110977</pub-id>
          <pub-id pub-id-type="medline">22371511</pub-id>
          <pub-id pub-id-type="pii">cmaj.110977</pub-id>
          <pub-id pub-id-type="pmcid">PMC3414599</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jones</surname>
              <given-names>MP</given-names>
            </name>
          </person-group>
          <article-title>Indicator and Stratification Methods for Missing Explanatory Variables in Multiple Linear Regression</article-title>
          <source>J Am Stat Assoc</source>
          <year>1996</year>
          <month>03</month>
          <volume>91</volume>
          <issue>433</issue>
          <fpage>222</fpage>
          <lpage>230</lpage>
          <pub-id pub-id-type="doi">10.1080/01621459.1996.10476680</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Ye</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Tan</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Yuen</surname>
              <given-names>PC</given-names>
            </name>
          </person-group>
          <article-title>Cross-Domain Missingness-Aware Time-Series Adaptation With Similarity Distillation in Medical Applications</article-title>
          <source>IEEE Trans Cybern</source>
          <year>2022</year>
          <month>05</month>
          <volume>52</volume>
          <issue>5</issue>
          <fpage>3394</fpage>
          <lpage>3407</lpage>
          <pub-id pub-id-type="doi">10.1109/TCYB.2020.3011934</pub-id>
          <pub-id pub-id-type="medline">32795976</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Khan</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Hoque</surname>
              <given-names>Abu Sayed Md Latiful</given-names>
            </name>
          </person-group>
          <article-title>SICE: an improved missing data imputation technique</article-title>
          <source>J Big Data</source>
          <year>2020</year>
          <volume>7</volume>
          <issue>1</issue>
          <fpage>37</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/32547903"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/s40537-020-00313-w</pub-id>
          <pub-id pub-id-type="medline">32547903</pub-id>
          <pub-id pub-id-type="pii">313</pub-id>
          <pub-id pub-id-type="pmcid">PMC7291187</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref27">
        <label>27</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ayilara</surname>
              <given-names>OF</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Sajobi</surname>
              <given-names>TT</given-names>
            </name>
            <name name-style="western">
              <surname>Sawatzky</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Bohm</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Lix</surname>
              <given-names>LM</given-names>
            </name>
          </person-group>
          <article-title>Impact of missing data on bias and precision when estimating change in patient-reported outcomes from a clinical registry</article-title>
          <source>Health Qual Life Outcomes</source>
          <year>2019</year>
          <month>06</month>
          <day>20</day>
          <volume>17</volume>
          <issue>1</issue>
          <fpage>106</fpage>
          <lpage>9</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://hqlo.biomedcentral.com/articles/10.1186/s12955-019-1181-2"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/s12955-019-1181-2</pub-id>
          <pub-id pub-id-type="medline">31221151</pub-id>
          <pub-id pub-id-type="pii">10.1186/s12955-019-1181-2</pub-id>
          <pub-id pub-id-type="pmcid">PMC6585083</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref28">
        <label>28</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Johnson</surname>
              <given-names>AEW</given-names>
            </name>
            <name name-style="western">
              <surname>Pollard</surname>
              <given-names>TJ</given-names>
            </name>
            <name name-style="western">
              <surname>Shen</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Lehman</surname>
              <given-names>LH</given-names>
            </name>
            <name name-style="western">
              <surname>Feng</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Ghassemi</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Moody</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Szolovits</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Celi</surname>
              <given-names>LA</given-names>
            </name>
            <name name-style="western">
              <surname>Mark</surname>
              <given-names>RG</given-names>
            </name>
          </person-group>
          <article-title>MIMIC-III, a freely accessible critical care database</article-title>
          <source>Sci Data</source>
          <year>2016</year>
          <month>05</month>
          <day>24</day>
          <volume>3</volume>
          <fpage>160035</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/27219127"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/sdata.2016.35</pub-id>
          <pub-id pub-id-type="medline">27219127</pub-id>
          <pub-id pub-id-type="pii">sdata201635</pub-id>
          <pub-id pub-id-type="pmcid">PMC4878278</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref29">
        <label>29</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tan</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Ye</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Ma</surname>
              <given-names>AJ</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Yip</surname>
              <given-names>TC</given-names>
            </name>
            <name name-style="western">
              <surname>Wong</surname>
              <given-names>GL</given-names>
            </name>
            <name name-style="western">
              <surname>Yuen</surname>
              <given-names>PC</given-names>
            </name>
          </person-group>
          <article-title>Explainable Uncertainty-Aware Convolutional Recurrent Neural Network for Irregular Medical Time Series</article-title>
          <source>IEEE Trans Neural Netw Learn Syst</source>
          <year>2021</year>
          <month>10</month>
          <volume>32</volume>
          <issue>10</issue>
          <fpage>4665</fpage>
          <lpage>4679</lpage>
          <pub-id pub-id-type="doi">10.1109/TNNLS.2020.3025813</pub-id>
          <pub-id pub-id-type="medline">33055037</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref30">
        <label>30</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Shickel</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Tighe</surname>
              <given-names>PJ</given-names>
            </name>
            <name name-style="western">
              <surname>Bihorac</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Rashidi</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>Deep EHR: A Survey of Recent Advances in Deep Learning Techniques for Electronic Health Record (EHR) Analysis</article-title>
          <source>IEEE J Biomed Health Inform</source>
          <year>2018</year>
          <month>09</month>
          <volume>22</volume>
          <issue>5</issue>
          <fpage>1589</fpage>
          <lpage>1604</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/29989977"/>
          </comment>
          <pub-id pub-id-type="doi">10.1109/JBHI.2017.2767063</pub-id>
          <pub-id pub-id-type="medline">29989977</pub-id>
          <pub-id pub-id-type="pmcid">PMC6043423</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref31">
        <label>31</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kingma</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Welling</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Auto-encoding variational bayes</article-title>
          <year>2014</year>
          <conf-name>International Conference on Learning Representations (ICLR)</conf-name>
          <conf-date>Apr 14-16, 2014</conf-date>
          <conf-loc>Banff, AB</conf-loc>
        </nlm-citation>
      </ref>
      <ref id="ref32">
        <label>32</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Raghavan</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Bollmann</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Jung</surname>
              <given-names>GS</given-names>
            </name>
          </person-group>
          <article-title>A critical investigation of recall and precision as measures of retrieval system performance</article-title>
          <source>ACM Trans Inf Syst</source>
          <year>1989</year>
          <month>07</month>
          <volume>7</volume>
          <issue>3</issue>
          <fpage>205</fpage>
          <lpage>229</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://dl.acm.org/doi/pdf/10.1145/65943.65945"/>
          </comment>
          <pub-id pub-id-type="doi">10.1145/65943.65945</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref33">
        <label>33</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>DeGroot</surname>
              <given-names>MH</given-names>
            </name>
            <name name-style="western">
              <surname>Fienberg</surname>
              <given-names>SE</given-names>
            </name>
          </person-group>
          <article-title>The Comparison and Evaluation of Forecasters</article-title>
          <source>J Roy Stat Soc D-Sta</source>
          <year>1983</year>
          <month>03</month>
          <volume>32</volume>
          <issue>1/2</issue>
          <fpage>12</fpage>
          <pub-id pub-id-type="doi">10.2307/2987588</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref34">
        <label>34</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Song</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Targeted backdoor attacks on deep learning systems using data poisoning</article-title>
          <source>ArXiv Preprint posted online on Dec 15, 2017</source>
          <year>2022</year>
          <pub-id pub-id-type="doi">10.48550/arXiv.1712.05526</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref35">
        <label>35</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Ma</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Aafer</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Zhai</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>W</given-names>
            </name>
          </person-group>
          <article-title>Trojaning attack on neural networks</article-title>
          <year>2018</year>
          <conf-name>Network and Distributed Systems Security Symposium (NDSS)</conf-name>
          <conf-date>Feb 18-21, 2018</conf-date>
          <conf-loc>San Diego, CA</conf-loc>
          <pub-id pub-id-type="doi">10.14722/ndss.2018.23291</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref36">
        <label>36</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Turner</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Tsipras</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Madry</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Clean-label backdoor attacks</article-title>
          <source>Massachusetts Institute of Technology</source>
          <access-date>2022-07-17</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://people.csail.mit.edu/madry/lab/cleanlabel.pdf">https://people.csail.mit.edu/madry/lab/cleanlabel.pdf</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref37">
        <label>37</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Nguyen</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Tran</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Input-aware dynamic backdoor attack</article-title>
          <source>NIPS'20: Proceedings of the 34th International Conference on Neural Information Processing Systems</source>
          <year>2020</year>
          <conf-name>34th International Conference on Neural Information Processing Systems</conf-name>
          <conf-date>Dec 6-12, 2020</conf-date>
          <conf-loc>Vancouver, BC</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://dl.acm.org/doi/pdf/10.5555/3495724.3496015"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref38">
        <label>38</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Li</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>He</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Lyu</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Invisible backdoor attack with sample-specific triggers</article-title>
          <source>Proceedings of the IEEE/CVF International Conference on Computer Vision</source>
          <year>2021</year>
          <conf-name>The IEEE/CVF International Conference on Computer Vision</conf-name>
          <conf-date>Sep 22-24, 2021</conf-date>
          <conf-loc>Online</conf-loc>
          <fpage>16463</fpage>
          <lpage>16472</lpage>
          <pub-id pub-id-type="doi">10.1109/iccv48922.2021.01615</pub-id>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
