<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMI</journal-id>
      <journal-id journal-id-type="nlm-ta">JMIR Med Inform</journal-id>
      <journal-title>JMIR Medical Informatics</journal-title>
      <issn pub-type="epub">2291-9694</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v8i12e21790</article-id>
      <article-id pub-id-type="pmid">33284119</article-id>
      <article-id pub-id-type="doi">10.2196/21790</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>Comprehensive Computer-Aided Decision Support Framework to Diagnose Tuberculosis From Chest X-Ray Images: Data Mining Study</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Eysenbach</surname>
            <given-names>Gunther</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Sun</surname>
            <given-names>Wenzheng</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Kaya</surname>
            <given-names>M. Onur</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author">
          <name name-style="western">
            <surname>Owais</surname>
            <given-names>Muhammad</given-names>
          </name>
          <degrees>MSc</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-7679-081X</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author">
          <name name-style="western">
            <surname>Arsalan</surname>
            <given-names>Muhammad</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-1868-5207</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author">
          <name name-style="western">
            <surname>Mahmood</surname>
            <given-names>Tahir</given-names>
          </name>
          <degrees>MSc</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-1691-9532</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author">
          <name name-style="western">
            <surname>Kim</surname>
            <given-names>Yu Hwan</given-names>
          </name>
          <degrees>MSc</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-6769-8261</ext-link>
        </contrib>
        <contrib id="contrib5" contrib-type="author" corresp="yes">
          <name name-style="western">
            <surname>Park</surname>
            <given-names>Kang Ryoung</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <address>
            <institution>Division of Electronics and Electrical Engineering</institution>
            <institution>Dongguk University</institution>
            <addr-line>30 Pildong-ro 1-gil, Jung-gu</addr-line>
            <addr-line>Seoul, 04620</addr-line>
            <country>Republic of Korea</country>
            <phone>82 10 3111 7022</phone>
            <fax>82 2 2277 8735</fax>
            <email>parkgr@dgu.edu</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-1214-9510</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>Division of Electronics and Electrical Engineering</institution>
        <institution>Dongguk University</institution>
        <addr-line>Seoul</addr-line>
        <country>Republic of Korea</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Kang Ryoung Park <email>parkgr@dgu.edu</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <month>12</month>
        <year>2020</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>7</day>
        <month>12</month>
        <year>2020</year>
      </pub-date>
      <volume>8</volume>
      <issue>12</issue>
      <elocation-id>e21790</elocation-id>
      <history>
        <date date-type="received">
          <day>25</day>
          <month>6</month>
          <year>2020</year>
        </date>
        <date date-type="rev-request">
          <day>3</day>
          <month>11</month>
          <year>2020</year>
        </date>
        <date date-type="rev-recd">
          <day>5</day>
          <month>11</month>
          <year>2020</year>
        </date>
        <date date-type="accepted">
          <day>9</day>
          <month>11</month>
          <year>2020</year>
        </date>
      </history>
      <copyright-statement>©Muhammad Owais, Muhammad Arsalan, Tahir Mahmood, Yu Hwan Kim, Kang Ryoung Park. Originally published in JMIR Medical Informatics (http://medinform.jmir.org), 07.12.2020.</copyright-statement>
      <copyright-year>2020</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on http://medinform.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="http://medinform.jmir.org/2020/12/e21790/" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>Tuberculosis (TB) is one of the most infectious diseases that can be fatal. Its early diagnosis and treatment can significantly reduce the mortality rate. In the literature, several computer-aided diagnosis (CAD) tools have been proposed for the efficient diagnosis of TB from chest radiograph (CXR) images. However, the majority of previous studies adopted conventional handcrafted feature-based algorithms. In addition, some recent CAD tools utilized the strength of deep learning methods to further enhance diagnostic performance. Nevertheless, all these existing methods can only classify a given CXR image into binary class (either TB positive or TB negative) without providing further descriptive information.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>The main objective of this study is to propose a comprehensive CAD framework for the effective diagnosis of TB by providing visual as well as descriptive information from the previous patients’ database.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>To accomplish our objective, first we propose a fusion-based deep classification network for the CAD decision that exhibits promising performance over the various state-of-the-art methods. Furthermore, a multilevel similarity measure algorithm is devised based on multiscale information fusion to retrieve the best-matched cases from the previous database.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>The performance of the framework was evaluated based on 2 well-known CXR data sets made available by the US National Library of Medicine and the National Institutes of Health. Our classification model exhibited the best diagnostic performance (0.929, 0.937, 0.921, 0.928, and 0.965 for F1 score, average precision, average recall, accuracy, and area under the curve, respectively) and outperforms the performance of various state-of-the-art methods.</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>This paper presents a comprehensive CAD framework to diagnose TB from CXR images by retrieving the relevant cases and their clinical observations from the previous patients’ database. These retrieval results assist the radiologist in making an effective diagnostic decision related to the current medical condition of a patient. Moreover, the retrieval results can facilitate the radiologists in subjectively validating the CAD decision.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>tuberculosis</kwd>
        <kwd>computer-aided diagnosis</kwd>
        <kwd>chest radiograph</kwd>
        <kwd>lung disease</kwd>
        <kwd>neural network</kwd>
        <kwd>classification-based retrieval</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <p>According to a World Health Organization (WHO) report, tuberculosis (TB) is a major global health problem that causes severe medical conditions among millions of people annually. It ranks along with the HIV as a leading cause of mortality worldwide [<xref ref-type="bibr" rid="ref1">1</xref>]. In 2014, approximately 9.6 million new TB cases were reported as per the WHO report, which ultimately caused 1.5 million deaths [<xref ref-type="bibr" rid="ref1">1</xref>]. Today, early diagnosis and proper treatment can cure almost all the TB cases. Various types of laboratory tests have been developed to diagnose TB [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref3">3</xref>]. Among these tests, sputum smear microscopy is the most common, in which bacteria are examined from sputum samples using a microscope [<xref ref-type="bibr" rid="ref2">2</xref>]. Developed in the last few years, molecular diagnostics [<xref ref-type="bibr" rid="ref3">3</xref>] are the new techniques to diagnose TB. However, they may not be suitable in real-time screening applications. Currently, chest radiography is the most common test to detect pulmonary TB worldwide [<xref ref-type="bibr" rid="ref4">4</xref>]. It has become cheaper and easier to use with the advent of digital chest radiography [<xref ref-type="bibr" rid="ref5">5</xref>]. However, all these diagnostic tests are assessed by specialized radiologists, who must expend significant time and effort to make an accurate diagnostic decision. Therefore, such subjective methods may not be suitable for real-time screening.</p>
      <p>Over the past few years, researchers have made a significant contribution to the development of computer-aided diagnosis (CAD) tools related to chest radiography [<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref7">7</xref>]. Such automated tools can detect the various type of chest abnormalities within seconds and can aid in population screening applications, particularly in scenarios which lack medical expertise. Fortunately, the recent development in artificial intelligence has presented a remarkable breakthrough in the performance of these tools. Deep learning algorithms, specifically artificial neural networks [<xref ref-type="bibr" rid="ref8">8</xref>], are the state-of-the-art achievement in the artificial intelligence domain. These algorithms offer more reliable methods to distinguish positive and negative TB cases from chest radiographs (CXR) images in a fully automated manner. In recent decades, several ground-breaking CAD methods have been proposed for TB diagnosis [<xref ref-type="bibr" rid="ref9">9</xref>-<xref ref-type="bibr" rid="ref24">24</xref>]. Most of the previous studies used segmentation-, detection-, and classification-based approaches to make the ultimate diagnostic decisions. All these methods indicated a binary decision (either TB positive or TB negative) without providing further descriptive information that may assist medical experts to validate the CAD decision. As the CAD decision can also be erroneous in some scenarios, a method to perform its cross-validation is necessary. Therefore, further research is required to achieve the practical performance and usability of such diagnostic systems in the real world. A comprehensive analysis of these existing studies [<xref ref-type="bibr" rid="ref9">9</xref>-<xref ref-type="bibr" rid="ref24">24</xref>] in comparison with our proposed method can be found in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p>
      <p>Recently, various types of artificial neural networks have been proposed in the domain of general image processing to achieve the maximum performance in terms of accuracy (ACC) and computational cost. Among these models, convolutional neural networks (CNNs) [<xref ref-type="bibr" rid="ref25">25</xref>] attract special attention because of their outstanding performance in many general and medical image recognition applications [<xref ref-type="bibr" rid="ref26">26</xref>,<xref ref-type="bibr" rid="ref27">27</xref>]. The entire structure of a CNN model consists of an input layer, hidden layers, and a final output layer. Among all these layers, hidden layers are considered the main components of the CNN model and primarily consist of a series of convolutional layers that include trainable filters of different sizes and depths. These filters are trained by performing a training procedure to extract the deep features from a training data set. When the training procedure is completed, the trained network can analyze the given testing data and generate the desired output.</p>
      <p>In this paper, a novel CAD framework is proposed to diagnose TB from a given CXR image and provide the appropriate visual and descriptive information from a previous database, which can further assist radiologists to subjectively validate the computer decision. Thus, both subjective and CAD decisions will complement each other and ultimately result in effective diagnosis and treatment. The performance of our proposed framework was evaluated using 2 well-known CXR data sets [<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref28">28</xref>]. The overall performance of our method is substantially higher than that of various state-of-the-art methods. The main contributions of our work can be summarized as follows:</p>
      <list list-type="order">
        <list-item>
          <p>To the best of our knowledge, this is the first comprehensive CAD framework in chest radiography based on multiscale information fusion that effectively diagnoses TB by providing visual and descriptive information based on a previous patients’ database.</p>
        </list-item>
        <list-item>
          <p>We propose an ensemble classification model obtained by integrating 2 CNNs named shallow CNN (SCNN) to capture the low-level features such as edge information and a deep CNN (DCNN) to extract high-level features such as TB patterns.</p>
        </list-item>
        <list-item>
          <p>Furthermore, a multilevel similarity measure (MLSM) algorithm is proposed based on multiscale information fusion to retrieve the best-matched cases from a previous database by computing a weighted structural similarity (SSIM) score of multilevel features.</p>
        </list-item>
        <list-item>
          <p>The cross-data analysis (trained with one data set and tested with another data set, and vice versa) is a key measure to access the generalizability of a CAD tool. However, in the medical image analysis domain, most of the existing studies [<xref ref-type="bibr" rid="ref9">9</xref>-<xref ref-type="bibr" rid="ref15">15</xref>,<xref ref-type="bibr" rid="ref18">18</xref>,<xref ref-type="bibr" rid="ref19">19</xref>,<xref ref-type="bibr" rid="ref21">21</xref>-<xref ref-type="bibr" rid="ref24">24</xref>] did not analyze the performance of their methods in cross data set. Therefore, to further highlight the discriminative power of the proposed model in real-world scenarios, we also analyzed its performance in a cross data set.</p>
        </list-item>
      </list>
      <p>The remainder of the paper is structured as follows. In the “Methods” section, we describe our proposed framework. Subsequently, the experimental results along with the data set, the experimental setup, and the performance evaluation metrics are provided in the “Results” section. Finally, the “Discussion” section presents the comprehensive discussions of our paper including the principal findings.</p>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <p>This section presents a comprehensive description of our proposed framework in the following sequential order. First, we provide a brief overview of the proposed method to describe its end-to-end workflow. Subsequently, a detailed explanation of our proposed classification model and similarity measuring algorithm is presented in subsequent subsections.</p>
      <sec>
        <title>Overview of Our Proposed Framework</title>
        <p>In general, the overall performance of the image classification and retrieval framework is directly related to the mechanism of feature extraction, which is adopted to transform the visual data from high-level semantics to low-level features. These low-level features incorporate the distinctive information that can easily distinguish the instances of multiple classes. Recently, deep learning methods provide a fully automated means to extract the optimal features from available training data sets and lead to a substantial performance gain. In this study, we used the strengths of such deep learning methods to develop a comprehensive CAD tool to diagnose TB from CXR images. A comprehensive representation of the proposed framework is shown in <xref rid="figure1" ref-type="fig">Figure 1</xref>. The complete framework comprised a classification stage, a retrieval phase to perform the diagnostic decision, and retrieval of the descriptive evidence, respectively. In the first phase, our proposed ensemble-shallow–deep CNN (ensemble-SDCNN) model was trained to make the diagnostic decision for the given CXR image I by predicting its class label (CL) as either TB positive or TB negative. Such a diagnostic decision was made into 2 stages: feature extraction and classification. The detailed explanation of the proposed ensemble-SDCNN model and its workflow is provided in the subsequent subsection.</p>
        <p>In the second phase, a classification-driven retrieval was performed for the input query image. The ultimate objective of this phase was to retrieve the relevant cases (such as CXR images) corresponding to the given CXR image with the inclusion of clinical observations (such as textual description) from the previous patients’ database. Such retrieval results can assist radiologists to subjectively validate the computer diagnostic decision, which ultimately results in an effective diagnostic decision. Initially, based on the predicted CL (in the first phase), a set of positive or negative feature vectors was selected from features database based on the following predefined criteria: F = F<sup>+</sup>, if CL = TB positive; otherwise F = F<sup>–</sup>, where F<sup>+</sup>and F<sup>–</sup>present the set of positive (F<sup>+</sup> = {f<sub>1</sub><sup>+</sup>, f<sub>2</sub><sup>+</sup>, ..., f<sub>p</sub><sup>+</sup>}) and negative features maps (F<sup>–</sup> = {f<sub>1</sub><sup>–</sup>, f<sub>2</sub><sup>–</sup>, ..., f<sub>q</sub><sup>–</sup>}) in the features database, respectively, and p and q are the total numbers of positive and negative cases, respectively.</p>
        <p>Both F<sup>+</sup> and F<sup>–</sup> were extracted from TB-positive and TB-negative CXR-database (previously collected CXR images of different patients), respectively, and stored as a features database. In the subsequent step, our proposed MLSM algorithm was applied to select a subset of n best-matched features from this selected set of positive or negative features maps (ie, F={F<sup>+</sup>} or {F<sup>–</sup>}) in the first phase. Such feature matching was performed for the extracted multilevel features f′ of input query image I (as explained in a later subsection). Finally, the selected subset of n best-matched features was used to select the corresponding CXR images and their clinical readings from CXR-database and information database, respectively.</p>
        <fig id="figure1" position="float">
          <label>Figure 1</label>
          <caption>
            <p>Comprehensive flow diagram of the proposed classification and retrieval framework. In the first stage, the given input CXR image is categorized as either TB positive or TB negative. In the second stage, the n best relevant cases are retrieved from the previous database based on our proposed MLSM algorithm. The parameter n is a user given input and controls the total number of retrieved cases from the previous record related to a current medical condition. CXR: chest radiograph; DB: database; MLSM: multilevel similarity measure; SDCNN: shallow–deepCNN; TB: tuberculosis.</p>
          </caption>
          <graphic xlink:href="medinform_v8i12e21790_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Classification Network</title>
        <p>The first phase of our proposed framework involved classifying the given CXR image as either TB positive or TB negative by predicting its CL. To accomplish this task, we proposed a jointly connected ensemble-SDCNN model by performing a features-level fusion of 2 different networks, SCNN and DCNN (<xref rid="figure2" ref-type="fig">Figure 2</xref>). In general, a shallow network captures low-level features such as edge information while a deep model is used to exploit high-level information such as overall shape patterns. In our radiograph image analysis study, the experimental results prove that the combination of low- and high-level features results in better performance compared with using only high-level features. Therefore, both networks were combined in parallel (by connecting their input and last output layers with each other; <xref rid="figure2" ref-type="fig">Figure 2</xref>) to create a single end-to-end trainable network. An existing DCNN model called a residual network (ResNet18) [<xref ref-type="bibr" rid="ref29">29</xref>] was selected based on its substantial classification performance and the optimal number of parameters in comparison with the other CNN models. After selecting an optimal DCNN model, we further enhanced its performance by connecting our proposed SCNN model in parallel to it. Several experiments were performed to select the optimal number of convolutional and fully connected (FC) layers (and their hyper parameters) for the SCNN. The ultimate objective of these experiments was to construct an optimal shallow network (according to the number of parameters) that could maximize the overall classification performance of the complete network.</p>
        <p>A complete layer-wise configuration of these models is shown in <xref ref-type="table" rid="table1">Table 1</xref>. This information can assist in exploring the parametric configuration of these models more precisely. Moreover, <xref rid="figure2" ref-type="fig">Figure 2</xref> shows the overall architecture of the proposed ensemble-SDCNN model based on shallow and deep networks. Both SCNN and DCNN models processed the given CXR image in a parallel order to extract low- and high-level features, respectively. In the SCNN, the Conv1 layer (first convolutional layer with a total of 128 filters of size 7 × 7) explored the input image I in both horizontal and vertical directions and generated the output feature map, F<sub>SN1</sub> of size 73 × 73 × 128. This output feature map was further processed through the Conv2 layer (second convolutional layer with a total of 64 filters of size 5 × 5) and converted into a new features map F<sub>SN2</sub> of size 35 × 35 × 64. Thereafter, the FC1 layer (first fully connected layer including a total of 32 output nodes) identified the significant hidden patterns in F<sub>SN2</sub> by combining all the learned features into a single features vector f<sub>SN</sub> of size 1 × 1 × 32. Thus, we obtained a low-dimension features vector f<sub>SN</sub>that held a more diverse representation of the low-level features compared with F<sub>SN2</sub>.</p>
        <fig id="figure2" position="float">
          <label>Figure 2</label>
          <caption>
            <p>Overall architecture of our ensemble-SDCNN model by connecting 2 different networks, SCNN and DCNN. Both networks process the input image I simultaneously (in the testing phase) and extract 2 different feature vectors, which are concatenated and finally used to make a diagnostic decision by predicting the CL. CL: class label; CNN: convolutional neural network; DCNN: deep CNN; SCNN: shallow CNN; SDCNN: shallow–deep CNN.</p>
          </caption>
          <graphic xlink:href="medinform_v8i12e21790_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <table-wrap position="float" id="table1">
          <label>Table 1</label>
          <caption>
            <p>Layer-wise configuration details of the proposed ensemble-SDCNN<sup>a</sup> model.<sup>b</sup></p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="30"/>
            <col width="170"/>
            <col width="200"/>
            <col width="200"/>
            <col width="200"/>
            <col width="200"/>
            <col width="0"/>
            <thead>
              <tr valign="bottom">
                <td colspan="2">Layer name</td>
                <td>Output size<sup>c</sup></td>
                <td>Filter size<sup>d</sup></td>
                <td>Iterations</td>
                <td>Parameters</td>
                <td>
                  <break/>
                </td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td colspan="7">
                  <bold>DCNN<sup>e</sup> model</bold>
                </td>
              </tr>
              <tr valign="bottom">
                <td>
                  <break/>
                </td>
                <td>Input</td>
                <td>(224,224,3)</td>
                <td>N/A<sup>f</sup></td>
                <td>—<sup>g</sup></td>
                <td>—</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Conv1</td>
                <td>(112,112,64)</td>
                <td>(7,7,64)</td>
                <td>1</td>
                <td>9600</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Max pooling</td>
                <td>(56,56,64)</td>
                <td>(3,3)</td>
                <td>1</td>
                <td>—</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td> IM<sup>h</sup>-based RU1<sup>i</sup></td>
                <td>(56,56,64)</td>
                <td>(3,3,64)</td>
                <td>2</td>
                <td>74,112</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>IM-based RU2</td>
                <td>(56,56,64)</td>
                <td>(3,3,64)</td>
                <td>2</td>
                <td>74,112</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td> CM<sup>j</sup>-based RU3</td>
                <td>(28,28,128)</td>
                <td>(3,3,128); (1,1,128)</td>
                <td>2; 1</td>
                <td>230,528</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>IM-based RU4</td>
                <td>(28,28,128)</td>
                <td>(3,3,128)</td>
                <td>2</td>
                <td>295,680</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>CM-based RU5</td>
                <td>(14,14,256)</td>
                <td>(3,3,256);<break/>(1,1,256)</td>
                <td>2; 1</td>
                <td>919,808</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>IM-based RU6</td>
                <td>(14,14,256)</td>
                <td>(3,3,256)</td>
                <td>2</td>
                <td>1,181,184</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>CM-based RU7</td>
                <td>(7,7,512)</td>
                <td>(3,3,512);<break/>(1,1,512)</td>
                <td>2; 1</td>
                <td>3,674,624</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>IM-based RU8</td>
                <td>(7,7,512)</td>
                <td>(3,3,512)</td>
                <td>2</td>
                <td>4,721,664</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Avg pooling</td>
                <td>(1,1,512)</td>
                <td>(7,7)</td>
                <td>1</td>
                <td>—</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td colspan="7">
                  <bold>SCNN<sup>k</sup> model</bold>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Conv1</td>
                <td>(112,112,128)</td>
                <td>(7,7,128)</td>
                <td>1</td>
                <td>19,200</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Conv2</td>
                <td>(35,35,64)</td>
                <td>(5,5,64)</td>
                <td>1</td>
                <td>204,992</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>FC1</td>
                <td>(1,1,32)</td>
                <td>(5,5,64)</td>
                <td>1</td>
                <td>2,508,832</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Depth concat</td>
                <td>(1,1,544)</td>
                <td>—</td>
                <td>1</td>
                <td>—</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>FC2</td>
                <td>(1,1,2)</td>
                <td>—</td>
                <td>1</td>
                <td>1090</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>SoftMax</td>
                <td>(1,1,2)</td>
                <td>—</td>
                <td>1</td>
                <td>—</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Classification</td>
                <td>2</td>
                <td>—</td>
                <td>1</td>
                <td>—</td>
                <td>
                  <break/>
                </td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table1fn1">
              <p><sup>a</sup>SDCNN: shallow–deep CNN.</p>
            </fn>
            <fn id="table1fn2">
              <p><sup>b</sup>Total learnable parameters: 13,915,426.</p>
            </fn>
            <fn id="table1fn3">
              <p><sup>c</sup>Output size (image width, image height, # of channels),</p>
            </fn>
            <fn id="table1fn4">
              <p><sup>d</sup>Kernel size (kernel width, kernel height, # of filters), Max pooling (kernel width, kernel height), Avg pooling (kernel width, kernel height).</p>
            </fn>
            <fn id="table1fn5">
              <p><sup>e</sup>DCNN: deep CNN.</p>
            </fn>
            <fn id="table1fn6">
              <p><sup>f</sup>N/A: not applicable.</p>
            </fn>
            <fn id="table1fn7">
              <p><sup>g</sup>—: not available.</p>
            </fn>
            <fn id="table1fn8">
              <p><sup>h</sup>IM: identity mapping.</p>
            </fn>
            <fn id="table1fn9">
              <p><sup>i</sup>RU: residual unit.</p>
            </fn>
            <fn id="table1fn10">
              <p><sup>j</sup>CM: convolutional mapping.</p>
            </fn>
            <fn id="table1fn11">
              <p><sup>k</sup>SCNN: shallow CNN.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <p>Similarly, for the DCNN, the input image I passes through a large number of convolutional layers (as compared with the SCNN) to exploit the high-level features. Our selected DCNN model was composed of multiple residual units (RUs) that consisted of identity mapping–based or convolutional mapping–based shortcut connections to each pair of 3 × 3 filters [<xref ref-type="bibr" rid="ref29">29</xref>]. These shortcut connections caused the network to converge more efficiently compared with other sequential networks without including any shortcut connection. Moreover, a detailed explanation of these RUs is provided in [<xref ref-type="bibr" rid="ref30">30</xref>]. <xref rid="figure2" ref-type="fig">Figure 2</xref> also depicts an abstract representation of our selected DCNN model. Primarily, the input image I underwent the first convolutional layer, Conv1, with a total 64 filters of size 7 × 7. Subsequently, a Max pooling layer (with a window size 3 × 3) further down sampled the output of Conv1 and generated an intermediate features map F<sub>DN1</sub> of size 56 × 56 × 64. Thereafter, a stack of 8 consecutive RUs (including 5 identity mapping–based RUs and 3 convolutional mapping–based RUs, as shown in <xref rid="figure2" ref-type="fig">Figure 2</xref>) further exploited high-level features. Furthermore, each RU converted the preceding features map into a new one by exploiting much deeper features in comparison with the previous layer. In <xref rid="figure2" ref-type="fig">Figure 2</xref>, all the intermediate features maps (ie, F<sub>DN2</sub>, F<sub>DN3</sub>, F<sub>DN4</sub>, and F<sub>DN5</sub>) after each pair of RU show the progressive effect of different RUs. We observed that the depth of these features maps increased progressively, and the spatial size decreased after passing through the RUs. Ultimately, a low-dimension feature vector, f<sub>DN</sub>, of size 1 × 1 × 512 was obtained after processing the final features map, F<sub>DN5</sub> (obtained from the last RU), through an average pooling layer. This low-dimension feature vector exhibited a high-level abstraction of the input image I and substantially contributed, together with f<sub>SN</sub>, to the prediction of the final CL.</p>
        <p>After extracting both low- and high-level features, a depth concatenation layer (labeled as Depth concat in <xref rid="figure2" ref-type="fig">Figure 2</xref> and <xref ref-type="table" rid="table1">Table 1</xref>) performed the feature-level fusion by combining both f<sub>SN</sub> and f<sub>DN</sub> along the depth direction and generated a final features vector, f, of size 1 × 1 × 544. Finally, a stack of the FC2, SoftMax, and the classification layers (<xref rid="figure2" ref-type="fig">Figure 2</xref>) acted as a multilayer perceptron classifier and predicted the CL for the given image I using the ultimate features vector f. In this stack, the FC2 layer (including the number of nodes equal to the total number of classes) identified the larger patterns in fby combining all the features values. It multiplied f by a weight matrix W, and then added a bias vector b, where y = W·f + b, with y = [y<sub>i</sub>&#124;<sub>i=1,2</sub>]. Subsequently, the SoftMax layer converted the output of FC2 in terms of probability by applying the softmax function as y′<sub>i</sub>=e<sup>yi</sup>/Σ<sup>2</sup><sub>i=1</sub> [<xref ref-type="bibr" rid="ref8">8</xref>]. Ultimately, the classification layer obtained (y′<sub>i</sub>)from the SoftMax layer was assigned each input to one of the 2 mutually exclusive classes (ie, TB positive and TB negative) using a cross-entropy (CE) loss function as Loss<sub>CE</sub>(W,b) = Σ<sup>2</sup><sub>i=1</sub> c<sub>i</sub>ln(y′<sub>i</sub>) [<xref ref-type="bibr" rid="ref8">8</xref>]. Here, (W, b) are the network trainable parameters and c<sub>i</sub> is the indicator of the actual class label of the <italic>i</italic>th class during the training procedure. Meanwhile, in the testing phase, the network generated a single CL (as either TB positive or TB negative) corresponding to each input image I.</p>
        <p>There is also an existing SDCNN model [<xref ref-type="bibr" rid="ref31">31</xref>] (proposed for effective breast cancer diagnosis). However, there is a substantial difference between our proposed and the existing model [<xref ref-type="bibr" rid="ref31">31</xref>] in terms of architecture, application, and computational complexity. In [<xref ref-type="bibr" rid="ref31">31</xref>], the authors proposed an ensemble of 2 existing ResNet50 [<xref ref-type="bibr" rid="ref29">29</xref>] models to extract the deep features and then used a gradient boosted tree classifier to make the diagnostic decision. In addition, a 4-layer FC network, namely SCNN (which includes FC convolutional layers), was proposed for image reconstruction to increase the data samples in the preprocessing stage. By contrast, in our work, we proposed an ensemble of SCNN (which includes 2 convolutional layers [no FC] and 1 FC layer) and DCNN models as shown in <xref rid="figure2" ref-type="fig">Figure 2</xref> to extract low- and high-level features, respectively. Then, an FC classifier (also known as a multilayer perceptron) was used to make the final diagnostic decision using both low- and high-level features. Furthermore, the SCNN [<xref ref-type="bibr" rid="ref31">31</xref>] is an image reconstruction network (ie, both input and output are images), whereas our proposed SCNN is a classification network (ie, input is image, and output is feature vector). Therefore, the architecture of both SCNN models is completely different from each other. In addition, our DCNN model is based on ResNet18 that includes a substantially lower number of trainable parameters than ResNet50 as used in [<xref ref-type="bibr" rid="ref31">31</xref>], that is, 11.2M (ResNet18) &#60;&#60; 23.5M (ResNet50). In this way, the total number of trainable parameters of the proposed ensemble-SDCNN is substantially lower than the existing SDCNN [<xref ref-type="bibr" rid="ref31">31</xref>], that is, 13.9M (proposed) &#60;&#60; 47M [<xref ref-type="bibr" rid="ref31">31</xref>]. <xref rid="figure3" ref-type="fig">Figure 3</xref> further highlights the overall structural difference between our proposed and the existing model [<xref ref-type="bibr" rid="ref31">31</xref>].</p>
        <fig id="figure3" position="float">
          <label>Figure 3</label>
          <caption>
            <p>Overall structural comparison of our proposed ensemble-SDCNN (left) and existing SDCNN model (right). MLP: multilayer perceptron; GBT: gradient boosted tree.</p>
          </caption>
          <graphic xlink:href="medinform_v8i12e21790_fig3.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Multilevel Similarity Measure Algorithm</title>
        <p>In the medical domain, the visually correlated images occasionally depict different illnesses, whereas the images for a similar ailment have distinctive appearances. Therefore, estimating the similarity by contemplating the multilevel features is more advantageous in content-based medical image retrieval systems rather than using single-level features. Most of the existing systems often use a single-level similarity measure (SLSM) method to perform the content-based medical image retrieval task. However, it can miss the potentially useful information that is required in discriminating the different diseases in visually correlated images. To overcome these challenges, we proposed an MLSM algorithm to retrieve the best-matched cases from the previous patients’ database by fusing multilevel features starting from a low-level visual to a high-level semantic scale. The similarity at multiple features levels was calculated using a well-known matching algorithm called SSIM [<xref ref-type="bibr" rid="ref32">32</xref>], as it quantified the visibility of errors (differences) between 2 samples more appropriately compared with other simple matching schemes such as mean square error, peak signal-to-noise ratio (PSNR), and Euclidean distance. A generalized mathematical expression to calculate the SSIM score between 2 samples (x and y) is given as follows:</p>
        <disp-formula>SSIM(x,y) = ([2µ<sub>x</sub>µ<sub>x</sub> + c1][2σ<sub>xy</sub> + c<sub>2</sub>])/[µ<sup>2</sup><sub>x</sub> + µ<sup>2</sup><sub>y</sub> + c<sub>1</sub>] [σ<sup>2</sup><sub>x</sub> + σ<sup>2</sup><sub>y</sub> + c<sub>2</sub>]     <bold>(1)</bold></disp-formula>
        <p>where [µ<sub>x</sub>·µ<sub>y</sub>], [σ<sub>x</sub>·σ<sub>y</sub>], and σ<sub>xy</sub> are the local mean, standard deviation, and cross-covariance of the given samples, respectively; and c<sub>1</sub> and c<sub>2</sub> are constants to avoid instabilities such as infinity errors and undefined solutions.</p>
        <p>In our MLSM algorithm, multilevel features were extracted from the 8 different locations of the ensemble-SDCNN model (<xref rid="figure4" ref-type="fig">Figure 4</xref>). Each features map in <xref rid="figure4" ref-type="fig">Figure 4</xref> was obtained by calculating the depth-wise averaging of each stack of feature maps (extracted from a particular location). Moreover, this newly obtained feature map corresponding to each specific location was further presented with a pseudocolor scheme to highlight the activated regions more appropriately. In <xref rid="figure4" ref-type="fig">Figure 4</xref>, f′ presents a set of these multilevel features maps (ie, {F′<sub>SN1</sub>, F′<sub>SN2</sub>, F′<sub>DN1</sub>, F′<sub>DN2</sub>, F′<sub>DN3</sub>, F′<sub>DN4</sub>, F′<sub>DN5</sub>, f*}) corresponding to the given query image I. Similarly, f<sup>+</sup><sub>i</sub> or f<sup>–</sup><sub>i</sub> notates a set of multilevel features maps (ie, {F<sub>SN1</sub>, F<sub>SN2</sub>, F<sub>DN1</sub>, F<sub>DN2</sub>, F<sub>DN3</sub>, F<sub>DN4</sub>, F<sub>DN5</sub>, f}) for the <italic>i</italic>th positive or negative sample image in CXR-database, respectively. The selection of f<sup>+</sup><sub>i</sub> or f<sup>–</sup><sub>i</sub> was conducted based on the CL prediction, which was performed by our proposed network in the first phase. For example, in a positive prediction (ie, CL = TB positive) for the input query image I, the MLSM score between the query image I and set of p positive sample images I<sup>+</sup> (stored in CXR-database) is calculated as follows:</p>
        <disp-formula>MLSM = Σ<sup>8</sup><sub>k=1</sub>w<sub>k</sub>SSIM(f′{k},f<sup>+</sup><sub>i</sub>{k}) <sub>i=1, 2, …, p</sub>     <bold>(2)</bold></disp-formula>
        <fig id="figure4" position="float">
          <label>Figure 4</label>
          <caption>
            <p>Complete workflow diagram of our proposed MLSM algorithm using the multilevel features (extracted from the different parts of the proposed ensemble-SDCNN model) in retrieving the best-matched cases from a previous patients’ database. DCNN: deep convolutional neural network; MLSM: multilevel similarity measure; SCNN: shallow convolutional neural network; SSIM: structure similarity.</p>
          </caption>
          <graphic xlink:href="medinform_v8i12e21790_fig4.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <p>Similarly, in a negative prediction (ie, CL = TB negative), the MLSM score between the query image I and set of q negative sample images I<sup>–</sup> (also stored in CXR-database) is calculated as follows:</p>
        <disp-formula>MLSM = Σ<sup>8</sup><sub>k=1</sub>w<sub>k</sub>SSIM(f′{k},f<sup>–</sup><sub>i</sub>{k}) <sub>i=1, 2, …, q</sub>     <bold>(3)</bold></disp-formula>
        <p>In both mathematical expressions, w<sub>1</sub>, w<sub>2</sub>, w<sub>3</sub>, …, w<sub>8</sub> are the weights of SSIM measured at different levels and their total sum is equal to one (ie, Σ<sup>8</sup><sub>i=1</sub> w<sub>i</sub>=1). The optimal weights were obtained by maximizing the intraclass SSIM score for some selected pairs of positive CXR images. Each pair (I<sup>+</sup><sub>i</sub>, I<sup>+</sup><sub>j</sub>) was selected from the positive data samples based on the highly correlated clinical observations between 2 CXR images. These observations were provided in our selected data sets as a text file for each data sample. As our main objective was to diagnose TB by retrieving similar abnormal cases from a previous database, we only considered positive CXR images in calculating the optimal weights rather than using normal images.</p>
        <p>Finally, the overall objective function to maximize the intraclass similarity is defined as follows:</p>
        <disp-formula>w*=max(Σ<sub>i,j</sub>εTBpositiveΣ<sup>8</sup><sub>k=1</sub>w<sub>k</sub> SSIM[f<sup>+</sup><sub>i</sub>{k},f<sup>+</sup><sub>j</sub>{k}]) /N<sup>+</sup>     <bold>(4)</bold></disp-formula>
        <p>where N<sup>+</sup>is the total number of pair images selected from the positive data samples. In our experiment, the total number of pairs was 30 (ie, N<sup>+</sup> = 30). After performing the optimization according to Equation (4), we obtained the optimal values of w<sub>1</sub>, w<sub>2</sub>, w<sub>3</sub>, …, w<sub>8</sub> as 0.069, 0.179, 0.087, 0.133, 0.071, 0.123, 0.299, and 0.039, respectively. Finally, these optimal weights were used to calculate the MLSM scores between f′ and F<sup>+</sup> (set of positive features maps in features database) or F<sup>–</sup> (set of negative features maps in features database) depending on the predicted CL in the classification stage. Thereafter, the indices of n best-matched features were selected based on the maximum MLSM scores. These indices were eventually used to select the corresponding CXR images and their clinical readings from CXR-database and information database, respectively. Thus, n best-matched cases were retrieved from the previous patients’ database, which could assist radiologists in making an effective diagnostic decision after performing the subjective validation of the computer decision.</p>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>Data Set and Preprocessing</title>
        <p>Our proposed diagnostic framework was validated using 2 publicly available data sets: Montgomery County (MC) and Shenzhen (SZ) [<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref28">28</xref>]. The MC data set is a subset of a larger CXR repository collected within the TB control program of the Department of Health and Human Services of Montgomery County, Maryland, USA. All these images are in 12-bit grayscale, captured using a Eureka stationary X-ray machine. This data set comprises a total of 138 posteroanterior CXR images, among which there are 80 normal and 58 abnormal images with the manifestations of TB disease. The abnormal images encompass a vast range of abnormalities related to pulmonary TB. The SZ data set is collected from the Shenzhen No. 3 People’s Hospital in Shenzhen, Guangdong Providence, China. This data set includes a total of 326 normal and 336 abnormal CXR images, which include different types of abnormalities related to pulmonary TB. All these images are also in 12-bit grayscale and were captured using the Philips DR DigitalDiagnost system. In both data sets, a radiologist report is also provided for each CXR image as a text file, containing the clinical observation related to chest abnormalities along with the patient’s age and gender information. After collecting both data sets, we resized all the images to a spatial dimension of 224 × 224 (according to the fixed input layer size of our ensemble-SDCNN model).</p>
      </sec>
      <sec>
        <title>Implementation Details</title>
        <p>The proposed framework was implemented using a standard deep learning toolbox available in the MATLAB R2019a (MathWorks, Inc.) framework [<xref ref-type="bibr" rid="ref33">33</xref>]. It provides a complete framework for developing and testing different types of artificial neural networks and using existing pretrained networks. All the experiments were performed on a desktop computer with a 3.50-GHz Intel Core i7-3770K CPU [<xref ref-type="bibr" rid="ref34">34</xref>], 16-GB RAM, an NVIDIA GeForce GTX 1070 graphics card [<xref ref-type="bibr" rid="ref35">35</xref>], and Windows 10 operating system (Microsoft). Our proposed and other baseline models were trained through back-propagation (a procedure to determine the optimal parameters of a model) using a well-known optimization algorithm called the stochastic gradient descent [<xref ref-type="bibr" rid="ref36">36</xref>]. It iteratively trains the network by computing the optimal learnable parameters (such as filter weights and biases) that are included in different layers of the network. The following hyper-parameters were selected for our proposed and all the comparative CNN-based methods: learning rate as 0.001 with a drop factor of 0.1. Moreover, the min-batch size was selected as 10 (ie, feeding a stack of 10 images per gradient update in each iteration), L2-regularization as 0.0001, and a momentum factor as 0.9.</p>
      </sec>
      <sec>
        <title>Evaluation Metrics and Protocol</title>
        <p>After the training, the quantitative performance of our proposed framework was evaluated based on the following metrics: ACC, average precision (AP), average recall (AR), F1 score (F1), and finally the area under the curve (AUC) [<xref ref-type="bibr" rid="ref37">37</xref>]. These well-known metrics can quantify the overall performance of a deep learning model from many perspectives. The mathematical definition of all these metrics is provided in <xref ref-type="table" rid="table2">Table 2</xref>.</p>
        <p>In our binary classification problem, true positive (TP) and true negative (TN) were the outcomes of our model for correctly predicted positive and negative cases, respectively, whereas false positive (FP) and false negative (FN) could be interpreted as the incorrectly predicted positive and negative cases, respectively. Finally, these 4 different outcomes were further used in assessing the overall performance of a model in terms of ACC, AP, AR, F1, and AUC. We performed a fivefold cross-validation in all the experiments by randomly selecting 80% of data (110/138 [79.7%] of MC data and 530/662 [80.0%] SZ data) for training and the remaining 20% (28/138 [20.2%] of MC data and 132/662 [19.9%] SZ data) for testing. As most of the previous studies considered fivefold cross-validations, we followed a similar data splitting protocol. However, the fivefold cross-validation was not possible for the evaluation of the cross–data set performance, as a complete data set was used for training and others for testing. However, we performed cross-data validation using the MC data set as training and the SZ data set as testing, and vice versa.</p>
        <table-wrap position="float" id="table2">
          <label>Table 2</label>
          <caption>
            <p>Mathematical definition of our selected performance evaluation metrics.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="404"/>
            <col width="596"/>
            <thead>
              <tr valign="top">
                <td>Metric name</td>
                <td>Mathematical equation</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Accuracy (ACC)</td>
                <td>(TP<sup>a</sup> + TN<sup>b</sup>)/(TP + TN + FP<sup>c</sup> + FN<sup>d</sup>)</td>
              </tr>
              <tr valign="top">
                <td>Average precision (AP)</td>
                <td>TP/(TP + FP)</td>
              </tr>
              <tr valign="top">
                <td>Average recall (AR)</td>
                <td>TP/(TP + FN)</td>
              </tr>
              <tr valign="top">
                <td>F1 score (F1)</td>
                <td>2×([AP × AR]/[AP + AR])</td>
              </tr>
              <tr valign="top">
                <td>Area under the curve (AUC)</td>
                <td>0.5 × (TP/[TP + FP] + TN/[TN + FP])</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table2fn1">
              <p><sup>a</sup>TP: true positive.</p>
            </fn>
            <fn id="table2fn2">
              <p><sup>b</sup>TN: true negative.</p>
            </fn>
            <fn id="table2fn3">
              <p><sup>c</sup>FP: false positive.</p>
            </fn>
            <fn id="table2fn4">
              <p><sup>d</sup>FN: false negative.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Our Results and an Ablation Study</title>
        <p>The overall performance of our diagnostic framework was directly related to the classification performance of the proposed ensemble-SDCNN model. As in our classification-driven retrieval framework, the first step was to predict the CL for the given query image and then explore that predicted class database to retrieve the relevant cases. Consequently, the correct prediction would ultimately result in correct retrieval and the incorrect prediction in incorrect retrieval. Therefore, we comprehensively assessed the classification performance of the proposed model for both data sets and their combinations. <xref ref-type="table" rid="table3">Table 3</xref> shows the performance of our classification model along with an ablation study to highlight the significance of each submodel in enhancing the overall performance. Therefore, the individual performance of both SCNN and DCNN models was also computed as an ablation study. The experimental results indicated that the combination of SCNN and DCNN resulted in a substantial performance gain (ie, 8.8%, 8.12%, 9.42%, 8.76%, and 5.68% for the average F1, AP, AR, ACC, and AUC, respectively) compared with their individual performances. We further performed a <italic>t</italic> test [<xref ref-type="bibr" rid="ref38">38</xref>] and Cohen <italic>d</italic> [<xref ref-type="bibr" rid="ref39">39</xref>] analysis to signify the performance gain of our SDCNN model in contrast to the DCNN (second-best model). In these 2 performance analysis measures, a large number of experimental results appropriately discriminated the performances of 2 systems.</p>
        <p>Therefore, the detailed performance results of both ensemble-SDCNN and DCNN for all the different folds were used to perform the <italic>t</italic> test and Cohen <italic>d</italic> analysis. In the <italic>t</italic> test analysis, all the <italic>P</italic>-values (ie, .012, .011, .015, .014, and .012 in the case of average F1, AP, AR, ACC, and AUC, respectively) were less than .05. These results implied the discriminative performance of our ensemble-SDCNN against the SCNN with a 95% confidence score. In the Cohen <italic>d</italic> analysis, the performance difference between 2 systems was measured in terms of effect size [<xref ref-type="bibr" rid="ref40">40</xref>], which is generally categorized as small (approximately 0.2-0.3), medium (approximately 0.5), and large (≥0.8). The large effect size indicated a substantial performance difference between the 2 systems. In this analysis, all the effect sizes (ie, 0.6, 0.6, 0.6, 0.5, and 0.5 for the average F1, AP, AR, ACC, and AUC, respectively) were greater than and equal to 0.5, which also indicated the substantial performance difference between the ensemble-SDCNN and SCNN models.</p>
        <table-wrap position="float" id="table3">
          <label>Table 3</label>
          <caption>
            <p>Classification performance of our proposed ensemble-SDCNN<sup>a</sup> model including the submodels as an ablation study.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="30"/>
            <col width="370"/>
            <col width="120"/>
            <col width="120"/>
            <col width="120"/>
            <col width="120"/>
            <col width="120"/>
            <thead>
              <tr valign="top">
                <td colspan="2">Data sets and models</td>
                <td>F1</td>
                <td>AP<sup>b</sup></td>
                <td>AR<sup>c</sup></td>
                <td>ACC<sup>d</sup></td>
                <td>AUC<sup>e</sup></td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td colspan="2">
                  <bold>MC<sup>f</sup></bold>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td rowspan="3">
                  <break/>
                </td>
                <td> SCNN<sup>g,h</sup></td>
                <td>0.765</td>
                <td>0.775</td>
                <td>0.757</td>
                <td>0.769</td>
                <td>0.817</td>
              </tr>
              <tr valign="top">
                <td> DCNN<sup>i,j</sup></td>
                <td>0.88</td>
                <td>0.888</td>
                <td>0.872</td>
                <td>0.878</td>
                <td>0.932</td>
              </tr>
              <tr valign="top">
                <td>ensemble-SDCNN</td>
                <td>0.929</td>
                <td>0.937</td>
                <td>0.921</td>
                <td>0.928</td>
                <td>0.965</td>
              </tr>
              <tr valign="top">
                <td colspan="2">
                  <bold>SZ<sup>k</sup></bold>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td rowspan="3">
                  <break/>
                </td>
                <td>SCNN</td>
                <td>0.802</td>
                <td>0.803</td>
                <td>0.802</td>
                <td>0.802</td>
                <td>0.868</td>
              </tr>
              <tr valign="top">
                <td>DCNN</td>
                <td>0.892</td>
                <td>0.892</td>
                <td>0.892</td>
                <td>0.891</td>
                <td>0.939</td>
              </tr>
              <tr valign="top">
                <td>ensemble-SDCNN</td>
                <td>0.908</td>
                <td>0.909</td>
                <td>0.908</td>
                <td>0.908</td>
                <td>0.948</td>
              </tr>
              <tr valign="top">
                <td colspan="2">
                  <bold>MC + SZ</bold>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td rowspan="3">
                  <break/>
                </td>
                <td>SCNN</td>
                <td>0.79</td>
                <td>0.793</td>
                <td>0.788</td>
                <td>0.789</td>
                <td>0.841</td>
              </tr>
              <tr valign="top">
                <td>DCNN</td>
                <td>0.891</td>
                <td>0.892</td>
                <td>0.89</td>
                <td>0.89</td>
                <td>0.943</td>
              </tr>
              <tr valign="top">
                <td>ensemble-SDCNN</td>
                <td>0.9</td>
                <td>0.902</td>
                <td>0.898</td>
                <td>0.899</td>
                <td>0.95</td>
              </tr>
              <tr valign="top">
                <td colspan="2">
                  <bold>MC train and SZ test</bold>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td rowspan="3">
                  <break/>
                </td>
                <td>SCNN</td>
                <td>0.557</td>
                <td>0.559</td>
                <td>0.555</td>
                <td>0.557</td>
                <td>0.541</td>
              </tr>
              <tr valign="top">
                <td>DCNN</td>
                <td>0.54</td>
                <td>0.574</td>
                <td>0.51</td>
                <td>0.517</td>
                <td>0.737</td>
              </tr>
              <tr valign="top">
                <td>ensemble-SDCNN</td>
                <td>0.795</td>
                <td>0.798</td>
                <td>0.793</td>
                <td>0.792</td>
                <td>0.853</td>
              </tr>
              <tr valign="top">
                <td colspan="2">
                  <bold>SZ train and MC test</bold>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td rowspan="3">
                  <break/>
                </td>
                <td>SCNN</td>
                <td>0.625</td>
                <td>0.624</td>
                <td>0.626</td>
                <td>0.616</td>
                <td>0.601</td>
              </tr>
              <tr valign="top">
                <td>DCNN</td>
                <td>0.7</td>
                <td>0.702</td>
                <td>0.698</td>
                <td>0.71</td>
                <td>0.754</td>
              </tr>
              <tr valign="top">
                <td>ensemble-SDCNN</td>
                <td>0.811</td>
                <td>0.808</td>
                <td>0.813</td>
                <td>0.797</td>
                <td>0.873</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table3fn1">
              <p><sup>a</sup>SDCNN: shallow–deep CNN.</p>
            </fn>
            <fn id="table3fn2">
              <p><sup>b</sup>AP: average precision.</p>
            </fn>
            <fn id="table3fn3">
              <p><sup>c</sup>AR: average recall.</p>
            </fn>
            <fn id="table3fn4">
              <p><sup>d</sup>ACC: accuracy.</p>
            </fn>
            <fn id="table3fn5">
              <p><sup>e</sup>AUC: area under the curve.</p>
            </fn>
            <fn id="table3fn6">
              <p><sup>f</sup>MC: Montgomery County.</p>
            </fn>
            <fn id="table3fn7">
              <p><sup>g</sup>Ablation study performance by only considering SCNN for classification.</p>
            </fn>
            <fn id="table3fn8">
              <p><sup>h</sup>SCNN: shallow CNN.</p>
            </fn>
            <fn id="table3fn9">
              <p><sup>i</sup>Ablation study performance by only considering DCNN for classification.</p>
            </fn>
            <fn id="table3fn10">
              <p><sup>j</sup>DCNN: deep CNN.</p>
            </fn>
            <fn id="table3fn11">
              <p><sup>k</sup>SZ: Shenzhen.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <p><xref rid="figure5" ref-type="fig">Figure 5</xref> depicts the receiver operating characteristic curves of the proposed model for all the data sets. Each curve plots the TPR versus the FPR of our model at different classification thresholds beginning from 0 to 1 at 0.001 increments. Among all the classification thresholds, the optimal threshold was obtained based on the operating points (as highlighted with red closed circles) existing over the operating line. We attained the optimal threshold value of 0.507 for all the data sets. This implied that any CXR image with a classification probability larger than .507 was reported as a positive case. Finally, based on these receiver operating characteristic curves, we calculated the AUC results of our model for each data set (<xref ref-type="table" rid="table3">Table 3</xref>). We observed that the MC, SZ, and MC + SZ data sets had comparable AUCs of 0.965, 0.948, and 0.95, respectively. However, the performance of the cross–data set AUC was lower than that of the MC and SZ because of high intraclass and interclass variances between 2 different data sets, but the comparative performance (as reported in the subsequent section) of our model was still greater than the existing state-of-the-art methods for all the data sets.</p>
        <fig id="figure5" position="float">
          <label>Figure 5</label>
          <caption>
            <p>Receiver operating characteristic curves of our ensemble-SDCNN model for all the datasets. Each curve plots true-positive rate (TPR) vs false-positive rate (FPR) of our model at different classification thresholds beginning from 0 to 1 in 0.001 increments. MC: Montgomery County; SDCNN: shallow–deep convolutional neural network; SZ: Shenzhen.</p>
          </caption>
          <graphic xlink:href="medinform_v8i12e21790_fig5.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <p>To determine the optimal ratio of the SCNN features with the DCNN, we performed several experiments for all the data sets by considering the different feature lengths of f<sub>SN</sub> concatenated with f<sub>DN</sub>. In this analysis, the feature lengths began from 0 to 512 with the increment of 8 features per experiment. <xref rid="figure6" ref-type="fig">Figure 6</xref> shows the F1 and AUC results (average performance of all the data sets) according to different features length of f<sub>SN</sub>. In addition, the black line depicts the growing number of the total parameters of our proposed model with the increasing length of f<sub>SN</sub>. The figure indicates that our model exhibited the best performance (ie, maximum F1 of 0.871 and AUC of 0.918 as indicated by the vertical red line) and required the optimal number of total parameters as 1.39 × 10<sup>7</sup> for f<sub>SN</sub>=32. Although the total number of trainable parameters of our model was slightly higher (approximately 2.7 million) than that of the DCNN, a substantial performance difference was observed, particularly for the cross data set (<xref ref-type="table" rid="table3">Table 3</xref>).</p>
        <fig id="figure6" position="float">
          <label>Figure 6</label>
          <caption>
            <p>Average performance of the proposed ensemble-SDCNN model by considering different lengths of SCNN features with DCNN features (beginning from 0 to 512 with the increment of eight features in each experiment). AUC: area under the curve; DCNN: deep convolutional neural network; SDCNN: shallow–deep convolutional neural network; SCNN: shallow convolutional neural network.</p>
          </caption>
          <graphic xlink:href="medinform_v8i12e21790_fig6.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <p>In our classification-driven framework, both classification and retrieval performances were similar. However, we also evaluated the retrieval performance without performing the class prediction to validate the superiority of our classification-driven approach. In <xref ref-type="table" rid="table4">Table 4</xref>, the experimental results indicate that our classification-driven approach exhibited higher retrieval accuracies than the retrieval without class prediction. Moreover, our retrieval approach was computationally more efficient than that without class prediction as feature matching was performed using only the predicted class database rather than the entire database as in the retrieval without class prediction. In conclusion, these comparative results (<xref ref-type="table" rid="table3">Tables 3</xref> and <xref ref-type="table" rid="table4">4</xref>) implied that our jointly connected model exhibited superior performance in making the effective diagnostic decision and retrieving the best-matched cases from the previous database.</p>
        <table-wrap position="float" id="table4">
          <label>Table 4</label>
          <caption>
            <p>Comparative retrieval performance with and without predicting the class label (CL).</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="30"/>
            <col width="370"/>
            <col width="150"/>
            <col width="150"/>
            <col width="150"/>
            <col width="150"/>
            <thead>
              <tr valign="top">
                <td colspan="2">Retrieval and data sets</td>
                <td>F1</td>
                <td>AP<sup>a</sup></td>
                <td>AR<sup>b</sup></td>
                <td>ACC<sup>c</sup></td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td colspan="2">
                  <bold>Without class prediction</bold>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td rowspan="5">
                  <break/>
                </td>
                <td>MC<sup>d</sup></td>
                <td>0.844</td>
                <td>0.861</td>
                <td>0.828</td>
                <td>0.847</td>
              </tr>
              <tr valign="top">
                <td>SZ<sup>e</sup></td>
                <td>0.891</td>
                <td>0.892</td>
                <td>0.89</td>
                <td>0.89</td>
              </tr>
              <tr valign="top">
                <td>MC + SZ</td>
                <td>0.88</td>
                <td>0.882</td>
                <td>0.878</td>
                <td>0.879</td>
              </tr>
              <tr valign="top">
                <td>MC train and SZ test</td>
                <td>0.534</td>
                <td>0.538</td>
                <td>0.53</td>
                <td>0.533</td>
              </tr>
              <tr valign="top">
                <td>SZ train and MC test</td>
                <td>0.729</td>
                <td>0.737</td>
                <td>0.72</td>
                <td>0.739</td>
              </tr>
              <tr valign="top">
                <td colspan="2">
                  <bold>With class prediction</bold>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td rowspan="5">
                  <break/>
                </td>
                <td>MC</td>
                <td>0.929</td>
                <td>0.937</td>
                <td>0.921</td>
                <td>0.928</td>
              </tr>
              <tr valign="top">
                <td>SZ</td>
                <td>0.908</td>
                <td>0.909</td>
                <td>0.908</td>
                <td>0.908</td>
              </tr>
              <tr valign="top">
                <td>MC + SZ</td>
                <td>0.9</td>
                <td>0.902</td>
                <td>0.898</td>
                <td>0.899</td>
              </tr>
              <tr valign="top">
                <td>MC train and SZ test</td>
                <td>0.795</td>
                <td>0.798</td>
                <td>0.793</td>
                <td>0.792</td>
              </tr>
              <tr valign="top">
                <td>SZ train and MC test</td>
                <td>0.811</td>
                <td>0.808</td>
                <td>0.813</td>
                <td>0.797</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table4fn1">
              <p><sup>a</sup>AP: average precision.</p>
            </fn>
            <fn id="table4fn2">
              <p><sup>b</sup>AR: average recall.</p>
            </fn>
            <fn id="table4fn3">
              <p><sup>c</sup>ACC: accuracy.</p>
            </fn>
            <fn id="table4fn4">
              <p><sup>d</sup>MC: Montgomery County.</p>
            </fn>
            <fn id="table4fn5">
              <p><sup>e</sup>SZ: Shenzhen.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Comparative Analysis</title>
        <p>Several CAD methods are presented in the literature for diagnosing pulmonary TB in CXR images. To make a fair comparison, we considered the following state-of-the-art methods [<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref15">15</xref>,<xref ref-type="bibr" rid="ref17">17</xref>,<xref ref-type="bibr" rid="ref21">21</xref>,<xref ref-type="bibr" rid="ref22">22</xref>,<xref ref-type="bibr" rid="ref41">41</xref>,<xref ref-type="bibr" rid="ref42">42</xref>], because these approaches selected the same data sets and experimental protocols as considered in our study. Moreover, in some recent studies [<xref ref-type="bibr" rid="ref21">21</xref>], the authors adopted existing CNN models to classify the different types of pulmonary abnormalities including TB. However, these studies considered different data sets and experimental protocols. For a fair and detailed comparison, we evaluated the performance of these methods for our selected data sets and experimental protocol. Additionally, we calculated the performance of other CNN models [<xref ref-type="bibr" rid="ref29">29</xref>,<xref ref-type="bibr" rid="ref43">43</xref>-<xref ref-type="bibr" rid="ref45">45</xref>] proposed for the general image-classification domain rather than radiology. The objective of this comparative analysis was to estimate the performance of the existing state-of-the-art CNN models in CXR image analyses. All these comparative analysis results are shown in <xref ref-type="table" rid="table5">Table 5</xref>.</p>
        <table-wrap position="float" id="table5">
          <label>Table 5</label>
          <caption>
            <p>Comparative performance analysis of the proposed ensemble-SDCNN<sup>a</sup> model with various state-of-the-art methods.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="250"/>
            <col width="50"/>
            <col width="50"/>
            <col width="50"/>
            <col width="50"/>
            <col width="50"/>
            <col width="0"/>
            <col width="50"/>
            <col width="50"/>
            <col width="50"/>
            <col width="50"/>
            <col width="50"/>
            <col width="0"/>
            <col width="50"/>
            <col width="50"/>
            <col width="50"/>
            <col width="50"/>
            <col width="50"/>
            <thead>
              <tr valign="top">
                <td rowspan="2">Comparative methods</td>
                <td colspan="5">MC<sup>b</sup></td>
                <td colspan="7">SZ<sup>c</sup></td>
                <td colspan="5">MC + SZ</td>
              </tr>
              <tr valign="top">
                <td>F1</td>
                <td>AP<sup>d</sup></td>
                <td>AR<sup>e</sup></td>
                <td>ACC<sup>f</sup></td>
                <td colspan="2">AUC<sup>g</sup></td>
                <td>F1</td>
                <td>AP</td>
                <td>AR</td>
                <td>ACC</td>
                <td>AUC</td>
                <td colspan="2">F1</td>
                <td>AP</td>
                <td>AR</td>
                <td>ACC</td>
                <td>AUC</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>LBP<sup>h</sup> and SVM<sup>i,j</sup> [<xref ref-type="bibr" rid="ref46">46</xref>]</td>
                <td>0.537</td>
                <td>0.58</td>
                <td>0.5</td>
                <td>0.58</td>
                <td colspan="2">0.675</td>
                <td>0.76</td>
                <td>0.76</td>
                <td>0.76</td>
                <td>0.76</td>
                <td>0.83</td>
                <td colspan="2">0.729</td>
                <td>0.729</td>
                <td>0.729</td>
                <td>0.729</td>
                <td>0.763</td>
              </tr>
              <tr valign="top">
                <td>HoG<sup>k</sup> and SVM<sup>i</sup> [<xref ref-type="bibr" rid="ref47">47</xref>]</td>
                <td>0.797</td>
                <td>0.796</td>
                <td>0.798</td>
                <td>0.797</td>
                <td colspan="2">0.863</td>
                <td>0.85</td>
                <td>0.85</td>
                <td>0.85</td>
                <td>0.85</td>
                <td>0.90</td>
                <td colspan="2">0.822</td>
                <td>0.823</td>
                <td>0.821</td>
                <td>0.821</td>
                <td>0.882</td>
              </tr>
              <tr valign="top">
                <td>ShuffleNet<sup>i</sup> [<xref ref-type="bibr" rid="ref43">43</xref>]</td>
                <td>0.747</td>
                <td>0.771</td>
                <td>0.727</td>
                <td>0.748</td>
                <td colspan="2">0.84</td>
                <td>0.875</td>
                <td>0.876</td>
                <td>0.873</td>
                <td>0.873</td>
                <td>0.937</td>
                <td colspan="2">0.884</td>
                <td>0.885</td>
                <td>0.883</td>
                <td>0.884</td>
                <td>0.936</td>
              </tr>
              <tr valign="top">
                <td>InceptionV3<sup>i</sup> [<xref ref-type="bibr" rid="ref44">44</xref>]</td>
                <td>0.739</td>
                <td>0.773</td>
                <td>0.711</td>
                <td>0.74</td>
                <td colspan="2">0.828</td>
                <td>0.882</td>
                <td>0.883</td>
                <td>0.881</td>
                <td>0.881</td>
                <td>0.942</td>
                <td colspan="2">0.887</td>
                <td>0.89</td>
                <td>0.884</td>
                <td>0.885</td>
                <td>0.944</td>
              </tr>
              <tr valign="top">
                <td>MobileNetV2<sup>i</sup> [<xref ref-type="bibr" rid="ref45">45</xref>]</td>
                <td>0.762</td>
                <td>0.769</td>
                <td>0.755</td>
                <td>0.769</td>
                <td colspan="2">0.833</td>
                <td>0.876</td>
                <td>0.878</td>
                <td>0.875</td>
                <td>0.875</td>
                <td>0.941</td>
                <td colspan="2">0.886</td>
                <td>0.888</td>
                <td>0.883</td>
                <td>0.884</td>
                <td>0.946</td>
              </tr>
              <tr valign="top">
                <td>Santosh et al [<xref ref-type="bibr" rid="ref41">41</xref>]</td>
                <td>—<sup>l</sup></td>
                <td>—</td>
                <td>—</td>
                <td>0.79</td>
                <td colspan="2">0.88</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>0.86</td>
                <td>0.93</td>
                <td colspan="2">—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>Hwang et al [<xref ref-type="bibr" rid="ref17">17</xref>]</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>0.674</td>
                <td colspan="2">0.884</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>0.837</td>
                <td>0.926</td>
                <td colspan="2">—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>ResNet50<sup>i</sup> [<xref ref-type="bibr" rid="ref29">29</xref>]</td>
                <td>0.788</td>
                <td>0.796</td>
                <td>0.78</td>
                <td>0.79</td>
                <td colspan="2">0.886</td>
                <td>0.877</td>
                <td>0.877</td>
                <td>0.877</td>
                <td>0.876</td>
                <td>0.94</td>
                <td colspan="2">0.88</td>
                <td>0.881</td>
                <td>0.878</td>
                <td>0.879</td>
                <td>0.921</td>
              </tr>
              <tr valign="top">
                <td>ResNet101<sup>i</sup> [<xref ref-type="bibr" rid="ref29">29</xref>]</td>
                <td>0.8</td>
                <td>0.821</td>
                <td>0.782</td>
                <td>0.798</td>
                <td colspan="2">0.895</td>
                <td>0.864</td>
                <td>0.865</td>
                <td>0.862</td>
                <td>0.861</td>
                <td>0.934</td>
                <td colspan="2">0.859</td>
                <td>0.862</td>
                <td>0.857</td>
                <td>0.858</td>
                <td>0.923</td>
              </tr>
              <tr valign="top">
                <td>Alfadhli et al [<xref ref-type="bibr" rid="ref14">14</xref>]</td>
                <td>—</td>
                <td>0.81</td>
                <td>0.79</td>
                <td>0.791</td>
                <td colspan="2">0.89</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>GoogLeNet<sup>i</sup> [<xref ref-type="bibr" rid="ref20">20</xref>,<xref ref-type="bibr" rid="ref21">21</xref>]</td>
                <td>0.834</td>
                <td>0.851</td>
                <td>0.818</td>
                <td>0.834</td>
                <td colspan="2">0.902</td>
                <td>0.852</td>
                <td>0.853</td>
                <td>0.851</td>
                <td>0.851</td>
                <td>0.921</td>
                <td colspan="2">0.843</td>
                <td>0.846</td>
                <td>0.84</td>
                <td>0.84</td>
                <td>0.914</td>
              </tr>
              <tr valign="top">
                <td>Lopes and Valiati [<xref ref-type="bibr" rid="ref21">21</xref>]</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>0.826</td>
                <td colspan="2">0.926</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>0.847</td>
                <td>0.904</td>
                <td colspan="2">—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>Vajda et al [<xref ref-type="bibr" rid="ref42">42</xref>]</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>0.783</td>
                <td colspan="2">0.87</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>Pasa et al [<xref ref-type="bibr" rid="ref22">22</xref>]</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>0.79</td>
                <td colspan="2">0.811</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>0.844</td>
                <td>0.9</td>
                <td colspan="2">—</td>
                <td>—</td>
                <td>—</td>
                <td>0.862</td>
                <td>0.925</td>
              </tr>
              <tr valign="top">
                <td>Govindarajan and Swaminathan [<xref ref-type="bibr" rid="ref15">15</xref>]</td>
                <td>0.876</td>
                <td>—</td>
                <td>0.877</td>
                <td>0.878</td>
                <td colspan="2">0.94</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>Proposed</td>
                <td>0.929</td>
                <td>0.937</td>
                <td>0.921</td>
                <td>0.928</td>
                <td colspan="2">0.965</td>
                <td>0.908</td>
                <td>0.909</td>
                <td>0.908</td>
                <td>0.908</td>
                <td>0.948</td>
                <td colspan="2">0.9</td>
                <td>0.902</td>
                <td>0.898</td>
                <td>0.899</td>
                <td>0.95</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table5fn1">
              <p><sup>a</sup>SDCNN: shallow–deep CNN.</p>
            </fn>
            <fn id="table5fn2">
              <p><sup>b</sup>MC: Montgomery County.</p>
            </fn>
            <fn id="table5fn3">
              <p><sup>c</sup>SZ: Shenzhen.</p>
            </fn>
            <fn id="table5fn4">
              <p><sup>d</sup>AP: average precision.</p>
            </fn>
            <fn id="table5fn5">
              <p><sup>e</sup>AR: average recall.</p>
            </fn>
            <fn id="table5fn6">
              <p><sup>f</sup>ACC: accuracy.</p>
            </fn>
            <fn id="table5fn7">
              <p><sup>g</sup>AUC: area under the curve.</p>
            </fn>
            <fn id="table5fn8">
              <p><sup>h</sup>LBP: local binary pattern.</p>
            </fn>
            <fn id="table5fn9">
              <p><sup>i</sup>We evaluated the performance of these models using our selected data sets and experimental protocol.</p>
            </fn>
            <fn id="table5fn10">
              <p><sup>j</sup>SVM: support vector machine.</p>
            </fn>
            <fn id="table5fn11">
              <p><sup>k</sup>HoG: histogram of oriented gradients.</p>
            </fn>
            <fn id="table5fn12">
              <p><sup>l</sup>—: not available. These results were not reported in some existing studies.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <p>We observed that our method exhibited a superior performance (in terms of all the performance measures and data sets) compared with all the other baseline methods. In addition to deep learning–based methods, we evaluated and compared the performance of 2 known handcrafted feature-based methods [<xref ref-type="bibr" rid="ref46">46</xref>,<xref ref-type="bibr" rid="ref47">47</xref>]. To evaluate the performance of these 2 methods [<xref ref-type="bibr" rid="ref46">46</xref>,<xref ref-type="bibr" rid="ref47">47</xref>], we used the following default parameters as provided by the MATLAB framework [<xref ref-type="bibr" rid="ref33">33</xref>]: size of histogram of oriented gradients cell as 8 × 8 with block size of 2 × 2 and number of overlapping cells between adjacent blocks as 1 block and the number of orientation bins as 9. In local binary patterns (LBPs) [<xref ref-type="bibr" rid="ref46">46</xref>], the number of neighbor pixels considered was 8, with the linear interpolation method applied to compute pixel neighbors. Whereas in LBP histogram parameters, cell size was selected as 1 × 1 by applying L2-normalization to each LBP cell histogram. Thus, our comparative analysis was more detailed than the various existing studies [<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref17">17</xref>,<xref ref-type="bibr" rid="ref21">21</xref>,<xref ref-type="bibr" rid="ref22">22</xref>]. For the MC data set, the performance gain of our model in contrast to Govindarajan and Swaminathan [<xref ref-type="bibr" rid="ref15">15</xref>] (second-best) was greater than 4.4%, 5%, and 2.5% for AR, ACC, and AUC, respectively. Similarly, the difference in the performance of our model from a second-best model called InceptionV3 [<xref ref-type="bibr" rid="ref44">44</xref>] (for the SZ data set) was more than 2.6%, 2.6%, 2.7%, 2.7%, and 0.6% for F1, AP, AR, ACC, and AUC, respectively. Moreover, for the combined data set (MC + SZ), the performance gain of our model in contrast to InceptionV3 [<xref ref-type="bibr" rid="ref44">44</xref>] (second-best) was equal to 2.1%, 1.9%, 2.4%, 2.3%, and 0.4% for F1, AP, AR, ACC, and AUC, respectively. Hence, the performance of all these existing baseline methods validated the superiority of our proposed model with a substantial performance difference.</p>
        <p>Moreover, comparative studies on the analysis of the cross–data set performance are rare. The majority of the studies only considered a similar data set for training and testing. Cross–data set testing is an important analysis to demonstrate the general capability of a model and its potential applicability in a real-world environment. Therefore, similar comparative results are also evaluated (in a cross data set) for different baseline models for a detailed performance comparison with the proposed ensemble-SDCNN model. In this analysis, the MC data set was used to train the model and SZ was used to test, and vice versa. <xref ref-type="table" rid="table6">Table 6</xref> shows the results of these cross–data set analyses along with comparative studies.</p>
        <p>These comparative results indicated that our model had outperformed the various deep learning and handcrafted feature-based TB diagnostic methods. For the SZ data set, which was used for training, the accuracies were slightly higher than those for the MC data set. The main reason for this was the presence of more training data samples compared with the MC data set. For the scenario in which the MC data set was the training set and the SZ the testing set, the performance of our model in contrast to that of Santosh and Antani [<xref ref-type="bibr" rid="ref16">16</xref>] (second best) was higher than 3.3%, 3.2%, and 3.3% for AR, ACC, and AUC, respectively, and the comparative performance difference of our model with that of Santosh and Antani [<xref ref-type="bibr" rid="ref16">16</xref>] (for SZ as training and MC as testing data sets) was also higher than 2.3%, 1.7%, and 2.3% for AR, ACC, and AUC, respectively. All these experimental results highlighted the potential applicability of our model in real-world diagnostics related to chest abnormalities.</p>
        <table-wrap position="float" id="table6">
          <label>Table 6</label>
          <caption>
            <p>Results of comparative performance analysis of our proposed method with various baseline methods for cross data sets.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="30"/>
            <col width="370"/>
            <col width="120"/>
            <col width="120"/>
            <col width="120"/>
            <col width="120"/>
            <col width="120"/>
            <thead>
              <tr valign="top">
                <td colspan="2">Data sets and our methods</td>
                <td>F1</td>
                <td>AP<sup>a</sup></td>
                <td>AR<sup>b</sup></td>
                <td>ACC<sup>c</sup></td>
                <td>AUC<sup>d</sup></td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td colspan="2">
                  <bold>MC<sup>e</sup> train and SZ<sup>f</sup> test</bold>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td rowspan="10">
                  <break/>
                </td>
                <td>LBP<sup>g</sup> and SVM<sup>h,i</sup> [<xref ref-type="bibr" rid="ref46">46</xref>]</td>
                <td>0.496</td>
                <td>0.492</td>
                <td>0.5</td>
                <td>0.492</td>
                <td>0.69</td>
              </tr>
              <tr valign="top">
                <td>HoG<sup>j</sup> and SVM<sup>i</sup> [<xref ref-type="bibr" rid="ref47">47</xref>]</td>
                <td>0.664</td>
                <td>0.695</td>
                <td>0.635</td>
                <td>0.639</td>
                <td>0.762</td>
              </tr>
              <tr valign="top">
                <td>ShuffleNet<sup>i</sup> [<xref ref-type="bibr" rid="ref43">43</xref>]</td>
                <td>0.661</td>
                <td>0.715</td>
                <td>0.615</td>
                <td>0.61</td>
                <td>0.709</td>
              </tr>
              <tr valign="top">
                <td>InceptionV3<sup>i</sup> [<xref ref-type="bibr" rid="ref44">44</xref>]</td>
                <td>0.708</td>
                <td>0.717</td>
                <td>0.7</td>
                <td>0.698</td>
                <td>0.761</td>
              </tr>
              <tr valign="top">
                <td>MobileNetV2<sup>i</sup> [<xref ref-type="bibr" rid="ref45">45</xref>]</td>
                <td>0.613</td>
                <td>0.678</td>
                <td>0.559</td>
                <td>0.565</td>
                <td>0.78</td>
              </tr>
              <tr valign="top">
                <td>ResNet50<sup>i</sup> [<xref ref-type="bibr" rid="ref29">29</xref>]</td>
                <td>0.686</td>
                <td>0.707</td>
                <td>0.667</td>
                <td>0.663</td>
                <td>0.77</td>
              </tr>
              <tr valign="top">
                <td>ResNet101<sup>i</sup> [<xref ref-type="bibr" rid="ref29">29</xref>]</td>
                <td>0.674</td>
                <td>0.677</td>
                <td>0.671</td>
                <td>0.672</td>
                <td>0.772</td>
              </tr>
              <tr valign="top">
                <td>GoogLeNet<sup>i</sup> [<xref ref-type="bibr" rid="ref20">20</xref>,<xref ref-type="bibr" rid="ref21">21</xref>]</td>
                <td>0.592</td>
                <td>0.595</td>
                <td>0.589</td>
                <td>0.591</td>
                <td>0.65</td>
              </tr>
              <tr valign="top">
                <td>Santosh and Antani [<xref ref-type="bibr" rid="ref16">16</xref>]</td>
                <td>—<sup>k</sup></td>
                <td>—</td>
                <td>0.76</td>
                <td>0.76</td>
                <td>0.82</td>
              </tr>
              <tr valign="top">
                <td>Proposed</td>
                <td>0.795</td>
                <td>0.798</td>
                <td>0.793</td>
                <td>0.792</td>
                <td>0.853</td>
              </tr>
              <tr valign="top">
                <td colspan="2">
                  <bold>SZ train and MC test</bold>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td rowspan="10">
                  <break/>
                </td>
                <td>LBP and SVM<sup>i</sup> [<xref ref-type="bibr" rid="ref46">46</xref>]</td>
                <td>0.537</td>
                <td>0.58</td>
                <td>0.5</td>
                <td>0.58</td>
                <td>0.552</td>
              </tr>
              <tr valign="top">
                <td>HoG and SVM<sup>i</sup> [<xref ref-type="bibr" rid="ref47">47</xref>]</td>
                <td>0.559</td>
                <td>0.573</td>
                <td>0.546</td>
                <td>0.594</td>
                <td>0.601</td>
              </tr>
              <tr valign="top">
                <td>ShuffleNet<sup>i</sup> [<xref ref-type="bibr" rid="ref43">43</xref>]</td>
                <td>0.633</td>
                <td>0.643</td>
                <td>0.624</td>
                <td>0.652</td>
                <td>0.683</td>
              </tr>
              <tr valign="top">
                <td>InceptionV3<sup>i</sup> [<xref ref-type="bibr" rid="ref44">44</xref>]</td>
                <td>0.681</td>
                <td>0.722</td>
                <td>0.644</td>
                <td>0.688</td>
                <td>0.748</td>
              </tr>
              <tr valign="top">
                <td>MobileNetV2<sup>i</sup> [<xref ref-type="bibr" rid="ref45">45</xref>]</td>
                <td>0.668</td>
                <td>0.772</td>
                <td>0.589</td>
                <td>0.652</td>
                <td>0.797</td>
              </tr>
              <tr valign="top">
                <td>ResNet50<sup>i</sup> [<xref ref-type="bibr" rid="ref29">29</xref>]</td>
                <td>0.64</td>
                <td>0.642</td>
                <td>0.638</td>
                <td>0.616</td>
                <td>0.787</td>
              </tr>
              <tr valign="top">
                <td>ResNet101<sup>i</sup> [<xref ref-type="bibr" rid="ref29">29</xref>]</td>
                <td>0.641</td>
                <td>0.726</td>
                <td>0.574</td>
                <td>0.638</td>
                <td>0.698</td>
              </tr>
              <tr valign="top">
                <td>GoogLeNet<sup>i</sup> [<xref ref-type="bibr" rid="ref20">20</xref>,<xref ref-type="bibr" rid="ref21">21</xref>]</td>
                <td>0.648</td>
                <td>0.691</td>
                <td>0.609</td>
                <td>0.659</td>
                <td>0.754</td>
              </tr>
              <tr valign="top">
                <td>Santosh and Antani [<xref ref-type="bibr" rid="ref16">16</xref>]</td>
                <td>—</td>
                <td>—</td>
                <td>0.79</td>
                <td>0.78</td>
                <td>0.85</td>
              </tr>
              <tr valign="top">
                <td>Proposed</td>
                <td>0.811</td>
                <td>0.808</td>
                <td>0.813</td>
                <td>0.797</td>
                <td>0.873</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table6fn1">
              <p><sup>a</sup>AP: average precision.</p>
            </fn>
            <fn id="table6fn2">
              <p><sup>b</sup>AR: average recall.</p>
            </fn>
            <fn id="table6fn3">
              <p><sup>c</sup>ACC: accuracy.</p>
            </fn>
            <fn id="table6fn4">
              <p><sup>d</sup>AUC: area under the curve.</p>
            </fn>
            <fn id="table6fn5">
              <p><sup>e</sup>MC: Montgomery County.</p>
            </fn>
            <fn id="table6fn6">
              <p><sup>f</sup>SZ: Shenzhen.</p>
            </fn>
            <fn id="table6fn7">
              <p><sup>g</sup>LBP: local binary pattern.</p>
            </fn>
            <fn id="table6fn8">
              <p><sup>h</sup>SVM: support vector machine.</p>
            </fn>
            <fn id="table6fn9">
              <p><sup>i</sup>We also evaluated the performance of these models (for the cross data set) using our selected data sets and experimental protocol.</p>
            </fn>
            <fn id="table6fn10">
              <p><sup>j</sup>HoG: histogram of oriented gradients.</p>
            </fn>
            <fn id="table6fn11">
              <p><sup>k</sup>—: not available. The results were not provided in this comparative study for these performance metrics.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <p>This article presents an interactive CAD framework based on multiscale information fusion to diagnose TB in CXR images and retrieve the relevant cases (CXR images) from a previous patients’ database including clinical observations. In this framework, a classification model is primarily proposed to classify the given CXR image as either a positive or a negative sample. Subsequently, classification-based retrieval is performed to retrieve the relevant cases and corresponding clinical readings based on our newly proposed MLSM algorithm. The proposed model substantially improves diagnostic performance by performing the fusion of both low- and high-level features. The network processes the input image through different layers and finally activates the class-specific discriminative region [<xref ref-type="bibr" rid="ref48">48</xref>] as key-features maps. <xref rid="figure7" ref-type="fig">Figure 7</xref> shows such activation maps extracted from the 7 different layers (ie, F<sub>SN1</sub>, F<sub>SN2</sub>, F<sub>DN1</sub>, F<sub>DN2</sub>, F<sub>DN3</sub>, F<sub>DN4</sub>, and F<sub>DN5</sub> as labeled in <xref rid="figure2" ref-type="fig">Figure 2</xref>) of our model for both positive and negative sample images. As <xref rid="figure7" ref-type="fig">Figure 7</xref> shows, each activation map is generated by calculating the average of all the extracted maps from a specific location. All the activation maps overlay on their corresponding input image after resizing and applying a pseudo-color scheme (blue to red, equivalent to lower to higher activated region) to produce a better visualization of the activated regions.</p>
      <fig id="figure7" position="float">
        <label>Figure 7</label>
        <caption>
          <p>Extracted features maps from the different parts of the proposed ensemble-SDCNN model for both TB positive and negative cases. DCNN: deep convolutional neural network; SDCNN: shallow–deep convolutional neural network; SCNN: shallow convolutional neural network; TB: tuberculosis.</p>
        </caption>
        <graphic xlink:href="medinform_v8i12e21790_fig7.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
      </fig>
      <p><xref rid="figure7" ref-type="fig">Figure 7</xref> indicates that the class-specific discriminative regions of the given input image become more prominent after processing through the successive layers of the network. A semilocalized activation map (labeled as F<sub>DN5</sub> in <xref rid="figure7" ref-type="fig">Figure 7</xref>) is obtained from the last convolutional layer of the DCNN model, which includes the more distinctive high-level features for each class. Moreover, for the SCNN, the obtained activation map from the last convolutional layer (labeled as F<sub>SN2</sub> in <xref rid="figure7" ref-type="fig">Figure 7</xref>) encompasses the low-level features such as edge information. Finally, both low- and high-level features are used in making an effective diagnostic decision for the given CXR image. The experimental results (also provided in <xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>) proved that the diagnostic performance of our ensemble-SDCNN model is more effective than the various CNN models where only single-level features are used for class prediction.</p>
      <p>After an effective diagnostic decision, we can further retrieve the relevant cases based on our proposed MLSM algorithm, which considers the multilevel features in retrieving the best matches. <xref rid="figure8" ref-type="fig">Figure 8</xref> depicts the retrieval results of our proposed MLSM algorithm in comparison with the conventional Euclidean distance–based SLSM scheme. In <xref rid="figure8" ref-type="fig">Figure 8</xref>, these results comprise the 5 best-matched CXR images along with their corresponding high-level activation maps (labeled as F<sub>DN5</sub> in <xref rid="figure7" ref-type="fig">Figure 7</xref>) and clinical readings. Generally, a high correlation between the high-level activation maps (as F<sub>DN5</sub> in our study) of the query image and retrieved image implies the optimal performance of a retrieval system. With our MLSM algorithm, these activation maps (corresponding to retrieved cases) were more analogous (in terms of shape and location) to that of query image compared with the conventional SLSM scheme. This implied that our algorithm retrieved the highly correlated cases in terms of TB patterns, location, and clinical observation.</p>
      <fig id="figure8" position="float">
        <label>Figure 8</label>
        <caption>
          <p>Visualization of retrieval performance for the given input query image by considering SLSM and MLSM (our proposed model). MLSM: multilevel similarity measure; SLSM: single-level similarity measure.</p>
        </caption>
        <graphic xlink:href="medinform_v8i12e21790_fig8.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
      </fig>
      <p>In addition, we evaluated the objective similarity score in terms of the PSNR between the activation maps of the input query and 20 best-matched cases for both algorithms (MLSM and SLSM). The main purpose of this analysis was to quantitatively evaluate such feature-level similarities of both algorithms. A total of 28 images (28/138, 20.2% of the MC data set) from the MC data set and 132 images (132/662, 19.9% of the SZ data set) from the SZ data set were selected as the query database to perform this analysis. Using each query image one at a time, we retrieved the 20 best-matched cases corresponding to each algorithm. Thus, 20 different PSNR values were computed corresponding to these retrieved images for each matching algorithm. After these results for the entire selected query database were evaluated, an average PSNR performance was calculated to present the average performance of a single query image for each algorithm. <xref rid="figure9" ref-type="fig">Figure 9</xref> shows the comparative performance results of our proposed MLSM algorithm and the conventional SLSM scheme. We observed that our matching algorithm exhibited the higher features-level similarity scores in terms of the PSNR (for all the retrieved images and both data sets) in contrast to the SLSM scheme. Thus, our algorithm resulted in an optimal retrieval performance because of the significant correlation of high-level activation maps. All these results (<xref rid="figure8" ref-type="fig">Figures 8</xref> and <xref rid="figure9" ref-type="fig">9</xref>) were computed based on our selected classification-driven retrieval method. The experimental results provided in <xref ref-type="table" rid="table4">Table 4</xref> have already proved that our selected class prediction–based retrieval method outperforms the retrieval method without class prediction.</p>
      <p>In addition to the numerical results provided in <xref ref-type="table" rid="table4">Table 4</xref>, <xref rid="figure10" ref-type="fig">Figure 10</xref> further distinguishes the retrieved results of these 2 different approaches (ie, with and without class prediction) figuratively. <xref rid="figure10" ref-type="fig">Figure 10</xref> indicates that all the retrieved cases (for the given query image) were TPs in our class prediction–based retrieval method.</p>
      <p>However, in the retrieval without class prediction, the first and third best matches were FPs (highlighted by the red bounding box) while the remaining three cases were TPs. Such FP cases may lead to a vague diagnostic decision. Additionally, the numerical results (<xref ref-type="table" rid="table4">Table 4</xref>) indicated that the average number of FPs in retrieval without class prediction was substantially higher than our class-prediction retrieval method. Therefore, in this study, we considered a classification-driven retrieval by performing the class prediction in the first step and then retrieving the best-matched cases from the predicted class database rather than exploring the entire database. Ultimately, the classification results can aid in making a diagnostic decision and the retrieved CXR images can assist radiologists to further validate the computer decision. Furthermore, if the wrong prediction is made by the computer, the medical expert can check other relevant cases (ie, second-, third-, or fourth-best matches) that can be more relevant than the first best match. Thus, both classification and retrieval results can aid radiologists in making an effective diagnostic decision even in scenarios of small TB patterns that remain undetectable in the early stage. Such a comprehensive CAD framework may assist radiologists in clinical practices and alleviate the burden of an increasing number of patients by providing an effective and timely diagnostic decision. Our trained model and the training and testing data splitting information are publicly available [<xref ref-type="bibr" rid="ref49">49</xref>] to enable other researchers to evaluate and compare its performance.</p>
      <fig id="figure9" position="float">
        <label>Figure 9</label>
        <caption>
          <p>PSNR-based objective similarity measures between the high-level activation maps of the query image and retrieved images to evaluate feature-level similarities of both algorithms (ie, MLSM and SLSM). MLSM: multilevel similarity measure; PSNR: peak signal-to-noise ratio; SLSM: single-level similarity measure.</p>
        </caption>
        <graphic xlink:href="medinform_v8i12e21790_fig9.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
      </fig>
      <fig id="figure10" position="float">
        <label>Figure 10</label>
        <caption>
          <p>Visualization of retrieval performance for the given input query image by considering both retrieval methods with class prediction and without class prediction. </p>
        </caption>
        <graphic xlink:href="medinform_v8i12e21790_fig10.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
      </fig>
    </sec>
  </body>
  <back>
    <app-group>
      <supplementary-material id="app1">
        <label>Multimedia Appendix 1</label>
        <p>Other supplementary material is provided in the attached word file [DOCX file (MS Word), 44 KB].</p>
        <media xlink:href="medinform_v8i12e21790_app1.docx" xlink:title="DOCX File , 44 KB"/>
      </supplementary-material>
      <supplementary-material id="app2">
        <label>Multimedia Appendix 2</label>
        <p>All the experimental results are provided in the attached excel file [XLSX file (MS Excel), 226 KB].</p>
        <media xlink:href="medinform_v8i12e21790_app2.xlsx" xlink:title="XLSX File  (Microsoft Excel File), 226 KB"/>
      </supplementary-material>
    </app-group>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">ACC</term>
          <def>
            <p>accuracy</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">AP</term>
          <def>
            <p>average precision</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">AR</term>
          <def>
            <p>average recall</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">AUC</term>
          <def>
            <p>area under the curve</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">CAD</term>
          <def>
            <p>computer-aided diagnosis</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">CL</term>
          <def>
            <p>class label</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb7">CNN</term>
          <def>
            <p>convolutional neural network</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb8">CXR</term>
          <def>
            <p>chest radiograph</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb9">DCNN</term>
          <def>
            <p>deep convolutional neural network</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb10">FN</term>
          <def>
            <p>false negatives</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb11">FP</term>
          <def>
            <p>false positives</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb12">FPR</term>
          <def>
            <p>false-positive rate</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb13">F1</term>
          <def>
            <p>F1 score</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb14">HoG</term>
          <def>
            <p>histogram of oriented gradients</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb15">LBP</term>
          <def>
            <p>local binary pattern</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb16">MC</term>
          <def>
            <p>Montgomery County</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb17">MLSM</term>
          <def>
            <p>multilevel similarity measure</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb18">PSNR</term>
          <def>
            <p>peak signal-to-noise ratio</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb19">ROC</term>
          <def>
            <p>receiver operating characteristic (curve)</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb20">SDCNN</term>
          <def>
            <p>shallow–deep convolutional neural network</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb21">SCNN</term>
          <def>
            <p>shallow convolutional neural network</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb22">SLSM</term>
          <def>
            <p>single-level similarity measure</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb23">SSIM</term>
          <def>
            <p>structure similarity</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb24">SVM</term>
          <def>
            <p>support vector machine.</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb25">SZ</term>
          <def>
            <p>Shenzhen</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb26">TB</term>
          <def>
            <p>tuberculosis</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb27">TN</term>
          <def>
            <p>true negative</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb28">TP</term>
          <def>
            <p>true positive</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb29">TPR</term>
          <def>
            <p>true-positive rate</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb30">WHO</term>
          <def>
            <p>World Health Organization</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <ack>
      <p>This work was supported in part by the Ministry of Science and ICT (MSIT), Korea, under the ITRC (Information Technology Research Center) support program (IITP-2020-2020-0-01789) supervised by the IITP (Institute for Information &#38; Communications Technology Promotion) and in part by the Bio and Medical Technology Development Program of the National Research Foundation of Korea (NRF) funded by the Korean government, the MSIT (NRF-2016M3A9E1915855).</p>
    </ack>
    <fn-group>
      <fn fn-type="con">
        <p>MO and KP designed the overall framework. Moreover, they wrote and revised the complete paper. MA, TM, and YK facilitated in designing comparative analysis and experiments.</p>
      </fn>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="web">
          <article-title>Global tuberculosis report</article-title>
          <source>World Health Organization</source>
          <year>2015</year>
          <access-date>2020-11-26</access-date>
          <publisher-loc>Geneva, Switzerland</publisher-loc>
          <publisher-name>World Health Organization</publisher-name>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://www.tbonline.info/posts/2015/10/28/global-tuberculosis-report-2015/">http://www.tbonline.info/posts/2015/10/28/global-tuberculosis-report-2015/</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Cheon</surname>
              <given-names>SA</given-names>
            </name>
            <name name-style="western">
              <surname>Cho</surname>
              <given-names>HH</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>HJ</given-names>
            </name>
            <name name-style="western">
              <surname>Park</surname>
              <given-names>TJ</given-names>
            </name>
          </person-group>
          <article-title>Recent tuberculosis diagnosis toward the end TB strategy</article-title>
          <source>J Microbiol Methods</source>
          <year>2016</year>
          <month>04</month>
          <volume>123</volume>
          <fpage>51</fpage>
          <lpage>61</lpage>
          <pub-id pub-id-type="doi">10.1016/j.mimet.2016.02.007</pub-id>
          <pub-id pub-id-type="medline">26853124</pub-id>
          <pub-id pub-id-type="pii">S0167-7012(16)30024-0</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Casela</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Cerqueira</surname>
              <given-names>SMA</given-names>
            </name>
            <name name-style="western">
              <surname>Casela</surname>
              <given-names>TDO</given-names>
            </name>
            <name name-style="western">
              <surname>Pereira</surname>
              <given-names>MA</given-names>
            </name>
            <name name-style="western">
              <surname>Santos</surname>
              <given-names>SQD</given-names>
            </name>
            <name name-style="western">
              <surname>Pozo</surname>
              <given-names>FAD</given-names>
            </name>
            <name name-style="western">
              <surname>Freire</surname>
              <given-names>SM</given-names>
            </name>
            <name name-style="western">
              <surname>Matos</surname>
              <given-names>ED</given-names>
            </name>
          </person-group>
          <article-title>Rapid molecular test for tuberculosis: impact of its routine use at a referral hospital</article-title>
          <source>J Bras Pneumol</source>
          <year>2018</year>
          <month>04</month>
          <volume>44</volume>
          <issue>2</issue>
          <fpage>112</fpage>
          <lpage>117</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.scielo.br/scielo.php?script=sci_arttext&#38;pid=S1806-37132018000200112&#38;lng=en&#38;nrm=iso&#38;tlng=en"/>
          </comment>
          <pub-id pub-id-type="doi">10.1590/s1806-37562017000000201</pub-id>
          <pub-id pub-id-type="medline">29791546</pub-id>
          <pub-id pub-id-type="pii">S1806-37132018000200112</pub-id>
          <pub-id pub-id-type="pmcid">PMC6044659</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Panteix</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Gutierrez</surname>
              <given-names>MC</given-names>
            </name>
            <name name-style="western">
              <surname>Boschiroli</surname>
              <given-names>ML</given-names>
            </name>
            <name name-style="western">
              <surname>Rouviere</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Plaidy</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Pressac</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Porcheret</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Chyderiotis</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Ponsada</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Van Oortegem</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Salloum</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Cabuzel</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Bañuls</surname>
              <given-names>AL</given-names>
            </name>
            <name name-style="western">
              <surname>Van de Perre</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Godreuil</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Pulmonary tuberculosis due to Mycobacterium microti: a study of six recent cases in France</article-title>
          <source>J Med Microbiol</source>
          <year>2010</year>
          <month>08</month>
          <volume>59</volume>
          <issue>8</issue>
          <fpage>984</fpage>
          <lpage>989</lpage>
          <pub-id pub-id-type="doi">10.1099/jmm.0.019372-0</pub-id>
          <pub-id pub-id-type="medline">20488936</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Schaefer-Prokop</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Neitzel</surname>
              <given-names>U</given-names>
            </name>
            <name name-style="western">
              <surname>Venema</surname>
              <given-names>HW</given-names>
            </name>
            <name name-style="western">
              <surname>Uffmann</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Prokop</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Digital chest radiography: an update on modern technology, dose containment and control of image quality</article-title>
          <source>Eur Radiol</source>
          <year>2008</year>
          <month>09</month>
          <volume>18</volume>
          <issue>9</issue>
          <fpage>1818</fpage>
          <lpage>30</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/18431577"/>
          </comment>
          <pub-id pub-id-type="doi">10.1007/s00330-008-0948-3</pub-id>
          <pub-id pub-id-type="medline">18431577</pub-id>
          <pub-id pub-id-type="pmcid">PMC2516181</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Raviglione</surname>
              <given-names>MC</given-names>
            </name>
            <name name-style="western">
              <surname>Flahault</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Use of Digital Technology to Enhance Tuberculosis Control: Scoping Review</article-title>
          <source>J Med Internet Res</source>
          <year>2020</year>
          <month>2</month>
          <day>13</day>
          <volume>22</volume>
          <issue>2</issue>
          <fpage>e15727</fpage>
          <pub-id pub-id-type="doi">10.2196/15727</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Gardezi</surname>
              <given-names>SJS</given-names>
            </name>
            <name name-style="western">
              <surname>Elazab</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Lei</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Breast Cancer Detection and Diagnosis Using Mammographic Data: Systematic Review</article-title>
          <source>J Med Internet Res</source>
          <year>2019</year>
          <month>07</month>
          <day>26</day>
          <volume>21</volume>
          <issue>7</issue>
          <fpage>e14464</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmir.org/2019/7/e14464/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/14464</pub-id>
          <pub-id pub-id-type="medline">31350843</pub-id>
          <pub-id pub-id-type="pii">v21i7e14464</pub-id>
          <pub-id pub-id-type="pmcid">PMC6688437</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Nielsen</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <source>Neural Networks and Deep Learning</source>
          <year>2015</year>
          <publisher-loc>San Francisco, CA</publisher-loc>
          <publisher-name>Determination Press</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jaeger</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Karargyris</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Candemir</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Folio</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Siegelman</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Callaghan</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Zhiyun</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Palaniappan</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Singh</surname>
              <given-names>RK</given-names>
            </name>
            <name name-style="western">
              <surname>Antani</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Thoma</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Yi-Xiang</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Pu-Xuan</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>McDonald</surname>
              <given-names>CJ</given-names>
            </name>
          </person-group>
          <article-title>Automatic Tuberculosis Screening Using Chest Radiographs</article-title>
          <source>IEEE Trans. Med. Imaging</source>
          <year>2014</year>
          <month>2</month>
          <volume>33</volume>
          <issue>2</issue>
          <fpage>233</fpage>
          <lpage>245</lpage>
          <pub-id pub-id-type="doi">10.1109/tmi.2013.2284099</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kumar</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Y-Y</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>K-C</given-names>
            </name>
            <name name-style="western">
              <surname>Tsai</surname>
              <given-names>I-C</given-names>
            </name>
            <name name-style="western">
              <surname>Huang</surname>
              <given-names>C-C</given-names>
            </name>
            <name name-style="western">
              <surname>Hung</surname>
              <given-names>N</given-names>
            </name>
          </person-group>
          <article-title>Distinguishing normal and pulmonary edema chest x-ray using Gabor filter and SVM</article-title>
          <year>2014</year>
          <conf-name>IEEE International Symposium on Bioelectronics and Bioinformatics (IEEE ISBB)</conf-name>
          <conf-date>3-6 April 2014</conf-date>
          <conf-loc>Chung Li, Taiwan</conf-loc>
          <fpage>117</fpage>
          <lpage>120</lpage>
          <pub-id pub-id-type="doi">10.1109/isbb.2014.6820918</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hogeweg</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Sanchez</surname>
              <given-names>CI</given-names>
            </name>
            <name name-style="western">
              <surname>Maduskar</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Philipsen</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Story</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Dawson</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Theron</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Dheda</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Peters-Bax</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>van Ginneken</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>Automatic Detection of Tuberculosis in Chest Radiographs Using a Combination of Textural, Focal, and Shape Abnormality Analysis</article-title>
          <source>IEEE Trans. Med. Imaging</source>
          <year>2015</year>
          <month>12</month>
          <volume>34</volume>
          <issue>12</issue>
          <fpage>2429</fpage>
          <lpage>2442</lpage>
          <pub-id pub-id-type="doi">10.1109/tmi.2015.2405761</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Carrillo-de-Gea</surname>
              <given-names>JM</given-names>
            </name>
            <name name-style="western">
              <surname>García-Mateos</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Fernández-Alemán</surname>
              <given-names>JL</given-names>
            </name>
            <name name-style="western">
              <surname>Hernández-Hernández</surname>
              <given-names>JL</given-names>
            </name>
          </person-group>
          <article-title>A Computer-Aided Detection System for Digital Chest Radiographs</article-title>
          <source>J Healthc Eng</source>
          <year>2016</year>
          <volume>2016</volume>
          <issue>1</issue>
          <fpage>8208923</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1155/2016/8208923"/>
          </comment>
          <pub-id pub-id-type="doi">10.1155/2016/8208923</pub-id>
          <pub-id pub-id-type="medline">27372536</pub-id>
          <pub-id pub-id-type="pii">8208923</pub-id>
          <pub-id pub-id-type="pmcid">PMC5058572</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Karargyris</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Siegelman</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Tzortzis</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Jaeger</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Candemir</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Xue</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Santosh</surname>
              <given-names>KC</given-names>
            </name>
            <name name-style="western">
              <surname>Vajda</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Antani</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Folio</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Thoma</surname>
              <given-names>GR</given-names>
            </name>
          </person-group>
          <article-title>Combination of texture and shape features to detect pulmonary abnormalities in digital chest X-rays</article-title>
          <source>Int J Comput Assist Radiol Surg</source>
          <year>2016</year>
          <month>01</month>
          <day>20</day>
          <volume>11</volume>
          <issue>1</issue>
          <fpage>99</fpage>
          <lpage>106</lpage>
          <pub-id pub-id-type="doi">10.1007/s11548-015-1242-x</pub-id>
          <pub-id pub-id-type="medline">26092662</pub-id>
          <pub-id pub-id-type="pii">10.1007/s11548-015-1242-x</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Alfadhli</surname>
              <given-names>FHO</given-names>
            </name>
            <name name-style="western">
              <surname>Mand</surname>
              <given-names>AA</given-names>
            </name>
            <name name-style="western">
              <surname>Sayeed</surname>
              <given-names>MD</given-names>
            </name>
            <name name-style="western">
              <surname>Sim</surname>
              <given-names>KS</given-names>
            </name>
            <name name-style="western">
              <surname>Al-Shabi</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Classification of tuberculosis with SURF spatial pyramid features</article-title>
          <year>2017</year>
          <conf-name>IEEE International Conference on Robotics, Automation and Sciences (ICORAS)</conf-name>
          <conf-date>27-29 November 2017</conf-date>
          <conf-loc>Melaka, Malaysia</conf-loc>
          <pub-id pub-id-type="doi">10.1109/icoras.2017.8308044</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Govindarajan</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Swaminathan</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Analysis of Tuberculosis in Chest Radiographs for Computerized Diagnosis using Bag of Keypoint Features</article-title>
          <source>J Med Syst</source>
          <year>2019</year>
          <month>03</month>
          <day>28</day>
          <volume>43</volume>
          <issue>4</issue>
          <fpage>87</fpage>
          <pub-id pub-id-type="doi">10.1007/s10916-019-1222-8</pub-id>
          <pub-id pub-id-type="medline">30820678</pub-id>
          <pub-id pub-id-type="pii">10.1007/s10916-019-1222-8</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Santosh</surname>
              <given-names>KC</given-names>
            </name>
            <name name-style="western">
              <surname>Antani</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Automated Chest X-Ray Screening: Can Lung Region Symmetry Help Detect Pulmonary Abnormalities?</article-title>
          <source>IEEE Trans. Med. Imaging</source>
          <year>2018</year>
          <month>5</month>
          <volume>37</volume>
          <issue>5</issue>
          <fpage>1168</fpage>
          <lpage>1177</lpage>
          <pub-id pub-id-type="doi">10.1109/tmi.2017.2775636</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hwang</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>H-E</given-names>
            </name>
            <name name-style="western">
              <surname>Jeong</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>H-J</given-names>
            </name>
          </person-group>
          <article-title>A novel approach for tuberculosis screening based on deep convolutional neural networks</article-title>
          <source>Proceedings of SPIE 9785, Medical Imaging 2016: Computer-Aided Diagnosis</source>
          <year>2016</year>
          <conf-name>SPIE 9785, Medical Imaging 2016: Computer-Aided Diagnosis</conf-name>
          <conf-date>24 March 2016</conf-date>
          <conf-loc>San Diego, CA, USA</conf-loc>
          <pub-id pub-id-type="doi">10.1117/12.2216198</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Shin</surname>
              <given-names>H-C</given-names>
            </name>
            <name name-style="western">
              <surname>Roberts</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Demner-Fushman</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Yao</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Summers</surname>
              <given-names>RM</given-names>
            </name>
          </person-group>
          <article-title>Learning to read chest x-rays: recurrent neural cascade model for automated image annotation</article-title>
          <year>2016</year>
          <conf-name>IEEE Conference on Computer Vision and Pattern Recognition</conf-name>
          <conf-date>27-30 June 2016</conf-date>
          <conf-loc>Las Vegas, NV, USA</conf-loc>
          <fpage>2497</fpage>
          <lpage>2506</lpage>
          <pub-id pub-id-type="doi">10.1109/cvpr.2016.274</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lakhani</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Sundaram</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>Deep Learning at Chest Radiography: Automated Classification of Pulmonary Tuberculosis by Using Convolutional Neural Networks</article-title>
          <source>Radiology</source>
          <year>2017</year>
          <month>08</month>
          <volume>284</volume>
          <issue>2</issue>
          <fpage>574</fpage>
          <lpage>582</lpage>
          <pub-id pub-id-type="doi">10.1148/radiol.2017162326</pub-id>
          <pub-id pub-id-type="medline">28436741</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Szegedy</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Jia</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Sermanet</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Reed</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Anguelov</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Erhan</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Vanhoucke</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Rabinovich</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Going deeper with convolutions</article-title>
          <year>2015</year>
          <conf-name>IEEE Conference Computer Vision Pattern Recognition</conf-name>
          <conf-date>7-12 June 2015</conf-date>
          <conf-loc>Boston, MA, USA</conf-loc>
          <fpage>1</fpage>
          <lpage>9</lpage>
          <pub-id pub-id-type="doi">10.1109/cvpr.2015.7298594</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lopes</surname>
              <given-names>UK</given-names>
            </name>
            <name name-style="western">
              <surname>Valiati</surname>
              <given-names>JF</given-names>
            </name>
          </person-group>
          <article-title>Pre-trained convolutional neural networks as feature extractors for tuberculosis detection</article-title>
          <source>Computers in Biology and Medicine</source>
          <year>2017</year>
          <month>10</month>
          <volume>89</volume>
          <issue>1</issue>
          <fpage>135</fpage>
          <lpage>143</lpage>
          <pub-id pub-id-type="doi">10.1016/j.compbiomed.2017.08.001</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Pasa</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Golkov</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Pfeiffer</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Cremers</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Pfeiffer</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Efficient Deep Network Architectures for Fast Chest X-Ray Tuberculosis Screening and Visualization</article-title>
          <source>Sci Rep</source>
          <year>2019</year>
          <month>04</month>
          <day>18</day>
          <volume>9</volume>
          <issue>1</issue>
          <fpage>6268</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41598-019-42557-4"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41598-019-42557-4</pub-id>
          <pub-id pub-id-type="medline">31000728</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41598-019-42557-4</pub-id>
          <pub-id pub-id-type="pmcid">PMC6472370</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Qin</surname>
              <given-names>ZZ</given-names>
            </name>
            <name name-style="western">
              <surname>Sander</surname>
              <given-names>MS</given-names>
            </name>
            <name name-style="western">
              <surname>Rai</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Titahong</surname>
              <given-names>CN</given-names>
            </name>
            <name name-style="western">
              <surname>Sudrungrot</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Laah</surname>
              <given-names>SN</given-names>
            </name>
            <name name-style="western">
              <surname>Adhikari</surname>
              <given-names>LM</given-names>
            </name>
            <name name-style="western">
              <surname>Carter</surname>
              <given-names>EJ</given-names>
            </name>
            <name name-style="western">
              <surname>Puri</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Codlin</surname>
              <given-names>AJ</given-names>
            </name>
            <name name-style="western">
              <surname>Creswell</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Using artificial intelligence to read chest radiographs for tuberculosis detection: A multi-site evaluation of the diagnostic accuracy of three deep learning systems</article-title>
          <source>Sci Rep</source>
          <year>2019</year>
          <month>10</month>
          <day>18</day>
          <volume>9</volume>
          <issue>1</issue>
          <fpage>15000</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41598-019-51503-3"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41598-019-51503-3</pub-id>
          <pub-id pub-id-type="medline">31628424</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41598-019-51503-3</pub-id>
          <pub-id pub-id-type="pmcid">PMC6802077</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Nash</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Kadavigere</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Andrade</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Sukumar</surname>
              <given-names>CA</given-names>
            </name>
            <name name-style="western">
              <surname>Chawla</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Shenoy</surname>
              <given-names>VP</given-names>
            </name>
            <name name-style="western">
              <surname>Pande</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Huddart</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Pai</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Saravu</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Deep learning, computer-aided radiography reading for tuberculosis: a diagnostic accuracy study from a tertiary hospital in India</article-title>
          <source>Sci Rep</source>
          <year>2020</year>
          <month>01</month>
          <day>14</day>
          <volume>10</volume>
          <issue>1</issue>
          <fpage>210</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41598-019-56589-3"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41598-019-56589-3</pub-id>
          <pub-id pub-id-type="medline">31937802</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41598-019-56589-3</pub-id>
          <pub-id pub-id-type="pmcid">PMC6959311</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Krizhevsky</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Sutskever</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Hinton</surname>
              <given-names>GE</given-names>
            </name>
          </person-group>
          <article-title>ImageNet classification with deep convolutional neural networks</article-title>
          <year>2012</year>
          <conf-name>25th International Conference on Neural Information Processing Systems</conf-name>
          <conf-date>3-6 December 2012</conf-date>
          <conf-loc>Lake Tahoe, NV, USA</conf-loc>
          <fpage>1097</fpage>
          <lpage>1105</lpage>
          <pub-id pub-id-type="doi">10.1145/3065386</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Triantafyllidis</surname>
              <given-names>AK</given-names>
            </name>
            <name name-style="western">
              <surname>Tsanas</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Applications of Machine Learning in Real-Life Digital Health Interventions: Review of the Literature</article-title>
          <source>J Med Internet Res</source>
          <year>2019</year>
          <month>04</month>
          <day>05</day>
          <volume>21</volume>
          <issue>4</issue>
          <fpage>e12286</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmir.org/2019/4/e12286/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/12286</pub-id>
          <pub-id pub-id-type="medline">30950797</pub-id>
          <pub-id pub-id-type="pii">v21i4e12286</pub-id>
          <pub-id pub-id-type="pmcid">PMC6473205</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref27">
        <label>27</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Identifying Lung Cancer Risk Factors in the Elderly Using Deep Neural Networks: Quantitative Analysis of Web-Based Survey Data</article-title>
          <source>J Med Internet Res</source>
          <year>2020</year>
          <month>03</month>
          <day>17</day>
          <volume>22</volume>
          <issue>3</issue>
          <fpage>e17695</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmir.org/2020/3/e17695/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/17695</pub-id>
          <pub-id pub-id-type="medline">32181751</pub-id>
          <pub-id pub-id-type="pii">v22i3e17695</pub-id>
          <pub-id pub-id-type="pmcid">PMC7109611</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref28">
        <label>28</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Candemir</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Jaeger</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Palaniappan</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Musco</surname>
              <given-names>JP</given-names>
            </name>
            <name name-style="western">
              <surname>Singh</surname>
              <given-names>RK</given-names>
            </name>
            <name name-style="western">
              <surname>Zhiyun</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Karargyris</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Antani</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Thoma</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>McDonald</surname>
              <given-names>CJ</given-names>
            </name>
          </person-group>
          <article-title>Lung Segmentation in Chest Radiographs Using Anatomical Atlases With Nonrigid Registration</article-title>
          <source>IEEE Trans. Med. Imaging</source>
          <year>2014</year>
          <month>2</month>
          <volume>33</volume>
          <issue>2</issue>
          <fpage>577</fpage>
          <lpage>590</lpage>
          <pub-id pub-id-type="doi">10.1109/tmi.2013.2290491</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref29">
        <label>29</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>He</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Ren</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Deep residual learning for image recognition</article-title>
          <year>2016</year>
          <conf-name>IEEE Conference on Computer Vision Pattern Recognition</conf-name>
          <conf-date>26 June-01 July 2016</conf-date>
          <conf-loc>Las Vegas, NV, USA</conf-loc>
          <fpage>770</fpage>
          <lpage>778</lpage>
          <pub-id pub-id-type="doi">10.1109/cvpr.2016.90</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref30">
        <label>30</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>He</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Ren</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Identity mappings in deep residual networks</article-title>
          <year>2016</year>
          <conf-name>European Conference on Computer Vision</conf-name>
          <conf-date>8-16 October 2016</conf-date>
          <conf-loc>Amsterdam, The Netherlands</conf-loc>
          <fpage>630</fpage>
          <lpage>645</lpage>
          <pub-id pub-id-type="doi">10.1007/978-3-319-46493-0_38</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref31">
        <label>31</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Gao</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Zheng</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Ruan</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Shang</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Patel</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>SD-CNN: A shallow-deep CNN for improved breast cancer diagnosis</article-title>
          <source>Comput Med Imaging Graph</source>
          <year>2018</year>
          <month>12</month>
          <volume>70</volume>
          <fpage>53</fpage>
          <lpage>62</lpage>
          <pub-id pub-id-type="doi">10.1016/j.compmedimag.2018.09.004</pub-id>
          <pub-id pub-id-type="medline">30292910</pub-id>
          <pub-id pub-id-type="pii">S0895-6111(18)30234-9</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref32">
        <label>32</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Bovik</surname>
              <given-names>AC</given-names>
            </name>
            <name name-style="western">
              <surname>Sheikh</surname>
              <given-names>HR</given-names>
            </name>
            <name name-style="western">
              <surname>Simoncelli</surname>
              <given-names>EP</given-names>
            </name>
          </person-group>
          <article-title>Image quality assessment: from error visibility to structural similarity</article-title>
          <source>IEEE Trans Image Process</source>
          <year>2004</year>
          <month>04</month>
          <volume>13</volume>
          <issue>4</issue>
          <fpage>600</fpage>
          <lpage>12</lpage>
          <pub-id pub-id-type="doi">10.1109/tip.2003.819861</pub-id>
          <pub-id pub-id-type="medline">15376593</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref33">
        <label>33</label>
        <nlm-citation citation-type="web">
          <source>Deep Learning Toolbox</source>
          <access-date>2020-05-01</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://in.mathworks.com/products/deeplearning.html">https://in.mathworks.com/products/deeplearning.html</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref34">
        <label>34</label>
        <nlm-citation citation-type="web">
          <source>Intel® Core i7-3770K Processor</source>
          <access-date>2020-05-01</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://ark.intel.com/content/www/us/en/ark/products/65523/intel-core-i7-3770k-processor-8m-cache-up-to-3-90-ghz.html">https://ark.intel.com/content/www/us/en/ark/products/65523/intel-core-i7-3770k-processor-8m-cache-up-to-3-90-ghz.html</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref35">
        <label>35</label>
        <nlm-citation citation-type="web">
          <source>GeForce GTX 1070</source>
          <access-date>2020-05-01</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.geforce.com/hardware/desktop-gpus/geforce-gtx1070/specifications">https://www.geforce.com/hardware/desktop-gpus/geforce-gtx1070/specifications</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref36">
        <label>36</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Li</surname>
              <given-names>X-L</given-names>
            </name>
          </person-group>
          <article-title>Preconditioned Stochastic Gradient Descent</article-title>
          <source>IEEE Trans. Neural Netw. Learning Syst</source>
          <year>2018</year>
          <month>5</month>
          <volume>29</volume>
          <issue>5</issue>
          <fpage>1454</fpage>
          <lpage>1466</lpage>
          <pub-id pub-id-type="doi">10.1109/tnnls.2017.2672978</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref37">
        <label>37</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>M</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>M.n</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>A Review on Evaluation Metrics for Data Classification Evaluations</article-title>
          <source>IJDKP</source>
          <year>2015</year>
          <month>03</month>
          <day>31</day>
          <volume>5</volume>
          <issue>2</issue>
          <fpage>01</fpage>
          <lpage>11</lpage>
          <pub-id pub-id-type="doi">10.5121/ijdkp.2015.5201</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref38">
        <label>38</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Livingston</surname>
              <given-names>EH</given-names>
            </name>
          </person-group>
          <article-title>Who was student and why do we care so much about his t-test?</article-title>
          <source>J Surg Res</source>
          <year>2004</year>
          <month>05</month>
          <day>01</day>
          <volume>118</volume>
          <issue>1</issue>
          <fpage>58</fpage>
          <lpage>65</lpage>
          <pub-id pub-id-type="doi">10.1016/j.jss.2004.02.003</pub-id>
          <pub-id pub-id-type="medline">15093718</pub-id>
          <pub-id pub-id-type="pii">S0022480404000496</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref39">
        <label>39</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Cohen</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>A power primer</article-title>
          <source>Psychological Bulletin</source>
          <year>1992</year>
          <volume>112</volume>
          <issue>1</issue>
          <fpage>155</fpage>
          <lpage>159</lpage>
          <pub-id pub-id-type="doi">10.1037/0033-2909.112.1.155</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref40">
        <label>40</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Nakagawa</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Cuthill</surname>
              <given-names>IC</given-names>
            </name>
          </person-group>
          <article-title>Effect size, confidence interval and statistical significance: a practical guide for biologists</article-title>
          <source>Biol Rev Camb Philos Soc</source>
          <year>2007</year>
          <month>11</month>
          <volume>82</volume>
          <issue>4</issue>
          <fpage>591</fpage>
          <lpage>605</lpage>
          <pub-id pub-id-type="doi">10.1111/j.1469-185X.2007.00027.x</pub-id>
          <pub-id pub-id-type="medline">17944619</pub-id>
          <pub-id pub-id-type="pii">BRV27</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref41">
        <label>41</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Santosh</surname>
              <given-names>KC</given-names>
            </name>
            <name name-style="western">
              <surname>Vajda</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Antani</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Thoma</surname>
              <given-names>GR</given-names>
            </name>
          </person-group>
          <article-title>Edge map analysis in chest X-rays for automatic pulmonary abnormality screening</article-title>
          <source>Int J Comput Assist Radiol Surg</source>
          <year>2016</year>
          <month>09</month>
          <day>19</day>
          <volume>11</volume>
          <issue>9</issue>
          <fpage>1637</fpage>
          <lpage>46</lpage>
          <pub-id pub-id-type="doi">10.1007/s11548-016-1359-6</pub-id>
          <pub-id pub-id-type="medline">26995600</pub-id>
          <pub-id pub-id-type="pii">10.1007/s11548-016-1359-6</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref42">
        <label>42</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Vajda</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Karargyris</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Jaeger</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Santosh</surname>
              <given-names>KC</given-names>
            </name>
            <name name-style="western">
              <surname>Candemir</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Xue</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Antani</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Thoma</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>Feature Selection for Automatic Tuberculosis Screening in Frontal Chest Radiographs</article-title>
          <source>J Med Syst</source>
          <year>2018</year>
          <month>06</month>
          <day>29</day>
          <volume>42</volume>
          <issue>8</issue>
          <fpage>146</fpage>
          <pub-id pub-id-type="doi">10.1007/s10916-018-0991-9</pub-id>
          <pub-id pub-id-type="medline">29959539</pub-id>
          <pub-id pub-id-type="pii">10.1007/s10916-018-0991-9</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref43">
        <label>43</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Zhou</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>ShuffleNet: An extremely efficient convolutional neural network for mobile devices</article-title>
          <year>2018</year>
          <conf-name>IEEE/CVF Conference on Computer Vision Pattern Recognition</conf-name>
          <conf-date>18-23 June 2018</conf-date>
          <conf-loc>Salt Lake City, UT, USA</conf-loc>
          <fpage>6848</fpage>
          <lpage>6856</lpage>
          <pub-id pub-id-type="doi">10.1109/cvpr.2018.00716</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref44">
        <label>44</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Szegedy</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Vanhoucke</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Ioffe</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Shlens</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Wojna</surname>
              <given-names>Z</given-names>
            </name>
          </person-group>
          <article-title>Rethinking the inception architecture for computer vision</article-title>
          <year>2016</year>
          <conf-name>IEEE Conference on Computer Vision Pattern Recognition</conf-name>
          <conf-date>27-30 June 2016</conf-date>
          <conf-loc>Las Vegas, NV, USA</conf-loc>
          <fpage>2818</fpage>
          <lpage>2826</lpage>
          <pub-id pub-id-type="doi">10.1109/cvpr.2016.308</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref45">
        <label>45</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sandler</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Howard</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Zhmoginov</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>L-C</given-names>
            </name>
          </person-group>
          <article-title>Mobilenetv2: Inverted residualslinear bottlenecks</article-title>
          <year>2018</year>
          <conf-name>IEEE/CVF Conference on Computer Vision Pattern Recognition</conf-name>
          <conf-date>18-23 June 2018</conf-date>
          <conf-loc>Salt Lake City, UT, USA</conf-loc>
          <fpage>4510</fpage>
          <lpage>4520</lpage>
          <pub-id pub-id-type="doi">10.1109/cvpr.2018.00474</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref46">
        <label>46</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Subrahmanyam</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Maheshwari</surname>
              <given-names>RP</given-names>
            </name>
            <name name-style="western">
              <surname>Balasubramanian</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Local maximum edge binary patterns: A new descriptor for image retrieval and object tracking</article-title>
          <source>Signal Processing</source>
          <year>2012</year>
          <month>6</month>
          <volume>92</volume>
          <issue>6</issue>
          <fpage>1467</fpage>
          <lpage>1479</lpage>
          <pub-id pub-id-type="doi">10.1016/j.sigpro.2011.12.005</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref47">
        <label>47</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Velmurugan</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Baboo</surname>
              <given-names>SS</given-names>
            </name>
          </person-group>
          <article-title>Image Retrieval using Harris Corners and Histogram of Oriented Gradients</article-title>
          <source>IJCA</source>
          <year>2011</year>
          <month>06</month>
          <day>30</day>
          <volume>24</volume>
          <issue>7</issue>
          <fpage>6</fpage>
          <lpage>10</lpage>
          <pub-id pub-id-type="doi">10.5120/2968-3968</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref48">
        <label>48</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhou</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Khosla</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Lapedriza</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Oliva</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Torralba</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Learning deep features for discriminative localization</article-title>
          <year>2016</year>
          <conf-name>IEEE Conference on Computer Vision Pattern Recognition</conf-name>
          <conf-date>27-30 June 2016</conf-date>
          <conf-loc>Las Vegas, NV, USA</conf-loc>
          <fpage>2921</fpage>
          <lpage>2929</lpage>
          <pub-id pub-id-type="doi">10.1109/cvpr.2016.319</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref49">
        <label>49</label>
        <nlm-citation citation-type="web">
          <source>Dongguk CAD framework for effective diagnosis of tuberculosis</source>
          <access-date>2020-05-01</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://dm.dgu.edu/link.html">http://dm.dgu.edu/link.html</ext-link>
          </comment>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
