<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.0" xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMI</journal-id>
      <journal-id journal-id-type="nlm-ta">JMIR Med Inform</journal-id>
      <journal-title>JMIR Medical Informatics</journal-title>
      <issn pub-type="epub">2291-9694</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v14i1e84417</article-id>
      <article-id pub-id-type="pmid"/>
      <article-id pub-id-type="doi">10.2196/84417</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>Ranking-Aware Multiple Instance Learning for Histopathology Slide Classification: Development and Validation Study</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Benis</surname>
            <given-names>Arriel</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Zhou</surname>
            <given-names>Yuanpin</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Lu</surname>
            <given-names>Yi</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Kim</surname>
            <given-names>Ho Heon</given-names>
          </name>
          <degrees>MS</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-7260-7504</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Hwang</surname>
            <given-names>Gisu</given-names>
          </name>
          <degrees>MS</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-1046-9286</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author">
          <name name-style="western">
            <surname>Jeong</surname>
            <given-names>Won Chan</given-names>
          </name>
          <degrees>BS</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0009-0008-1931-5957</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author" corresp="yes">
          <name name-style="western">
            <surname>Ko</surname>
            <given-names>Young Sin</given-names>
          </name>
          <degrees>MD, PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <address>
            <institution>AI R&amp;D Center, Seegene Medical Foundation</institution>
            <addr-line>17F, Seegene Medical Tower, 288 Dapsimni-ro, Dongdaemun-gu</addr-line>
            <addr-line>Seoul, 02637</addr-line>
            <country>Republic of Korea</country>
            <phone>82 2 2228 9349</phone>
            <email>noteasy@mf.seegene.com</email>
          </address>
          <xref rid="aff3" ref-type="aff">3</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-1319-4847</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>AI R&amp;D Center, Seegene Medical Foundation</institution>
        <addr-line>Seoul</addr-line>
        <country>Republic of Korea</country>
      </aff>
      <aff id="aff2">
        <label>2</label>
        <institution>Department of Medical Sciences, Graduate School, The Catholic University of Korea</institution>
        <addr-line>Seoul</addr-line>
        <country>Republic of Korea</country>
      </aff>
      <aff id="aff3">
        <label>3</label>
        <institution>Pathology Center, Seegene Medical Foundation</institution>
        <addr-line>Seoul</addr-line>
        <country>Republic of Korea</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Young Sin Ko <email>noteasy@mf.seegene.com</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <year>2026</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>4</day>
        <month>2</month>
        <year>2026</year>
      </pub-date>
      <volume>14</volume>
      <elocation-id>e84417</elocation-id>
      <history>
        <date date-type="received">
          <day>19</day>
          <month>9</month>
          <year>2025</year>
        </date>
        <date date-type="rev-request">
          <day>23</day>
          <month>10</month>
          <year>2025</year>
        </date>
        <date date-type="rev-recd">
          <day>14</day>
          <month>12</month>
          <year>2025</year>
        </date>
        <date date-type="accepted">
          <day>1</day>
          <month>1</month>
          <year>2026</year>
        </date>
      </history>
      <copyright-statement>©Ho Heon Kim, Gisu Hwang, Won Chan Jeong, Young Sin Ko. Originally published in JMIR Medical Informatics (https://medinform.jmir.org), 04.02.2026.</copyright-statement>
      <copyright-year>2026</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on https://medinform.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://medinform.jmir.org/2026/1/e84417" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>Multiple instance learning (MIL) is widely used for slide-level classification in digital pathology without requiring expert annotations. However, even partial expert annotations offer valuable supervision; few studies have effectively leveraged this information within MIL frameworks.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>This study aims to develop and evaluate a ranking-aware MIL framework, called rank induction, that effectively incorporates partial expert annotations to improve slide-level classification performance under realistic annotation constraints.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>We developed rank induction, a MIL approach that incorporates expert annotations using a pairwise rank loss inspired by RankNet. The method encourages the model to assign higher attention scores to annotated regions than to unannotated ones, guiding it to focus on diagnostically relevant patches. We evaluated rank induction on 2 public datasets (Camelyon16 and DigestPath2019) and an in-house dataset (Seegene Medical Foundation-stomach; SMF-stomach) and tested its robustness under 3 real-world conditions: low-data regimes, coarse within-slide annotations, and sparse slide-level annotations.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>Rank induction outperformed existing methodologies, achieving an area under the receiver operating characteristic curve (AUROC) of 0.839 on Camelyon16, 0.995 on DigestPath2019, and 0.875 on SMF-stomach. It remained robust under low-data conditions, maintaining an AUROC of 0.761 with only 60.2% (130/216) of the training data. When using coarse annotations (with 2240-pixel padding), performance slightly declined to 0.823. Remarkably, annotating just 20% (18/89) of the slides was enough to reach near-saturated performance (AUROC of 0.806, vs 0.839 with full annotations).</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>Incorporating expert annotations through ranking-based supervision improves MIL-based classification. Rank induction remains robust even with limited, coarse, or sparsely available annotations, demonstrating its practicality in real-world scenarios.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>multiple instance learning</kwd>
        <kwd>learning to rank</kwd>
        <kwd>digital pathology</kwd>
        <kwd>whole slide image</kwd>
        <kwd>weakly supervised learning</kwd>
        <kwd>mixed supervision</kwd>
        <kwd>data-efficient training</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <p>With the rise of digital pathology, whole slide images (WSIs) have become essential for computer-assisted diagnosis tasks such as tumor detection, subtype classification, and prognosis prediction [<xref ref-type="bibr" rid="ref1">1</xref>-<xref ref-type="bibr" rid="ref3">3</xref>]. Due to their extremely high resolution, WSIs are typically divided into smaller patches before being used in deep learning models. In fully supervised learning settings, this requires patch-level labels—an annotation process that is time consuming and labor intensive due to the scarcity of lesion-containing patches. To reduce the burden of exhaustive labeling, the multiple instance learning (MIL) framework has emerged as the dominant approach for slide-level prediction. MIL operates in a weakly supervised setting, using only slide-level labels and eliminating the need for detailed patch annotations. Despite this advantage, MIL models still require large datasets to achieve generalizable performance due to a weak supervision signal [<xref ref-type="bibr" rid="ref4">4</xref>]. In data-constrained settings, MIL training often becomes unstable, leading to a significant reduction in predictive accuracy. To address the limitations of weak supervision, recent studies can be broadly categorized based on the strength of supervision into three main streams: (1) slide-level approaches relying solely on slide labels, (2) pseudolabel-based methods that infer patch-level labels, and (3) expert-guided methods leveraging direct annotations.</p>
      <p>Slide-level approaches operate solely with slide-level supervision but aim to improve performance through semisupervised learning, self-supervised learning, or multiresolution architectures [<xref ref-type="bibr" rid="ref5">5</xref>-<xref ref-type="bibr" rid="ref12">12</xref>]. For instance, dual-stream multiple instance learning (DS-MIL) introduced a dual-stream architecture that models relationships between high- and low-attention instances to mitigate attention collapse and improve feature discrimination [<xref ref-type="bibr" rid="ref11">11</xref>]. Hierarchical image pyramid transformer (HIPT) leveraged hierarchical self-supervised vision transformers to capture multiscale contextual information across gigapixel WSIs, demonstrating that strong representation learning can alleviate weak supervision limitations [<xref ref-type="bibr" rid="ref12">12</xref>]. However, because these approaches still rely only on slide-level labels, the supervision signal remains weak and cannot fully guide the model toward diagnostically relevant regions.</p>
      <p>Pseudolabel-based methods generate inferred patch-level labels from attention maps to strengthen the weak supervision signal. A prominent data-efficient training using pseudolabel is clustering-constrained attention multiple instance learning (CLAM), which selects high- and low-attention instances within each slide to form pseudopositive or pseudonegative sets and trains an auxiliary instance classifier [<xref ref-type="bibr" rid="ref13">13</xref>]. Iterative multiple instance learning proposed an iterative framework that refines both instance and slide representations by retraining the feature extractor with attention-derived pseudolabels, though its performance depends on the stability of the initial MIL attention [<xref ref-type="bibr" rid="ref14">14</xref>]. In general, pseudolabeling techniques aim to fill the gap in patch-level labels by generating inferred annotations [<xref ref-type="bibr" rid="ref13">13</xref>-<xref ref-type="bibr" rid="ref18">18</xref>]. Despite their utility, pseudolabels are inherently noisy, as they are inferred rather than observed. Consequently, MIL pipelines that solely rely on pseudolabels often suffer from degraded performance.</p>
      <p>In contrast, expert-guided methods offer a promising way to improve both performance and supervision quality by directly integrating expert annotations. Nevertheless, this approach remains relatively underexplored compared to other categories. A notable study described the attention induction model, which uses pathologist-drawn lesion annotations to guide model attention [<xref ref-type="bibr" rid="ref19">19</xref>]. However, by enforcing attention scores proportional to the size of the annotated regions, this method inherently biases the model toward larger areas, increasing the risk of overlooking small but clinically critical diagnostic features.</p>
      <p>To overcome this limitation, we propose rank induction, a novel MIL framework that introduces a pathology-informed inductive bias. Rather than constraining attention based on region size, rank induction assumes that annotated regions should carry higher diagnostic importance than unannotated ones. By learning relative ranks between patches, our method provides finer-grained supervision, guiding the model to focus on diagnostically relevant areas even when annotations are sparse or training data are limited.</p>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <sec>
        <title>Problem Definition</title>
        <p>We extend the standard MIL framework by incorporating patch-level annotations from pathologists as explicit instance labels, either across all training slides or only a subset. In a typical MIL setting, each WSI is divided into <italic>K</italic> patches {x<sub>1</sub>,…,x<italic><sub>K</sub></italic>}, and only a slide-level label, indicating disease presence, is provided. A slide is labeled positive if at least 1 patch contains a lesion and negative if all patches are normal (for simplicity, we assume a fixed <italic>K</italic> across slides).</p>
        <p>In our extended setting, we additionally use binary labels {y<sub>1</sub>,...,y<sub>k</sub>} at the patch level, where y<sub>k</sub> ∈{0,1}, y<sub>k</sub>=1, if a patch overlaps with a pathologist-annotated region, and y<sub>k</sub>=0 otherwise. The core inductive bias we introduce is simple but clinically grounded: patches overlapping annotated regions should receive higher attention than unannotated ones. This reflects the assumption that tumor-confirmed regions are more relevant to slide-level diagnosis than those confirmed to be normal.</p>
      </sec>
      <sec>
        <title>Rank Induction</title>
        <p>As illustrated in <xref rid="figure1" ref-type="fig">Figure 1</xref>, rank induction consists of 2 key components: pairwise rank comparison derived from rich annotation supervision and attention thresholding to filter out residual noninformative patches. To implement our ranking-aware attention mechanism, we first define the computation of patch-wise attention. Let z<sub>k</sub> denote the feature vector of patch x<sub>k</sub> extracted by a backbone feature encoder. We compute a raw attention score s<sub>k</sub> from z<sub>k</sub> using an attention layer and normalize it using softmax: <inline-graphic xlink:href="medinform_v14i1e84417_fig7.png" xlink:type="simple" mimetype="image"/>. Inspired by RankNet [<xref ref-type="bibr" rid="ref20">20</xref>], rank induction introduces a ranking constraint that encourages higher attention scores for annotated patches than for nonannotated ones, such that s<sub>i</sub> &gt; s<sub>j</sub>, where s<sub>i</sub> is the attention score for an annotated patch and s<sub>j</sub> for a nonannotated patch.</p>
        <fig id="figure1" position="float">
          <label>Figure 1</label>
          <caption>
            <p>Overview of rank induction.</p>
          </caption>
          <graphic xlink:href="medinform_v14i1e84417_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <p>To avoid the normalization constraint of softmax (<inline-graphic xlink:href="medinform_v14i1e84417_fig8.png" xlink:type="simple" mimetype="image"/>), we apply the ranking loss directly on the raw attention scores s<sub>i</sub> ε R. We define the pairwise ranking preference using a sigmoid function:</p>
        <p>
          <graphic xlink:href="medinform_v14i1e84417_fig9.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </p>
        <p>Here, <italic>m</italic> is a margin parameter that enforces separation between positive and negative patches and σ is a scaling factor. We exclude same-class pairs (eg, lesion-lesion or normal-normal) and define the valid pair set: P = {(i,j)|y<sub>i</sub> = 1, y<sub>j</sub> = 0}. Let <inline-graphic xlink:href="medinform_v14i1e84417_fig10.png" xlink:type="simple" mimetype="image"/> be the ground truth ranking label. The rank loss is defined as:</p>
        <p>
          <graphic xlink:href="medinform_v14i1e84417_fig11.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </p>
        <p>We jointly optimize the slide-level binary cross-entropy loss and the rank loss:</p>
        <disp-formula>L = L<sub>BCE</sub> + λL<sub>rank</sub></disp-formula>
        <p>where λ is a weighting hyperparameter. For slides without annotated lesions (ie, negative slides), we omit the ranking loss and only apply the slide-level loss.</p>
        <p>To mitigate slide-level representation dilution, where a few high-attention lesion patches are overwhelmed by numerous low-attention nonlesion patches during the attention-weighted summation, we apply an attention thresholding strategy during training. Specifically, we threshold the normalized attention weights and then renormalize all (including zero-valued) weights as follows:</p>
        <p>
          <graphic xlink:href="medinform_v14i1e84417_fig12.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </p>
        <p>Here, <italic>T</italic> is a predefined threshold and <italic>K</italic> is the number of patches. The threshold <italic>T</italic> is introduced to prevent the slide-level representation from being dominated by the large number of negative instances when the bag size is large. For example, in a slide containing 2000 instances with only 10 lesion patches, even if attention is highly concentrated on those few positives, the aggregate contribution from thousands of near-zero attention weights of negative patches can still overshadow the lesion signal, leading to degraded slide representations. By applying <italic>T</italic>, we effectively suppress these noise-like weights and maintain a more balanced representation across instances.</p>
      </sec>
      <sec>
        <title>Datasets</title>
        <p>We evaluated rank induction on 2 public datasets (Camelyon16 [<xref ref-type="bibr" rid="ref21">21</xref>] and DigestPath2019 [<xref ref-type="bibr" rid="ref22">22</xref>]) and an in-house dataset (Seegene Medical Foundation-stomach; SMF-stomach). <xref ref-type="boxed-text" rid="box1">Textbox 1</xref> provides a detailed description of the datasets used in this study, with sample count statistics summarized in <xref ref-type="table" rid="table1">Table 1</xref>, which presents the statistics for the 3 datasets.</p>
        <boxed-text id="box1" position="float">
          <title>Description of the datasets used in this study.</title>
          <list list-type="bullet">
            <list-item>
              <p>Camelyon16: this dataset consists of 399 hematoxylin and eosin-stained whole slide images (WSIs) of lymph node sections for breast cancer metastasis detection. It is originally split into 67.7% (270/399) training slides and 32.3% (129/399) test slides. For model development and hyperparameter tuning, we further split the 270 training WSIs into 216 (80%) training slides and 54 (20%) validation slides (an 8:2 split). Each slide includes a slide-level diagnosis and polygon-based lesion annotations provided by expert pathologists. In this study, slides containing metastasis were labeled as abnormal, while negative slides were labeled as normal.</p>
            </list-item>
            <list-item>
              <p>DigestPath2019: we used the colonoscopy tissue segmentation dataset from the DigestPath Challenge 2019. The dataset originally includes 660 training images and 212 test images with slide-level labels (benign or malignant). However, because the test set is not publicly available, we used only the 660 labeled WSIs. We allocated 20% (132/660) of the slides as the test set and used the remaining 80% (528/660) for training and validation. Specifically, we split the 528 slides into 422 (79.9%) for training and 106 (20.1%) for validation. Each WSI is a JPEG image with an average resolution of approximately 5000×5000 pixels. For consistency, we mapped malignant samples to the abnormal class and benign samples to the normal class.</p>
            </list-item>
            <list-item>
              <p>SMF-stomach (in-house): we constructed an in-house gastric pathology dataset at Seegene Medical Foundation (SMF). Slide preparation and digitization followed standard operating procedures described by Kim et al [<xref ref-type="bibr" rid="ref23">23</xref>]. A distinct feature of this dataset, as illustrated in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>, is that at least 3 consecutive tissue sections are placed on a single glass slide; however, pixel-level annotations were not exhaustively applied to every section. A single board-certified pathologist reviewed the slides and provided polygon-based annotations only for representative regions; thus, unannotated tissue sections may coexist within the same WSI. Following the diagnostic criteria of Ko et al [<xref ref-type="bibr" rid="ref24">24</xref>], lesions were originally categorized as malignant, dysplasia, negative, and uncategorized. We excluded the uncategorized class due to its high heterogeneity. For the binary classification task, we mapped malignant and dysplasia to the abnormal class and negative to the normal class. The final dataset comprised 209 slides, allocated as 120 (57.4%) for training, 30 (14.4%) for validation, and 59 (28.2%) for testing.</p>
            </list-item>
          </list>
        </boxed-text>
        <table-wrap position="float" id="table1">
          <label>Table 1</label>
          <caption>
            <p>Dataset statistics (whole slide images).</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="210"/>
            <col width="140"/>
            <col width="110"/>
            <col width="140"/>
            <col width="110"/>
            <col width="140"/>
            <col width="150"/>
            <thead>
              <tr valign="top">
                <td>Dataset</td>
                <td colspan="2">Train</td>
                <td colspan="2">Validation</td>
                <td colspan="2">Test</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Abnormal</td>
                <td>Normal</td>
                <td>Abnormal</td>
                <td>Normal</td>
                <td>Abnormal</td>
                <td>Normal</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Camelyon16</td>
                <td>89</td>
                <td>127</td>
                <td>22</td>
                <td>32</td>
                <td>49</td>
                <td>80</td>
              </tr>
              <tr valign="top">
                <td>DigestPath2019</td>
                <td>160</td>
                <td>262</td>
                <td>40</td>
                <td>66</td>
                <td>50</td>
                <td>82</td>
              </tr>
              <tr valign="top">
                <td>SMF-stomach<sup>a</sup></td>
                <td>81</td>
                <td>39</td>
                <td>20</td>
                <td>10</td>
                <td>39</td>
                <td>20</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table1fn1">
              <p><sup>a</sup>SMF-stomach: Seegene Medical Foundation-stomach.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Data Preprocessing</title>
        <p>We applied the same preprocessing pipeline to all datasets. First, we performed tissue-background segmentation using the multilevel Otsu algorithm. Each WSI was converted to grayscale and thresholded to generate a binary mask that separated tissue from the white background. We retained only the foreground tissue regions and discarded blank or irrelevant areas.</p>
        <p>Next, we tessellated each foreground region into nonoverlapping patches of size 224×224 pixels. For Camelyon16 and SMF-stomach, we extracted patches at 20× magnification, whereas for DigestPath2019, we used the original image resolution.</p>
        <p>After tessellation, we computed the tissue coverage ratio for each patch and discarded patches containing less than 5% tissue, as these were considered mostly empty. Finally, we used a ResNet-50 [<xref ref-type="bibr" rid="ref24">24</xref>] pretrained on ImageNet-1k [<xref ref-type="bibr" rid="ref25">25</xref>] to extract features from each patch. Specifically, we used the output of the third residual block and applied an adaptive average pooling layer to obtain a 2048-dimensional feature vector.</p>
      </sec>
      <sec>
        <title>Statistical Methods</title>
        <p>To evaluate rank induction, we designed a series of experiments targeting performance, data efficiency, and annotation use. Our evaluation included 6 different experiment types, which are discussed in the following subsections.</p>
        <sec>
          <title>Model Comparison</title>
          <p>We trained and evaluated 7 different MIL models on 3 datasets: the baseline attention-based MIL (AB-MIL), the attention induction method, 2 variants of CLAM (single-branch [SB]; multibranch [MB]), DS-MIL, HIPT, and the proposed rank induction method.</p>
        </sec>
        <sec>
          <title>Low-Data Regime</title>
          <p>To test data efficiency, we trained models on subsets of the training data at 20%, 40%, 60%, 80%, and 100% of the full set. For each fraction, we trained all 7 models (AB-MIL, attention induction, CLAM-SB, CLAM-MB, DS-MIL, HIPT, and rank induction) from scratch and evaluated them on a common held-out test set.</p>
        </sec>
        <sec>
          <title>Annotation Granularity</title>
          <p>To simulate real-world coarse annotations, we degraded annotation precision by expanding the original lesion masks on Camelyon16 slides. At 40× magnification, we added symmetric paddings of 0, 448, 896, 1344, 1792, and 2240 pixels to each polygon annotation. For instance, a 448-pixel padding at 40× resolution adds 1 patch-width around the lesion, which corresponds to a 224-pixel patch at 20× magnification—potentially including normal tissue and simulating noisy annotations.</p>
        </sec>
        <sec>
          <title>Subset-Annotated Slide Setting</title>
          <p>In real-world clinical settings, not all slides can be exhaustively annotated. We tested rank induction under partial supervision by varying the proportion of positive slides with patch-level annotations (5%, 10%, 20%, 40%, 60%, and 80%) in the Camelyon16 training set (111 positive slides in total). The remaining positive slides were treated with slide-level labels only. Negative slides, by definition, had no lesion annotations.</p>
        </sec>
        <sec>
          <title>Ablation Test</title>
          <p>To assess the contribution of attention thresholding, we performed an ablation study with threshold values <italic>T</italic>∈{0,0.25,0.5,1,2,5}. The case of <italic>T</italic>=0 served as the reference condition (ie, without attention thresholding). We compared WSI classification performance and the cumulative attention weight assigned to nonlesion patches on the test set to examine how different thresholds affect model behavior.</p>
        </sec>
        <sec>
          <title>Attention Localization</title>
          <p>To assess interpretability, we compared how different MIL models spatially focus on tumor regions in Camelyon16. This is not a dense pixel-wise segmentation map; rather, it is a ranking over discrete patches, with scores defined for all patches in the slide [<xref ref-type="bibr" rid="ref13">13</xref>]. Because no established gold standard protocol exists for evaluating such attention maps, we quantified spatial alignment between attention and expert lesion annotations using 3 metrics: intersection over union (IoU), Dice score, and pointing game [<xref ref-type="bibr" rid="ref25">25</xref>]. For each WSI, we ranked all patches by attention weight and then selected only the top α% relative to the number of tumor-annotated patches in that slide (α∈{1, 5, 10, 20, 50, 75, 100}). To ensure that the ideal case could always achieve a score of 1, IoU and Dice score were computed by sampling up to the number of tumor-annotated patches, whereas the pointing game was evaluated by sampling from all patches in the slide. We then computed IoU and Dice score between the selected high-attention patches and the tumor-annotated patches, averaged over positive test slides. We also measured the pointing game by calculating a hit rate, defined as whether at least 1 of the top α highest attention patches fell within the annotated tumor region (classified as a hit). For all malignant test slides, localization accuracy was defined as hits divided by the sum of hits and misses.</p>
          <p>For model evaluation, we computed area under the receiver operating characteristic curve (AUROC), area under the precision-recall curve (AUPRC), and accuracy for slide-level classification. All results were averaged over a 10-fold Monte Carlo cross-validation. We assessed statistical significance using the Mann-Whitney <italic>U</italic> test, with <italic>P</italic> values &lt;.05 considered statistically significant.</p>
        </sec>
      </sec>
      <sec>
        <title>Experiment Details</title>
        <p>We implemented the models in Python using the PyTorch deep learning framework. WSI data preprocessing was carried out with OpenSlide, OpenCV, and Pillow libraries. Visualization of attention maps and result plots was generated using Seaborn and Matplotlib libraries (specific version information for each library can be found in the GitHub repository). The training and experiments were run on a computing system with a pair of NVIDIA A100 graphics processing units (80 GB memory each) and an Intel Xeon Gold 6338N central processing unit (128 cores, 2.20 GHz) with 512 GB of RAM. Additionally, because the original implementation of attention induction was not publicly available, we reimplemented the method based on the descriptions provided in the paper. Our implementation is available to facilitate reproducibility [<xref ref-type="bibr" rid="ref26">26</xref>].</p>
        <p>For optimization, we used the Adam optimizer with a learning rate of 2×10<sup>–4</sup> and a weight decay of 1×10<sup>–5</sup>. We trained each model for up to 200 epochs using an early stopping criterion: if the validation loss did not improve for 7 consecutive epochs (after an initial 20-epoch warmup period), training was stopped. In practice, most models converged well before 200 epochs under this criterion. For all MIL benchmarks, the ResNet-50 backbone was frozen during training. Crucially, we used a frozen ResNet-50 backbone for all MIL benchmarks. Although methods such as DS-MIL and HIPT typically leverage self-supervised pretraining or hierarchical encoders, we used ResNet-50 as the feature extractor for this experiment. This ensured a direct comparison of the MIL aggregation strategies themselves, isolating their performance from the variations in pretrained feature representations.</p>
        <p>We set the rank loss hyperparameters as follows: the scaling factor σ=1 and margin m=1 in the pairwise probability formula, and the rank loss weight λ=1. In each training iteration, to limit memory use and ensure computational efficiency when computing the rank loss, we randomly sampled up to 1024 positive and 1024 negative patches per slide. The attention threshold <italic>T</italic> was set to 1.</p>
      </sec>
      <sec>
        <title>Ethical Considerations</title>
        <p>This study used both publicly available deidentified datasets (Camelyon16 and DigestPath2019) and a private clinical dataset (SMF-stomach). The Camelyon16 and DigestPath2019 datasets consist of fully deidentified WSIs, publicly released for research under their respective challenge licenses; therefore, no institutional review board (IRB) approval or informed consent was required for their use. For the SMF-stomach dataset, this study was approved by the IRB of the Seegene Medical Foundation (SMF-IRB-2024-015). Given the anonymous and deidentified nature of the retrospective pathological images, informed consent was not required.</p>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>Model Performance</title>
        <p>We compared rank induction against 6 representative MIL benchmarks: AB-MIL, attention induction, CLAM, DS-MIL, and HIPT, regarding slide-level classification performance. On Camelyon16, rank induction achieved an AUROC of 0.839 (SD 0.050) and an AUPRC of 0.850 (SD 0.042), outperforming all baseline methods with consistently low variance (<xref ref-type="table" rid="table2">Table 2</xref>). We observed similar performance trends on the DigestPath2019 dataset. Although all models performed well due to the dataset’s relative simplicity, rank induction achieved the best results (AUROC=0.995; AUPRC=0.993) with stable variance. On the SMF-stomach dataset, rank induction demonstrated the best performance, achieving an AUROC of 0.875 (SD 0.008) and an AUPRC of 0.937 (SD 0.006). Due to space constraints, the full accuracy comparison for all datasets is reported in <xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>.</p>
        <table-wrap position="float" id="table2">
          <label>Table 2</label>
          <caption>
            <p>Comparison of model performance across methods on 2 public datasets and an in-house dataset.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="120"/>
            <col width="160"/>
            <col width="140"/>
            <col width="140"/>
            <col width="160"/>
            <col width="0"/>
            <col width="110"/>
            <col width="170"/>
            <thead>
              <tr valign="top">
                <td>Model</td>
                <td colspan="2">Camelyon16</td>
                <td colspan="3">DigestPath2019</td>
                <td colspan="2">SMF<sup>a</sup>-stomach</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>AUROC<sup>b</sup>, mean (SD)</td>
                <td>AUPRC<sup>c</sup>, mean (SD)</td>
                <td>AUROC, mean (SD)</td>
                <td>AUPRC, mean (SD)</td>
                <td colspan="2">AUROC, mean (SD)</td>
                <td>AUPRC, mean (SD)</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>AB-MIL<sup>d</sup></td>
                <td>0.741 (0.146)</td>
                <td>0.730 (0.183)</td>
                <td>0.993 (0.003)</td>
                <td>0.990 (0.005)</td>
                <td colspan="2">0.864 (0.031)</td>
                <td>0.928 (0.016)</td>
              </tr>
              <tr valign="top">
                <td>Attention induction</td>
                <td>0.742 (0.142)</td>
                <td>0.727 (0.179)</td>
                <td>0.994 (0.002)</td>
                <td>0.990 (0.003)</td>
                <td colspan="2">0.868 (0.031)<sup>e</sup></td>
                <td>0.930 (0.016)<sup>e</sup></td>
              </tr>
              <tr valign="top">
                <td>CLAM-SB<sup>f</sup></td>
                <td>0.732 (0.137)</td>
                <td>0.700 (0.179)</td>
                <td>0.977 (0.019)</td>
                <td>0.965 (0.026)</td>
                <td colspan="2">0.837 (0.023)</td>
                <td>0.919 (0.013)</td>
              </tr>
              <tr valign="top">
                <td>CLAM-MB<sup>g</sup></td>
                <td>0.794 (0.128)<sup>e</sup></td>
                <td>0.767 (0.168)<sup>e</sup></td>
                <td>0.976 (0.019)</td>
                <td>0.963 (0.026)</td>
                <td colspan="2">0.838 (0.032)</td>
                <td>0.916 (0.025)</td>
              </tr>
              <tr valign="top">
                <td>DS-MIL<sup>h</sup></td>
                <td>0.690 (0.170)</td>
                <td>0.666 (0.223)</td>
                <td>0.995 (0.004)<sup>e</sup></td>
                <td>0.993 (0.006)<sup>e</sup></td>
                <td colspan="2">0.839 (0.031)</td>
                <td>0.915 (0.016)</td>
              </tr>
              <tr valign="top">
                <td>HIPT<sup>i</sup></td>
                <td>0.483 (0.103)</td>
                <td>0.416 (0.136)</td>
                <td>0.962 (0.058)</td>
                <td>0.947 (0.091)</td>
                <td colspan="2">0.676 (0.064)</td>
                <td>0.751 (0.060)</td>
              </tr>
              <tr valign="top">
                <td>Rank induction</td>
                <td>
                  <italic>0.839</italic>
                  <italic>(</italic>
                  <italic>0.</italic>
                  <italic>050</italic>
                  <italic>)</italic>
                  <sup>j,k,</sup>
                  <sup>l,m,</sup>
                  <sup>o</sup>
                  <sup>,p</sup>
                </td>
                <td>
                  <italic>0.850</italic>
                  <italic>(</italic>
                  <italic>0.</italic>
                  <italic>042</italic>
                  <italic>)</italic>
                  <sup>j,k,</sup>
                  <sup>l,m,</sup>
                  <sup>o</sup>
                  <sup>,p</sup>
                </td>
                <td>
                  <italic>0.995</italic>
                  <italic>(</italic>
                  <italic>0.</italic>
                  <italic>002</italic>
                  <italic>)</italic>
                  <sup>j,k,</sup>
                  <sup>l,m,n,p</sup>
                </td>
                <td>
                  <italic>0.993</italic>
                  <italic>(</italic>
                  <italic>0.</italic>
                  <italic>002</italic>
                  <italic>)</italic>
                  <sup>j,k,</sup>
                  <sup>l,m,n,</sup>
                  <sup>p</sup>
                </td>
                <td colspan="2">
                  <italic>0.875</italic>
                  <italic>(</italic>
                  <italic>0.</italic>
                  <italic>008</italic>
                  <italic>)</italic>
                  <sup>j,m,</sup>
                  <sup>n,o,p</sup>
                </td>
                <td>
                  <italic>0.937</italic>
                  <italic>(</italic>
                  <italic>0.</italic>
                  <italic>006</italic>
                  <italic>)</italic>
                  <sup>j,m,</sup>
                  <sup>n,o,p</sup>
                </td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table2fn1">
              <p><sup>a</sup>SMF: Seegene Medical Foundation.</p>
            </fn>
            <fn id="table2fn2">
              <p><sup>b</sup>AUROC: area under the receiver operating characteristic curve.</p>
            </fn>
            <fn id="table2fn3">
              <p><sup>c</sup>AUPRC: area under the precision-recall curve.</p>
            </fn>
            <fn id="table2fn4">
              <p><sup>d</sup>AB-MIL: attention-based multiple instance learning.</p>
            </fn>
            <fn id="table2fn5">
              <p><sup>e</sup>Second-best result.</p>
            </fn>
            <fn id="table2fn6">
              <p><sup>f</sup>CLAM-SB: clustering-constrained attention multiple instance learning, single-branch.</p>
            </fn>
            <fn id="table2fn7">
              <p><sup>g</sup>CLAM-MB: clustering-constrained attention multiple instance learning, multibranch.</p>
            </fn>
            <fn id="table2fn8">
              <p><sup>h</sup>DS-MIL: dual-stream multiple instance learning.</p>
            </fn>
            <fn id="table2fn9">
              <p><sup>i</sup>HIPT: hierarchical image pyramid transformer.</p>
            </fn>
            <fn id="table2fn10">
              <p><sup>j</sup>Italicization indicates the best result.</p>
            </fn>
            <fn id="table2fn11">
              <p><sup>k</sup><italic>P</italic>&lt;.05 vs attention-based multiple instance learning.</p>
            </fn>
            <fn id="table2fn12">
              <p><sup>l</sup><italic>P</italic>&lt;.05 vs attention induction.</p>
            </fn>
            <fn id="table2fn13">
              <p><sup>m</sup><italic>P</italic>&lt;.05 vs clustering-constrained attention multiple instance learning, single-branch.</p>
            </fn>
            <fn id="table2fn14">
              <p><sup>n</sup><italic>P</italic>&lt;.05 vs clustering-constrained attention multiple instance learning, multibranch.</p>
            </fn>
            <fn id="table2fn15">
              <p><sup>o</sup><italic>P</italic>&lt;.05 vs dual-stream multiple instance learning.</p>
            </fn>
            <fn id="table2fn16">
              <p><sup>p</sup><italic>P</italic>&lt;.05 vs hierarchical image pyramid transformer.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Data-Efficient Learning With Expert Annotation</title>
        <p>To investigate data-efficient training with expert annotation under low-data regimes reflecting real-world constraints, we assessed model performance on Camelyon16 by varying the proportion of training data (<xref rid="figure2" ref-type="fig">Figure 2</xref>). With only 20% of the slides, all methods achieved AUROC scores in the range of 0.4 to 0.6, and statistically significant differences were observed between rank induction and the other benchmarks. Among them, rank induction achieved the highest AUROC. At 40% and 60% of the training data size, rank induction significantly outperformed other MIL methods. When trained with 80% of the data, rank induction achieved an AUROC of 0.786 (SD 0.042), surpassing HIPT (AUROC=0.464, SD 0.076), DS-MIL (AUROC=0.709, SD 0.154), AB-MIL (AUROC=0.707, SD 0.186), attention induction (AUROC=0.714, SD 0.193), CLAM-SB (AUROC=0.721, SD 0.180), and CLAM-MB (AUROC=0.847, SD 0.042).</p>
        <fig id="figure2" position="float">
          <label>Figure 2</label>
          <caption>
            <p>Area under the receiver operating characteristic curve (AUROC) performance across different training set sampling ratios on Camelyon16. AB-MIL: attention-based multiple instance learning; DS-MIL: dual-stream multiple instance learning; CLAM-MB: clustering-constrained attention multiple instance learning, multibranch; CLAM-SB: clustering-constrained attention multiple instance learning, single-branch; HIPT: hierarchical image pyramid transformer. Asterisks indicate statistical significance based on the Mann-Whitney U test: *<italic>P</italic>&lt;.05, **<italic>P</italic>&lt;.001, ***<italic>P</italic>&lt;.0001, and ****<italic>P</italic>&lt;.00001.</p>
          </caption>
          <graphic xlink:href="medinform_v14i1e84417_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Annotation Granularity</title>
        <p>We evaluated model robustness to noisy or imprecise annotations by progressively padding the pathologist-drawn tumor regions on Camelyon16 to simulate coarse lesion markings. Both rank induction and attention induction were trained under each noise condition. With the original annotations (0-pixel padding), rank induction achieved an AUROC of 0.839 (SD 0.050). Even at the highest padding level, its performance declined only slightly to an AUROC of 0.823 (SD 0.050; <xref rid="figure3" ref-type="fig">Figure 3</xref>). In comparison, attention induction dropped from an AUROC of 0.742 (SD 0.142) to 0.723 (SD 0.134) as annotation noise increased. Across all levels of annotation coarseness, rank induction consistently outperformed attention induction, although the differences were not statistically significant at some padding levels.</p>
        <fig id="figure3" position="float">
          <label>Figure 3</label>
          <caption>
            <p>Annotation granularity experiment on Camelyon16. (A) Comparison of slide-level classification performance (area under the receiver operating characteristic curve [AUROC]) between attention induction and rank induction. (B) Representative visualization of the artificially expanded annotations used in the experiment. The thin inner red line indicates the original precise lesion annotation provided by the pathologist (ground truth), while the bold outer red line represents the artificially expanded annotation boundary simulating coarser supervision granularity.</p>
          </caption>
          <graphic xlink:href="medinform_v14i1e84417_fig3.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Fraction of Expert Annotation</title>
        <p>To simulate real-world scenarios in which annotating all WSIs is infeasible, we evaluated model performance by varying the proportion of annotated positive WSIs in the training set. As a baseline (0% annotation), AB-MIL achieved an average AUROC of 0.741 (SD 0.146; <xref rid="figure4" ref-type="fig">Figure 4</xref>). With only 10% (9/89) of positive WSIs annotated in the train dataset, rank induction boosted the performance to an average AUROC of 0.782 (SD 0.070). Increasing the annotation coverage to 20% (18/89) raised the AUROC to 0.807 (SD 0.070), while using all available positive WSIs (N=89) yielded a final AUROC of 0.839 (SD 0.050). Notably, rank induction achieved near-saturated performance with 20% of the annotated data, demonstrating high data efficiency. Moreover, the introduction of even minimal annotation significantly reduced performance variance compared to the unannotated baseline. The performance curve began to plateau beyond approximately 20% annotation coverage, indicating diminishing returns in AUROC gains, while maintaining low variability.</p>
        <fig id="figure4" position="float">
          <label>Figure 4</label>
          <caption>
            <p>Impact of annotation fraction on area under the receiver operating characteristic curve (AUROC). WSI: whole slide image.</p>
          </caption>
          <graphic xlink:href="medinform_v14i1e84417_fig4.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Ablation Test</title>
        <p>The impact of varying the attention threshold <italic>T</italic> on model performance and attention distribution is presented in <xref ref-type="table" rid="table3">Table 3</xref>. As <italic>T</italic> increased from 0 to 2, both AUROC and AUPRC improved steadily, accompanied by a marked reduction in the average sum of attention assigned to nonlesion patches.</p>
        <table-wrap position="float" id="table3">
          <label>Table 3</label>
          <caption>
            <p>Effect of attention threshold <italic>T</italic> on whole slide image (WSI) classification performance and sum of attention weight for nonlesion patches.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="360"/>
            <col width="170"/>
            <col width="180"/>
            <col width="290"/>
            <thead>
              <tr valign="top">
                <td>Threshold</td>
                <td colspan="2">Test dataset (n=129)</td>
                <td>Positive WSI in the test dataset (n=49; sum of nonlesion attention), mean (SD)</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>AUROC<sup>a</sup></td>
                <td>AUPRC<sup>b</sup></td>
                <td>
                  <break/>
                </td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>0</td>
                <td>0.830</td>
                <td>0.846</td>
                <td>0.465 (0.420)</td>
              </tr>
              <tr valign="top">
                <td>0.25</td>
                <td>0.853</td>
                <td>0.878</td>
                <td>0.411 (0.434)</td>
              </tr>
              <tr valign="top">
                <td>0.5</td>
                <td>0.869</td>
                <td>0.881</td>
                <td>0.362 (0.423)</td>
              </tr>
              <tr valign="top">
                <td>1</td>
                <td>0.894</td>
                <td>0.905</td>
                <td>0.269 (0.387)</td>
              </tr>
              <tr valign="top">
                <td>2</td>
                <td>0.916</td>
                <td>0.922</td>
                <td>0.178 (0.354)</td>
              </tr>
              <tr valign="top">
                <td>5</td>
                <td>0.909</td>
                <td>0.914</td>
                <td>0.105 (0.339)</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table3fn1">
              <p><sup>a</sup>AUROC: area under the receiver operating characteristic curve.</p>
            </fn>
            <fn id="table3fn2">
              <p><sup>b</sup>AUPRC: area under the precision-recall curve.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <p>With no thresholding (<italic>T</italic>=0) or low thresholds (<italic>T</italic>&lt;1, where attention distributions remain nearly uniform), the AUROC remained relatively low. This suggests that weakly attended nonlesion patches diluted the slide-level representation. The best performance was observed at <italic>T</italic>=2 (AUROC=0.916; AUPRC=0.922), indicating that moderate thresholding effectively suppresses noise-like, low-value attention weights while preserving discriminative focus. Excessively high thresholds (eg, <italic>T</italic>=5) yielded a slight decline, reflecting an optimal balance near <italic>T</italic>=2.</p>
      </sec>
      <sec>
        <title>Attention Localization and Qualitative Analysis</title>
        <p>To evaluate the effect of rank induction on attention localization, we examined whether the attention weights were concentrated within expert-annotated regions. Across all 3 localization metrics, rank induction consistently demonstrated greater focus on the annotated areas compared to other MIL methods. At α=5, AB-MIL, attention induction, and rank induction achieved IoU scores of 0.029, 0.023, and 0.031 and Dice scores of 0.056, 0.045, and 0.059, respectively. When α increased to 75, the performance gap between rank induction and the baseline models became more pronounced (<xref rid="figure5" ref-type="fig">Figures 5</xref>A and 5B). A similar trend was observed in pointing game accuracy (<xref rid="figure5" ref-type="fig">Figure 5</xref>C). Qualitatively, both AB-MIL and attention induction produced noisy attention maps that highlighted not only tumor regions but also large areas of normal tissue, resulting in frequent false positives. In contrast, rank induction generated sharp and well-localized attention maps, accurately delineating lesion areas with minimal off-target activation—closely matching the ground truth annotations. This distinct separation between tumor and normal regions explains rank induction’s superior performance and enhances the interpretability of its predictions (<xref rid="figure6" ref-type="fig">Figure 6</xref>).</p>
        <fig id="figure5" position="float">
          <label>Figure 5</label>
          <caption>
            <p>Attention localization performance of attention-based multiple instance learning (AB-MIL), attention induction, and rank induction. (A) Intersection over union (IoU) score according to the top α% of patches. (B) Dice score according to the top α% of patches. (C) Average pointing game score. WSI: whole slide image.</p>
          </caption>
          <graphic xlink:href="medinform_v14i1e84417_fig5.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure6" position="float">
          <label>Figure 6</label>
          <caption>
            <p>Qualitative visualization of patch selection for attention-based multiple instance learning (AB-MIL), attention induction, and rank induction. (A) A positive WSI thumbnail. (B) Attention distributions from each model in the lesion (red box) and nonlesion (blue box) regions in (A). (C) Another positive WSI thumbnail. (D) Attention distributions from each model in the lesion and nonlesion regions in (C).</p>
          </caption>
          <graphic xlink:href="medinform_v14i1e84417_fig6.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <sec>
        <title>Principal Findings</title>
        <p>In this work, we present rank induction, the first method to integrate MIL with expert annotations using pairwise rank constraints. By enforcing a ranking loss that encourages higher attention scores for annotated patches than for nonannotated ones, our approach effectively strengthens the weak supervision signal inherent to MIL frameworks.</p>
        <p>We validated rank induction through comprehensive experiments on 2 public datasets and 1 in-house dataset, covering 5 key scenarios: full-data classification, low-data regimes, varying annotation granularity, subset annotation settings, and qualitative visualization. Across all settings, rank induction consistently outperformed baseline methods, including pseudo-labeling approaches and attention induction, while producing more faithful and interpretable attention heat maps.</p>
        <p>Notably, the proposed ranking mechanism is model-agnostic and can be used as a versatile module within other MIL architectures. As demonstrated in <xref ref-type="supplementary-material" rid="app3">Multimedia Appendix 3</xref>, integrating rank induction into existing methods such as DS-MIL and HIPT consistently improved their performance. In our main experiments, these baseline models (DS-MIL and HIPT) exhibited lower performance than that reported in their original studies. This discrepancy arises because we used a standard ImageNet-pretrained ResNet-50 backbone across all models to ensure a fair comparison of the MIL aggregation strategies themselves, rather than relying on the self-supervised pretraining typically required by those methods. However, even under this constrained setting, incorporating rank induction significantly boosted their classification accuracy, demonstrating the robustness and generalizability of our ranking-based supervision.</p>
        <p>Ultimately, these performance improvements stem from bridging weak and strong supervision through informative pairwise constraints that approximate expert guidance. Rank induction is grounded in a simple yet clinically intuitive assumption: models should prioritize annotated regions over nonannotated ones. This inductive bias aligns naturally with domain knowledge in pathology, establishing rank induction as an effective and practical solution for real-world applications.</p>
      </sec>
      <sec>
        <title>Real-World Application With Expert Annotations</title>
        <p>The pairwise ranking approach in rank induction addresses key challenges associated with expert annotations in real-world settings. In practice, tumor regions often exhibit irregular and heterogeneous morphology with poorly defined boundaries, making precise annotations difficult. Consequently, annotations are frequently coarse or incomplete. Despite this, rank induction maintained robust classification performance even with noisy annotations (<xref rid="figure3" ref-type="fig">Figure 3</xref>). This resilience likely stems from the pairwise training mechanism: although some comparisons may involve incorrectly labeled patches, most pairwise relationships remain valid, enabling the model to suppress noise and learn effectively.</p>
        <p>Moreover, due to time and resource limitations, only a subset of positive WSIs can typically be annotated in clinical workflows. In such scenarios, extracting maximum value from limited expert input is critical. Unlike attention induction, which relies on assigning exact attention weights to individual patches, rank induction leverages a larger number of patch-pair comparisons from the valid pair set <italic>P</italic>, which grows quadratically with the number of annotated patches. This design allows the model to achieve near-saturated performance even with partial annotations (<xref rid="figure4" ref-type="fig">Figure 4</xref>).</p>
        <p>Although pseudolabeling offers an alternative when expert annotations are unavailable, as seen in methods such as CLAM, it introduces uncertainty by relying on the model’s own predictions. These pseudolabels are inherently noisier than expert supervision and can lead to cascading errors. In contrast, rank induction avoids this risk by grounding supervision in verified expert annotations, leading to more stable and accurate slide-level predictions, especially in low-data or noisy-label conditions.</p>
      </sec>
      <sec>
        <title>Comparison of Fully Supervised Learning and Rank Induction</title>
        <p>In scenarios where exhaustive patch-level annotations are available for all WSIs, fully supervised learning remains a strong baseline. Using the same training settings as our MIL models, a patch-level convolutional neural network with a fine-tuned ResNet-50 backbone and max-pooling aggregation achieved an AUROC of 0.892 on Camelyon16. This result outperformed the AUROC of 0.839 achieved by rank induction, which utilizes a frozen backbone and MIL-based aggregation. This result illustrates the upper bound of fully supervised performance given complete expert annotations and domain-specific fine-tuning (<xref ref-type="supplementary-material" rid="app4">Multimedia Appendix 4</xref>).</p>
        <p>However, in realistic clinical environments, exhaustive annotation is rarely feasible. When only 10% (9/89) of positive WSIs were annotated in the train dataset, the fully supervised model trained on this subset achieved an AUROC of 0.729. In contrast, rank induction, using the same 10% of patch-level annotations and leveraging slide-level labels from the remaining 90%, achieved a higher AUROC of 0.775. The key difference lies in how each method handles unannotated slides. Fully supervised models are limited to the annotated subset and cannot benefit from weak labels. Rank induction, by integrating both patch- and slide-level supervision through pairwise ranking constraints, makes effective use of all available data, even with sparse annotations.</p>
        <p>These findings highlight a clear trade-off: while fully supervised learning performs best under ideal, fully annotated conditions, rank induction offers a more practical and annotation-efficient alternative. Its ability to generalize from partial supervision makes it especially suitable for real-world digital pathology applications, where annotations are incomplete or expensive.</p>
      </sec>
      <sec>
        <title>Limitations</title>
        <p>Our study has several limitations. First, although we validated our method on 3 datasets, including 1 in-house dataset (SMF-stomach), the diversity of organs and pathologies remains limited. Specifically, the SMF-stomach dataset was collected from a single institution, which may introduce biases related to specific staining protocols or scanner characteristics. To fully assess cross-domain robustness, future work should include evaluation on larger, multicenter cohorts with diverse scanners. Nonetheless, the inclusion of Camelyon16 partially addresses this concern, as it contains slides from multiple medical institutions, providing evidence of generalization.</p>
      </sec>
      <sec>
        <title>Conclusions</title>
        <p>In summary, rank induction provides a practical and effective solution for improving slide-level classification in digital pathology by bridging weak and strong supervision via pairwise ranking constraints. By leveraging only sparse expert annotations, it achieves a strong balance between annotation efficiency and model performance, outperforming traditional MIL and pseudolabeling approaches, particularly in low-data or noisy-label settings.</p>
        <p>Although fully supervised learning may yield higher accuracy when exhaustive annotations are available, rank induction offers a more scalable and annotation-efficient alternative for real-world use. Its robustness with limited annotations and its ability to highlight diagnostically relevant regions underscore the potential of rank-based supervision to improve both model interpretability and generalizability across diverse clinical environments.</p>
      </sec>
    </sec>
  </body>
  <back>
    <app-group>
      <supplementary-material id="app1">
        <label>Multimedia Appendix 1</label>
        <p>Example of a partially annotated whole slide image from the Seegene Medical Foundation-stomach dataset.</p>
        <media xlink:href="medinform_v14i1e84417_app1.png" xlink:title="PNG File , 511 KB"/>
      </supplementary-material>
      <supplementary-material id="app2">
        <label>Multimedia Appendix 2</label>
        <p>Comparison of slide-level classification accuracy across different multiple instance learning methods on 3 datasets.</p>
        <media xlink:href="medinform_v14i1e84417_app2.docx" xlink:title="DOCX File , 30 KB"/>
      </supplementary-material>
      <supplementary-material id="app3">
        <label>Multimedia Appendix 3</label>
        <p>Impact of integrating rank induction into existing multiple instance learning architectures (dual-stream multiple instance learning and hierarchical image pyramid transformer).</p>
        <media xlink:href="medinform_v14i1e84417_app3.docx" xlink:title="DOCX File , 31 KB"/>
      </supplementary-material>
      <supplementary-material id="app4">
        <label>Multimedia Appendix 4</label>
        <p>Performance comparison between fully supervised learning and rank induction under varying proportions of annotated positive whole slide images.</p>
        <media xlink:href="medinform_v14i1e84417_app4.png" xlink:title="PNG File , 41 KB"/>
      </supplementary-material>
    </app-group>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">AB-MIL</term>
          <def>
            <p>attention-based multiple instance learning</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">AUPRC</term>
          <def>
            <p>area under the precision-recall curve</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">AUROC</term>
          <def>
            <p>area under the receiver operating characteristic curve</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">CLAM</term>
          <def>
            <p>clustering-constrained attention multiple instance learning</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">CLAM-MB</term>
          <def>
            <p>clustering-constrained attention multiple instance learning, multibranch</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">CLAM-SB</term>
          <def>
            <p>clustering-constrained attention multiple instance learning, single-branch</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb7">DS-MIL</term>
          <def>
            <p>dual-stream multiple instance learning</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb8">HIPT</term>
          <def>
            <p>hierarchical image pyramid transformer</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb9">IoU</term>
          <def>
            <p>intersection over union</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb10">IRB</term>
          <def>
            <p>institutional review board</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb11">MIL</term>
          <def>
            <p>multiple instance learning</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb12">SMF-stomach</term>
          <def>
            <p>Seegene Medical Foundation-stomach</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb13">WSI</term>
          <def>
            <p>whole slide image</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <fn-group>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Song</surname>
              <given-names>AH</given-names>
            </name>
            <name name-style="western">
              <surname>Jaume</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Williamson</surname>
              <given-names>DF</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>MY</given-names>
            </name>
            <name name-style="western">
              <surname>Vaidya</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Miller</surname>
              <given-names>TR</given-names>
            </name>
            <name name-style="western">
              <surname>Mahmood</surname>
              <given-names>F</given-names>
            </name>
          </person-group>
          <article-title>Artificial intelligence for digital and computational pathology</article-title>
          <source>Nat Rev Bioeng</source>
          <year>2023</year>
          <month>10</month>
          <day>02</day>
          <volume>1</volume>
          <issue>12</issue>
          <fpage>930</fpage>
          <lpage>49</lpage>
          <pub-id pub-id-type="doi">10.1038/s44222-023-00096-8</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Saltz</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Gupta</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Hou</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Kurc</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Singh</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Nguyen</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Samaras</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Shroyer</surname>
              <given-names>KR</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Batiste</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Van Arnam</surname>
              <given-names>J</given-names>
            </name>
            <collab>Cancer Genome Atlas Research Network</collab>
            <name name-style="western">
              <surname>Shmulevich</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Rao</surname>
              <given-names>AU</given-names>
            </name>
            <name name-style="western">
              <surname>Lazar</surname>
              <given-names>AJ</given-names>
            </name>
            <name name-style="western">
              <surname>Sharma</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Thorsson</surname>
              <given-names>V</given-names>
            </name>
          </person-group>
          <article-title>Spatial organization and molecular correlation of tumor-infiltrating lymphocytes using deep learning on pathology images</article-title>
          <source>Cell Rep</source>
          <year>2018</year>
          <month>04</month>
          <day>03</day>
          <volume>23</volume>
          <issue>1</issue>
          <fpage>181</fpage>
          <lpage>93.e7</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://air.unimi.it/handle/2434/586507"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.celrep.2018.03.086</pub-id>
          <pub-id pub-id-type="medline">29617659</pub-id>
          <pub-id pub-id-type="pii">S2211-1247(18)30447-9</pub-id>
          <pub-id pub-id-type="pmcid">PMC5943714</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>van der Laak</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Litjens</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Ciompi</surname>
              <given-names>F</given-names>
            </name>
          </person-group>
          <article-title>Deep learning in histopathology: the path to the clinic</article-title>
          <source>Nat Med</source>
          <year>2021</year>
          <month>05</month>
          <day>14</day>
          <volume>27</volume>
          <issue>5</issue>
          <fpage>775</fpage>
          <lpage>84</lpage>
          <pub-id pub-id-type="doi">10.1038/s41591-021-01343-4</pub-id>
          <pub-id pub-id-type="medline">33990804</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41591-021-01343-4</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Campanella</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Hanna</surname>
              <given-names>MG</given-names>
            </name>
            <name name-style="western">
              <surname>Geneslaw</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Miraflor</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Werneck Krauss Silva</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Busam</surname>
              <given-names>KJ</given-names>
            </name>
            <name name-style="western">
              <surname>Brogi</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Reuter</surname>
              <given-names>VE</given-names>
            </name>
            <name name-style="western">
              <surname>Klimstra</surname>
              <given-names>DS</given-names>
            </name>
            <name name-style="western">
              <surname>Fuchs</surname>
              <given-names>TJ</given-names>
            </name>
          </person-group>
          <article-title>Clinical-grade computational pathology using weakly supervised deep learning on whole slide images</article-title>
          <source>Nat Med</source>
          <year>2019</year>
          <month>08</month>
          <day>15</day>
          <volume>25</volume>
          <issue>8</issue>
          <fpage>1301</fpage>
          <lpage>9</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/31308507"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41591-019-0508-1</pub-id>
          <pub-id pub-id-type="medline">31308507</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41591-019-0508-1</pub-id>
          <pub-id pub-id-type="pmcid">PMC7418463</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Meng</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Qiao</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Coupland</surname>
              <given-names>SE</given-names>
            </name>
            <name name-style="western">
              <surname>Zheng</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>DTFD-MIL: double-tier feature distillation multiple instance learning for histopathology whole slide image classification</article-title>
          <source>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source>
          <year>2022</year>
          <conf-name>CVPR 2022</conf-name>
          <conf-date>June 18-24, 2022</conf-date>
          <conf-loc>New Orleans, LA</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://ieeexplore.ieee.org/document/9880112?denied="/>
          </comment>
          <pub-id pub-id-type="doi">10.1109/cvpr52688.2022.01824</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhou</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Multiple instance learning with critical instance for whole slide image classification</article-title>
          <source>Proceedings of the IEEE 20th International Symposium on Biomedical Imaging</source>
          <year>2023</year>
          <conf-name>ISBI 2023</conf-name>
          <conf-date>April 18-21, 2023</conf-date>
          <conf-loc>Cartagena, Colombia</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://ieeexplore.ieee.org/abstract/document/10230595/"/>
          </comment>
          <pub-id pub-id-type="doi">10.1109/isbi53787.2023.10230595</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhou</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Deep hierarchical multiple instance learning for whole slide image classification</article-title>
          <source>Proceedings of the IEEE 19th International Symposium on Biomedical Imaging</source>
          <year>2022</year>
          <conf-name>ISBI 2022</conf-name>
          <conf-date>March 28-31, 2022</conf-date>
          <conf-loc>Kolkata, India</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://ieeexplore.ieee.org/abstract/document/9761678/"/>
          </comment>
          <pub-id pub-id-type="doi">10.1109/isbi52829.2022.9761678</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Qu</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Luo</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Guo</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Song</surname>
              <given-names>Z</given-names>
            </name>
          </person-group>
          <article-title>Weakly semi-supervised whole slide image classification by two-level cross consistency supervision</article-title>
          <source>ArXiv. Preprint posted online on April 16, 2025</source>
          <year>2026</year>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/2504.12132"/>
          </comment>
          <pub-id pub-id-type="doi">10.48550/arXiv.2504.12132</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ouyang</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Fu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Yan</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Shi</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Ling</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>He</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Guan</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>MergeUp-augmented semi-weakly supervised learning for WSI classification</article-title>
          <source>ArXiv. Preprint posted online on August 23, 2024</source>
          <year>2026</year>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/2408.12825"/>
          </comment>
          <pub-id pub-id-type="doi">10.48550/arXiv.2408.12825</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>MY</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>RJ</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Dillon</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Mahmood</surname>
              <given-names>F</given-names>
            </name>
          </person-group>
          <article-title>Semi-supervised histology classification using deep multiple instance learning and contrastive predictive coding</article-title>
          <source>ArXiv. Preprint posted online on October 23, 2019</source>
          <year>2026</year>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/1910.10825"/>
          </comment>
          <pub-id pub-id-type="doi">10.48550/arXiv.1910.10825</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Li</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Eliceiri</surname>
              <given-names>KW</given-names>
            </name>
          </person-group>
          <article-title>Dual-stream multiple instance learning network for whole slide image classification with self-supervised contrastive learning</article-title>
          <source>ArXiv. Preprint posted online on November 17, 2020</source>
          <year>2026</year>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/2011.08939"/>
          </comment>
          <pub-id pub-id-type="doi">10.48550/arXiv.2011.08939</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>RJ</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>TY</given-names>
            </name>
            <name name-style="western">
              <surname>Trister</surname>
              <given-names>AD</given-names>
            </name>
            <name name-style="western">
              <surname>Krishnan</surname>
              <given-names>RG</given-names>
            </name>
            <name name-style="western">
              <surname>Mahmood</surname>
              <given-names>F</given-names>
            </name>
          </person-group>
          <article-title>Scaling vision transformers to gigapixel images via hierarchical self-supervised learning</article-title>
          <source>ArXiv. Preprint posted online on June 6, 2022</source>
          <year>2026</year>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/2206.02647"/>
          </comment>
          <pub-id pub-id-type="doi">10.1109/cvpr52688.2022.01567</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>MY</given-names>
            </name>
            <name name-style="western">
              <surname>Williamson</surname>
              <given-names>DF</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>TY</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>RJ</given-names>
            </name>
            <name name-style="western">
              <surname>Barbieri</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Mahmood</surname>
              <given-names>F</given-names>
            </name>
          </person-group>
          <article-title>Data-efficient and weakly supervised computational pathology on whole-slide images</article-title>
          <source>Nat Biomed Eng</source>
          <year>2021</year>
          <month>06</month>
          <day>01</day>
          <volume>5</volume>
          <issue>6</issue>
          <fpage>555</fpage>
          <lpage>70</lpage>
          <pub-id pub-id-type="doi">10.1038/s41551-020-00682-w</pub-id>
          <pub-id pub-id-type="medline">33649564</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41551-020-00682-w</pub-id>
          <pub-id pub-id-type="pmcid">PMC8711640</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhou</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Che</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Yan</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Wei</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Ding</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Iterative multiple instance learning for weakly annotated whole slide image classification</article-title>
          <source>Phys Med Biol</source>
          <year>2023</year>
          <month>07</month>
          <day>19</day>
          <volume>68</volume>
          <issue>15</issue>
          <pub-id pub-id-type="doi">10.1088/1361-6560/acde3f</pub-id>
          <pub-id pub-id-type="medline">37311470</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhou</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Multiple instance learning with task-specific multi-level features for weakly annotated histopathological image classification</article-title>
          <source>Proceedings of the 2022 IEEE International Conference on Acoustics, Speech and Signal Processing</source>
          <year>2022</year>
          <conf-name>ICASSP 2022</conf-name>
          <conf-date>May 23-27, 2022</conf-date>
          <conf-loc>Singapore, Singapore</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://ieeexplore.ieee.org/abstract/document/9747121/"/>
          </comment>
          <pub-id pub-id-type="doi">10.1109/icassp43922.2022.9747121</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Myronenko</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Roth</surname>
              <given-names>HR</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <person-group person-group-type="editor">
            <name name-style="western">
              <surname>de Bruijne</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Cattin</surname>
              <given-names>PC</given-names>
            </name>
            <name name-style="western">
              <surname>Cotin</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Padoy</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Speidel</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Zheng</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Essert</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Accounting for Dependencies in Deep Learning Based Multiple Instance Learning for Whole Slide Imaging</article-title>
          <source>Medical Image Computing and Computer Assisted Intervention – MICCAI 2021</source>
          <year>2021</year>
          <publisher-loc>Cham, Switzerland</publisher-loc>
          <publisher-name>Springer International Publishing</publisher-name>
          <fpage>329</fpage>
          <lpage>38</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lerousseau</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Vakalopoulou</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Classe</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Adam</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Battistella</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Carré</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Estienne</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Henry</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Deutsch</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Paragios</surname>
              <given-names>N</given-names>
            </name>
          </person-group>
          <person-group person-group-type="editor">
            <name name-style="western">
              <surname>Martel</surname>
              <given-names>AL</given-names>
            </name>
            <name name-style="western">
              <surname>Abolmaesum</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Stoyanov</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Mateus</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Zuluaga</surname>
              <given-names>MA</given-names>
            </name>
            <name name-style="western">
              <surname>Zhou</surname>
              <given-names>SK</given-names>
            </name>
            <name name-style="western">
              <surname>Racoceanu</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Joskowicz</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>Weakly supervised multiple instance learning histopathological tumor segmentation</article-title>
          <source>Medical Image Computing and Computer Assisted Intervention – MICCAI 2020</source>
          <year>2020</year>
          <publisher-loc>Cham, Switzerland</publisher-loc>
          <publisher-name>Springer International Publishing</publisher-name>
          <fpage>470</fpage>
          <lpage>9</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Qu</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Luo</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Song</surname>
              <given-names>Z</given-names>
            </name>
          </person-group>
          <person-group person-group-type="editor">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Dou</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Fletcher</surname>
              <given-names>PT</given-names>
            </name>
            <name name-style="western">
              <surname>Speidel</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>DGMIL: distribution guided multiple instance learning for whole slide image classification</article-title>
          <source>Medical Image Computing and Computer Assisted Intervention – MICCAI 2022</source>
          <year>2022</year>
          <publisher-loc>Cham, Switzerland</publisher-loc>
          <publisher-name>Springer</publisher-name>
          <fpage>24</fpage>
          <lpage>34</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Koga</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Yokota</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Arihiro</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Hontani</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>Attention induction based on pathologist annotations for improving whole slide pathology image classifier</article-title>
          <source>J Pathol Inform</source>
          <year>2025</year>
          <month>01</month>
          <volume>16</volume>
          <fpage>100413</fpage>
          <pub-id pub-id-type="doi">10.1016/j.jpi.2024.100413</pub-id>
          <pub-id pub-id-type="medline">39845979</pub-id>
          <pub-id pub-id-type="pii">S2153-3539(24)00052-X</pub-id>
          <pub-id pub-id-type="pmcid">PMC11750489</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Burges</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>From RankNet to LambdaRank to LambdaMART: an overview</article-title>
          <source>Microsoft Research</source>
          <year>2010</year>
          <access-date>2026-01-11</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.microsoft.com/en-us/research/wp-content/uploads/2016/02/MSR-TR-2010-82.pdf">https://www.microsoft.com/en-us/research/wp-content/uploads/2016/02/MSR-TR-2010-82.pdf</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ehteshami Bejnordi</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Veta</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Johannes van Diest</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>van Ginneken</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Karssemeijer</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Litjens</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>van der Laak</surname>
              <given-names>JA</given-names>
            </name>
            <collab>the CAMELYON16 Consortium</collab>
            <name name-style="western">
              <surname>Hermsen</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Manson</surname>
              <given-names>QF</given-names>
            </name>
            <name name-style="western">
              <surname>Balkenhol</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Geessink</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Stathonikos</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>van Dijk</surname>
              <given-names>MC</given-names>
            </name>
            <name name-style="western">
              <surname>Bult</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Beca</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Beck</surname>
              <given-names>AH</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Khosla</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Gargeya</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Irshad</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Zhong</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Dou</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Heng</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Haß</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Bruni</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Wong</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Halici</surname>
              <given-names>U</given-names>
            </name>
            <name name-style="western">
              <surname>Öner</surname>
              <given-names>MÜ</given-names>
            </name>
            <name name-style="western">
              <surname>Cetin-Atalay</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Berseth</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Khvatkov</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Vylegzhanin</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Kraus</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Shaban</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Rajpoot</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Awan</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Sirinukunwattana</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Qaiser</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Tsang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Tellez</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Annuscheit</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Hufnagl</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Valkonen</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Kartasalo</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Latonen</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Ruusuvuori</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Liimatainen</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Albarqouni</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Mungal</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>George</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Demirci</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Navab</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Watanabe</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Seno</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Takenaka</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Matsuda</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Ahmady Phoulady</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Kovalev</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Kalinovsky</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Liauchuk</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Bueno</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Fernandez-Carrobles</surname>
              <given-names>MM</given-names>
            </name>
            <name name-style="western">
              <surname>Serrano</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Deniz</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Racoceanu</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Venâncio</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Diagnostic assessment of deep learning algorithms for detection of lymph node metastases in women with breast cancer</article-title>
          <source>JAMA</source>
          <year>2017</year>
          <month>12</month>
          <day>12</day>
          <volume>318</volume>
          <issue>22</issue>
          <fpage>2199</fpage>
          <lpage>210</lpage>
          <pub-id pub-id-type="doi">10.1001/jama.2017.14585</pub-id>
          <pub-id pub-id-type="medline">29234806</pub-id>
          <pub-id pub-id-type="pii">2665774</pub-id>
          <pub-id pub-id-type="pmcid">PMC5820737</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Da</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Huang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Zuo</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Hu</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Yi</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Guo</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>He</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Mei</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Shen</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Shi</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Krishnamurthi</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Song</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Graham</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Bashir</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Qin</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Tian</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Yin</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Metaxas</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>DigestPath: a benchmark dataset with challenge review for the pathological detection and segmentation of digestive-system</article-title>
          <source>Med Image Anal</source>
          <year>2022</year>
          <month>08</month>
          <volume>80</volume>
          <fpage>102485</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1016/j.media.20102485"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.media.2022.102485</pub-id>
          <pub-id pub-id-type="medline">35679692</pub-id>
          <pub-id pub-id-type="pii">S1361-8415(22)00132-3</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>HH</given-names>
            </name>
            <name name-style="western">
              <surname>Ko</surname>
              <given-names>YS</given-names>
            </name>
            <name name-style="western">
              <surname>Jeong</surname>
              <given-names>WC</given-names>
            </name>
            <name name-style="western">
              <surname>Yun</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Evaluating the robustness of slide-level AI predictions on out-of-focus whole slide images: a retrospective observational study</article-title>
          <source>J Pathol Inform</source>
          <year>2025</year>
          <month>11</month>
          <volume>19</volume>
          <fpage>100518</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S2153-3539(25)00104-X"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.jpi.2025.100518</pub-id>
          <pub-id pub-id-type="medline">41146682</pub-id>
          <pub-id pub-id-type="pii">S2153-3539(25)00104-X</pub-id>
          <pub-id pub-id-type="pmcid">PMC12554180</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ko</surname>
              <given-names>YS</given-names>
            </name>
            <name name-style="western">
              <surname>Choi</surname>
              <given-names>YM</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Park</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Ashraf</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Quiñones Robles</surname>
              <given-names>Willmer Rafell</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>MJ</given-names>
            </name>
            <name name-style="western">
              <surname>Jang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Yun</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Hwang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Jang</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Yi</surname>
              <given-names>MY</given-names>
            </name>
          </person-group>
          <article-title>Improving quality control in the routine practice for histopathological interpretation of gastrointestinal endoscopic biopsies using artificial intelligence</article-title>
          <source>PLoS One</source>
          <year>2022</year>
          <volume>17</volume>
          <issue>12</issue>
          <fpage>e0278542</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://dx.plos.org/10.1371/journal.pone.0278542"/>
          </comment>
          <pub-id pub-id-type="doi">10.1371/journal.pone.0278542</pub-id>
          <pub-id pub-id-type="medline">36520777</pub-id>
          <pub-id pub-id-type="pii">PONE-D-22-17965</pub-id>
          <pub-id pub-id-type="pmcid">PMC9754254</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Bargal</surname>
              <given-names>SA</given-names>
            </name>
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Brandt</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Shen</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Sclaroff</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Top-down neural attention by excitation backprop</article-title>
          <source>Int J Comput Vis</source>
          <year>2017</year>
          <month>12</month>
          <day>23</day>
          <volume>126</volume>
          <fpage>1084</fpage>
          <lpage>102</lpage>
          <pub-id pub-id-type="doi">10.1007/s11263-017-1059-x</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="web">
          <article-title>JMIR: Official repository for "Ranking-Aware Multiple Instance Learning for Histopathology Slide Classification: Development and Validation Study"</article-title>
          <source>GitHub</source>
          <access-date>2026-01-20</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://github.com/SMF-AI/rank_induction_JMIR">https://github.com/SMF-AI/rank_induction_JMIR</ext-link>
          </comment>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
