<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Med Inform</journal-id><journal-id journal-id-type="publisher-id">medinform</journal-id><journal-id journal-id-type="index">7</journal-id><journal-title>JMIR Medical Informatics</journal-title><abbrev-journal-title>JMIR Med Inform</abbrev-journal-title><issn pub-type="epub">2291-9694</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v14i1e73311</article-id><article-id pub-id-type="doi">10.2196/73311</article-id><article-categories><subj-group subj-group-type="heading"><subject>Original Paper</subject></subj-group></article-categories><title-group><article-title>Automatic Recognition and Prognostic Prediction of Colorectal Liver Metastases Using a Multi-Scale Deep Learning Framework: Model Development and Validation Study</article-title></title-group><contrib-group><contrib contrib-type="author"><name name-style="western"><surname>Long</surname><given-names>Houwen</given-names></name><degrees>BSc</degrees><xref ref-type="aff" rid="aff1"/></contrib><contrib contrib-type="author"><name name-style="western"><surname>Ying</surname><given-names>Feng</given-names></name><degrees>BSc</degrees><xref ref-type="aff" rid="aff1"/></contrib><contrib contrib-type="author"><name name-style="western"><surname>Wu</surname><given-names>Shi</given-names></name><degrees>MSc</degrees><xref ref-type="aff" rid="aff1"/></contrib><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Li</surname><given-names>Long</given-names></name><degrees>MSc</degrees><xref ref-type="aff" rid="aff1"/></contrib></contrib-group><aff id="aff1"><institution>Department of Anus and Intestine Surgery, Yongkang First People&#x2019;s Hospital Affiliated to Hangzhou Medical College</institution><addr-line>599 Jinshan West Road, Dongcheng Street</addr-line><addr-line>Yongkang</addr-line><country>China</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Benis</surname><given-names>Arriel</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Khan</surname><given-names>Md Tabrej</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Sivarajkumar</surname><given-names>Sonish</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Long Li, MSc, Department of Anus and Intestine Surgery, Yongkang First People&#x2019;s Hospital Affiliated to Hangzhou Medical College, 599 Jinshan West Road, Dongcheng Street, Yongkang, 321300, China, 86 0579-89279021; <email>zjykll1988@163.com</email></corresp></author-notes><pub-date pub-type="collection"><year>2026</year></pub-date><pub-date pub-type="epub"><day>7</day><month>4</month><year>2026</year></pub-date><volume>14</volume><elocation-id>e73311</elocation-id><history><date date-type="received"><day>01</day><month>03</month><year>2025</year></date><date date-type="rev-recd"><day>18</day><month>09</month><year>2025</year></date><date date-type="accepted"><day>05</day><month>10</month><year>2025</year></date></history><copyright-statement>&#x00A9; Houwen Long, Feng Ying, Shi Wu, Long Li. Originally published in JMIR Medical Informatics (<ext-link ext-link-type="uri" xlink:href="https://medinform.jmir.org">https://medinform.jmir.org</ext-link>), 7.4.2026. </copyright-statement><copyright-year>2026</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://medinform.jmir.org/">https://medinform.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://medinform.jmir.org/2026/1/e73311"/><abstract><sec><title>Background</title><p>Colorectal cancer liver metastasis (CRLM) presents considerable challenges in both diagnosis and prognosis, as conventional approaches often are limited by subjectivity, variability, and limited efficiency. Recent advances in deep learning have shown great potential for automated extraction of pathological features, offering improved diagnostic accuracy and more reliable prognostic predictions.</p></sec><sec><title>Objective</title><p>This study aimed to develop and validate a multi-model ensemble deep learning framework (colon cancer liver metastasis network [CLM-Net]) for the automatic recognition and prognostic prediction of CRLM from pathological images, thereby enhancing diagnostic accuracy and clinical applicability.</p></sec><sec sec-type="methods"><title>Methods</title><p>A total of 197 pathologically annotated CRLM cases were collected and integrated from publicly available datasets (Kaggle and The Cancer Imaging Archive) to construct high-quality training and independent test sets. CLM-Net was built upon base convolutional neural network architectures including VGG16, DeepLab-v3, and U-Net, incorporating multi-scale atrous convolutions, a squeeze-and-excitation attention mechanism, a conditional random field refinement module, and transfer learning strategies. The framework was comprehensively evaluated using 5-fold cross-validation and independent testing for classification, segmentation, and prognostic prediction tasks. For survival prediction, 1024-dimensional image feature vectors extracted by CLM-Net were analyzed using logistic regression and random forest classifiers, with Kaplan-Meier curves and log-rank tests used for survival analysis.</p></sec><sec sec-type="results"><title>Results</title><p>CLM-Net demonstrated superior performance in pathological image recognition, achieving 94% accuracy, 92% recall, 93% <italic>F</italic><sub>1</sub>-score, and an area under the receiver operating characteristic curve of 0.96 on the independent test set, outperforming single models. For survival prediction, CLM-Net with multi-scale attention achieved an area under the receiver operating characteristic curve of 0.864. Kaplan-Meier analysis revealed its significantly stronger risk stratification ability compared with VGG16, U-Net, and DeepLab-v3. Precision-recall curves and heatmaps further confirmed the model&#x2019;s high robustness and generalizability in unseen samples. Clinical evaluation by pathologists indicated strong interpretability and diagnostic utility, with a concordance rate of 90%.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>The proposed CLM-Net framework, through the integration of diverse deep learning architectures and mechanisms, markedly improved the recognition and prognostic prediction of CRLM, demonstrating excellent generalization and clinical translation potential. Future studies will focus on multi-center validation and integration of multimodal features to further optimize its role in precision medicine.</p></sec></abstract><kwd-group><kwd>colorectal liver metastasis</kwd><kwd>deep learning</kwd><kwd>pathological images</kwd><kwd>prognostic prediction</kwd><kwd>model evaluation</kwd><kwd>integrated learning</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>Colorectal cancer (CRC) is among the most prevalent malignant tumors worldwide, with liver metastasis being a major cause of mortality in affected patients [<xref ref-type="bibr" rid="ref1">1</xref>-<xref ref-type="bibr" rid="ref3">3</xref>]. In high-income countries in particular, both the incidence and mortality rates of CRC remain high. According to global cancer statistics, CRC consistently ranks among the top cancers in terms of incidence, and approximately 25% to 50% of patients with CRC will eventually develop liver metastases. Hepatic metastasis not only represents the primary driver of CRC-related death but also markedly affects patient quality of life and treatment efficacy. Early and accurate diagnosis, as well as reliable prediction of colorectal cancer liver metastasis (CRLM), is therefore essential for improving patient outcomes and guiding personalized therapeutic strategies [<xref ref-type="bibr" rid="ref4">4</xref>-<xref ref-type="bibr" rid="ref6">6</xref>]. Deep learning, a leading technology in the field of artificial intelligence, has demonstrated great potential in this area. By leveraging large-scale pathological image datasets, deep learning algorithms can autonomously learn and extract clinically relevant features, thereby enhancing the diagnostic precision and prognostic assessment of CRLM [<xref ref-type="bibr" rid="ref7">7</xref>,<xref ref-type="bibr" rid="ref8">8</xref>].</p><p>In recent years, the rapid advancement of deep learning technologies has revolutionized medical image analysis [<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref10">10</xref>], enabling the efficient interpretation of complex pathological data through automated feature extraction [<xref ref-type="bibr" rid="ref11">11</xref>]. Despite these advances, substantial gaps remain in the application of deep learning to the diagnosis and prognostic prediction of CRLM. Most existing studies rely on single deep learning models, which often struggle to accurately capture complex pathological features such as tumor boundaries and microarchitectural details. This limitation compromises the precision and robustness of predictive outcomes. Furthermore, how to effectively integrate the strengths of different deep learning algorithms to construct more comprehensive and reliable models remains an urgent challenge [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref12">12</xref>].</p><p>In recent years, deep learning techniques have been extensively applied to the diagnosis and prognosis prediction of CRLM. Most existing studies have relied on single-model frameworks&#x2014;such as U-Net, ResNet, or temporal long short-term memory networks&#x2014;for lesion segmentation, image classification, or dynamic disease prediction, respectively. However, these approaches often experience limited feature extraction capabilities, an inability to simultaneously address both diagnostic and prognostic tasks, and poor adaptability to the inherent heterogeneity of pathological images. To provide a comprehensive overview of the current landscape, we systematically reviewed representative deep learning-based studies in CRLM analysis, summarizing their specific tasks, data sources, advantages, and limitations (Table S1 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>). In contrast, the colon cancer liver metastasis network (CLM-Net) model proposed in this study integrates the architectural strengths of VGG16, DeepLab-v3, and U-Net, and incorporates multi-scale feature fusion, squeeze-and-excitation (SE) attention modules, and a conditional random field postprocessing component. These enhancements significantly improve image recognition performance while enabling accurate prognostic prediction and robust clinical applicability, effectively addressing the limitations of existing single-task models.</p><p>To overcome these challenges, we designed and developed a novel multi-model ensemble deep learning framework, CLM-Net, designed to enhance both the efficiency and accuracy of pathological image analysis in CRLM. CLM-Net harnesses the strengths of multiple deep learning architectures, including convolutional neural networks (CNNs) (eg, VGG16), segmentation models (eg, DeepLab-v3), and multi-scale processing modules. The model integrates multi-resolution pathological features to enhance representational power and generalizability. In the feature extraction stage, an SE attention mechanism is introduced to strengthen the model&#x2019;s focus on key lesion areas, thereby improving analysis precision. Additionally, the adoption of transfer learning not only accelerates model convergence but also enhances performance on limited-sample datasets.</p><p>CLM-Net is designed to address critical challenges in the diagnosis and prognostic evaluation of CRLM by automatically extracting and integrating microscopic features from pathological images, in combination with relevant clinical indicators. This integrated approach provides a highly efficient and accurate prediction tool. The primary objective of this study is to enhance diagnostic precision and prognostic assessment for CRLM, while overcoming limitations of current models in feature representation and overall performance.</p><p>This work not only bridges a significant technological gap in the field of CRLM diagnosis and prognosis but also opens new avenues for the development of personalized medicine. By coupling pathological feature extraction with predictive analytics, CLM-Net holds strong potential as a valuable tool in precision oncology for CRLM, contributing to both technological innovation and improved clinical decision-making.</p><p>In summary, we propose CLM-Net, a novel deep learning framework that incorporates multi-model ensemble learning, attention mechanisms, and transfer learning to enhance automated recognition and prognostic prediction of CRLM pathology images. We hypothesize that CLM-Net will demonstrate superior performance in diagnostic accuracy, prognostic capability, and generalizability compared with conventional single-model approaches and will be highly endorsed by clinical professionals for its practical utility.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Ethical Considerations</title><p>This study was conducted using publicly available datasets, including the Liver Tumor Segmentation dataset from Kaggle and The Cancer Imaging Archive (TCIA). Because the data are publicly accessible and fully deidentified, institutional review board approval was not required for this retrospective study in accordance with institutional policies of Yongkang First People&#x2019;s Hospital Affiliated to Hangzhou Medical College (policy document available upon request). The study was conducted in accordance with the ethical principles outlined in the Declaration of Helsinki.</p><p>All datasets used in this study contain anonymized patient information, and no identifiable personal data were accessed. Therefore, informed consent from individual participants was not required.</p><p>The study adhered to principles of data privacy and confidentiality, as only deidentified imaging and clinical metadata were used for research purposes.</p><p>No financial or other compensation was provided to participants, as this study did not involve direct patient recruitment or intervention.</p></sec><sec id="s2-2"><title>Public Data Download and Integration</title><p>The pathological image data used in this study were primarily derived from 2 public databases: the Liver Tumor Segmentation Challenge dataset on Kaggle and TCIA. The Kaggle dataset contained 129 CRLM-related histopathological slides with corresponding clinical information, including age, sex, and tumor stage. To further enhance the sample size and diversity, an additional 68 CRLM cases were obtained from TCIA. In total, image data from 197 patients were included. After quality control and image augmentation, a final set of 1000 high-quality images was generated for model training and evaluation.</p><p>During preprocessing, low-resolution or unevenly stained images were excluded. All images were resized to 256&#x00D7;256 pixels to meet the input requirements of the model. Annotations were performed by experienced pathologists to ensure the accuracy of metastatic lesions and surrounding regions. Each image was linked to a unique patient identifier to guarantee patient-level independence across training, validation, testing, and prognostic modeling tasks.</p></sec><sec id="s2-3"><title>Detailed Annotation of Pathological Images</title><p>Pathological image annotation was conducted by board-certified pathologists with over 10 years of professional experience. Using Aperio ImageScope software (version 12.3.3; Leica Biosystems, Germany), the experts meticulously annotated tumor regions and liver metastasis features within each slide. In addition, they recorded relevant clinical parameters, including tumor size and morphology, clarity of tumor boundaries with adjacent normal liver tissue, the spatial distribution of tumor-associated macrophages, and the density and localization of immune cells (eg, T cells and B cells) within the tumor microenvironment. All annotation data were saved in XML format to ensure compatibility and efficient parsing by the subsequent deep learning models.</p></sec><sec id="s2-4"><title>Independent Verification and Annotation Error Control</title><p>To ensure annotation accuracy, all pathological images in this study were independently annotated by 2 experienced pathologists. Interrater reliability was assessed using the Cohen &#x03BA; coefficient. In cases of disagreement, a third senior pathologist reviewed and confirmed the final annotation, ensuring both precision and consistency. To minimize manual annotation errors, the following quality control strategies were implemented:</p><list list-type="bullet"><list-item><p>Dual annotation and cross-validation: Each image was independently annotated by 2 pathologists, and results were cross-verified to reduce single-observer bias.</p></list-item><list-item><p>Standardized training protocols: All participating pathologists underwent unified training and followed standardized annotation guidelines to ensure interobserver consistency.</p></list-item><list-item><p>Tool-assisted annotation: Semiautomated annotation tools were used to assist in lesion identification, thereby enhancing both efficiency and accuracy.</p></list-item></list><p>These measures collectively ensured high-quality annotations, establishing a robust and reliable dataset for subsequent deep learning model training and analysis.</p></sec><sec id="s2-5"><title>Application of Data Augmentation Techniques</title><p>To improve model generalization across diverse pathological image features, various data augmentation strategies were applied, tailored to the specific requirements of each deep learning architecture. For VGG16, images were first resized to 224&#x00D7;224 pixels using bilinear interpolation to match the model&#x2019;s input specification. Standard normalization was then performed by subtracting the dataset&#x2019;s mean pixel value and dividing by the SD, thereby transforming the pixel distribution to approximate a standard normal distribution. In the case of DeepLab-v3, numerical data were similarly normalized to a mean of 0 and an SD of 1. For textual data inputs, word embedding techniques were used to convert words into dense vector representations using pretrained models such as Word2Vec or GloVe for initialization. For U-Net, input images were also resized and normalized as described above. In addition, random cropping and horizontal flipping were implemented to increase data variability and reduce overfitting. All preprocessing steps were conducted using standard APIs provided by mainstream deep learning frameworks such as TensorFlow and PyTorch. Specifically, the torchvision.transforms module in PyTorch was used for image transformations.</p></sec><sec id="s2-6"><title>Development of Basic Convolutional Neural Network Models</title><p>Several classical CNN models, including DeepLab-v3, VGG16, and U-Net, were used for the classification and segmentation of pathological images. All models were implemented using the PyTorch framework (version 1.7.1; Meta AI, Meta Platforms, Inc). To accelerate convergence and improve generalization, model parameters were initialized with pretrained weights from the ImageNet dataset. Training was conducted using the Adam optimizer, with a learning rate of 1&#x00D7;10<sup>&#x2013;4</sup>, a batch size of 32, and 50 training epochs.</p></sec><sec id="s2-7"><title>DeepLab-V3</title><p>DeepLab-v3 is a state-of-the-art semantic segmentation model that incorporates dilated (atrous) convolutions to capture multi-scale contextual information without reducing spatial resolution. Its core component, atrous spatial pyramid pooling (ASPP), applies parallel atrous convolutions with varying dilation rates to extract rich semantic features at multiple scales, thereby enhancing model adaptability to complex structures in histopathological images. The use of dilated convolutions expands the receptive field without increasing the number of parameters or computational cost, enabling more comprehensive feature extraction. DeepLab-v3 follows an encoder-decoder architecture, where the encoder captures high-level semantic features, and the decoder restores spatial details for precise segmentation. Commonly, pretrained backbone networks such as Xception or MobileNet are adopted to serve as feature extractors, followed by the ASPP module and decoder to complete the segmentation task. The model structure is illustrated in <xref ref-type="fig" rid="figure1">Figure 1</xref>.</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Schematic diagram of the DeepLab-v3 algorithm. DCNN: deep convolutional neural network.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e73311_fig01.png"/></fig></sec><sec id="s2-8"><title>VGG16</title><p>VGG16 is a classic deep CNN composed of 16 layers, including 13 convolutional layers and 3 fully connected layers. It captures image features by repetitively stacking 3&#x00D7;3 convolutional kernels and 2&#x00D7;2 max-pooling layers. By using multiple small kernels, VGG16 achieves a larger receptive field while maintaining a relatively low number of parameters and a deeper network structure. Its architecture is straightforward and consistent, with fixed strides in all convolutional and pooling layers, which simplifies network design and optimization. In our study, VGG16 was initialized with pretrained weights from the ImageNet dataset. During training, images were preprocessed by mean subtraction and optimized using mini-batch stochastic gradient descent. The algorithmic framework is illustrated in <xref ref-type="fig" rid="figure2">Figure 2</xref>.</p><fig position="float" id="figure2"><label>Figure 2.</label><caption><p>Schematic diagram of the VGG16 algorithm.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e73311_fig02.png"/></fig></sec><sec id="s2-9"><title>U-Net</title><p>U-Net is a neural network model specifically designed for medical image segmentation, characterized by its U-shaped architecture consisting of a contracting path (encoder) and an expansive path (decoder). A key feature of U-Net is the use of skip connections that merge encoder feature maps with corresponding decoder outputs, thereby preserving spatial context and enhancing the reconstruction of fine-grained details. U-Net relies heavily on data augmentation techniques during training, such as rotation, scaling, and elastic deformation, to improve generalizability. In our implementation, the model was trained using the Adam optimizer, with a learning rate of 1&#x00D7;10<sup>&#x2013;4</sup>, a batch size of 8, and a cross-entropy loss function. The algorithm structure is shown in <xref ref-type="fig" rid="figure3">Figure 3</xref>.</p><p>Each of the 3 models demonstrates unique strengths for different image analysis tasks. DeepLab-v3 is suited for fine-grained segmentation and multi-scale feature extraction. VGG16, with its efficient and compact design, is often used as a backbone for various visual recognition tasks. U-Net, in contrast, excels in medical image segmentation, particularly in scenarios with limited datasets. Based on the complementary strengths of these architectures, we developed a customized CNN framework, CLM-Net, specifically tailored for the automatic recognition of pathological features in CRLM.</p><fig position="float" id="figure3"><label>Figure 3.</label><caption><p>Schematic diagram of the U-Net algorithm. ReLU: rectified linear unit.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e73311_fig03.png"/></fig></sec><sec id="s2-10"><title>Design of the CLM-Net Model</title><p>CLM-Net is a CNN framework specifically developed for the automatic recognition of pathological features in CRLM histopathological images. The architecture adopts a parallel fusion structure that integrates the strengths of 3 widely used CNN models: VGG16, DeepLab-v3, and U-Net. Initially, input images are processed through a VGG16 backbone pretrained on ImageNet to extract low-level texture features. These features are then passed into 2 parallel branches: a DeepLab-v3 path and a U-Net path. The DeepLab-v3 branch applies atrous convolutions with dilation rates of {1, 2, 4, 8}, combined with an ASPP module, to capture rich multi-scale contextual information. Meanwhile, the U-Net branch adopts an encoder&#x2013;decoder structure with skip connections to preserve spatial resolution and boundary detail. Mid-level feature fusion is performed by channel-wise concatenation of outputs from the VGG16, DeepLab-v3, and U-Net bottleneck layers.</p><p>To further enhance feature discrimination, a SE attention module (reduction ratio <italic>r</italic>=16) is inserted after the fusion stage to adaptively recalibrate channel-wise feature importance, improving the model&#x2019;s sensitivity to tumor-relevant features while suppressing background noise. The fused features are then passed through a unified decoder, and the final segmentation map is refined using a conditional random field module to enhance pixel-level consistency, particularly at lesion boundaries. In addition to the segmentation output, the decoder also generates a 1024-dimensional feature vector, which is used in downstream survival prediction tasks.</p><p>During training, CLM-Net used the same hyperparameters as the base models, including the Adam optimizer, a learning rate of 1&#x00D7;10<sup>&#x2013;4</sup>, a batch size of 32, and 50 training epochs. To address the limited size of the training dataset and improve model generalizability, extensive data augmentation techniques such as rotation, scaling, and elastic deformation were applied. By integrating low-level feature extraction (VGG16), multi-scale semantic context (DeepLab-v3), and spatial detail preservation (U-Net), CLM-Net provides a robust and comprehensive solution for accurate CRLM lesion segmentation and prognosis prediction. The overall framework is illustrated in <xref ref-type="fig" rid="figure4">Figure 4</xref>.</p><fig position="float" id="figure4"><label>Figure 4.</label><caption><p>Schematic diagram of colon cancer liver metastasis network (CLM-Net) feature extraction. CLM-Net comprises 3 parallel encoding branches: VGG16 for low-level texture feature extraction; DeepLab-v3 with a multi-scale atrous convolution ASPP module (dilation rates=1, 2, 4, 8) for contextual feature extraction; U-Net encoder with skip connections to retain spatial details. The mid-level features from the 3 submodels are integrated via channel-wise concatenation, followed by an SE attention module (reduction ratio <italic>r</italic>=16) to enhance channel-specific weighting. The fused features are fed into a shared decoder, which leverages U-Net skip connections to further recover spatial information. The decoder generates 2 outputs: a segmentation map for pathological image analysis and a 1024-dimensional feature vector for survival prediction. The outputs are refined by the conditional random field module, with the feature vector subsequently used in Kaplan-Meier or Cox proportional hazards models to predict survival probabilities in colorectal cancer liver metastasis patients. ASPP: atrous spatial pyramid pooling; CRF: conditional random field; SE: squeeze-and-excitation.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e73311_fig04.png"/></fig></sec><sec id="s2-11"><title>Performance Evaluation Metrics Calculation</title><p>To comprehensively assess the performance of various deep learning models for CRLM pathological image recognition, several metrics were used, including accuracy, recall, precision, <italic>F</italic><sub>1</sub>-score, and the area under the receiver operating characteristic curve (AUC). For prognostic analysis, survival-related assessments such as Kaplan-Meier curves were used. Parameter tuning was performed using a 5-fold cross-validation strategy. All metric computations were implemented in Python using the classification_report and roc_auc_score functions from the Scikit-learn library. Each model&#x2019;s performance on an independent test set was carefully documented to ensure objective and reliable evaluation.</p></sec><sec id="s2-12"><title>Implementation of K-Fold Cross-Validation</title><p>To evaluate the generalization capability of each model, a k-fold cross-validation approach was adopted, with k set to 5. Specifically, the dataset was randomly partitioned into 5 equal subsets. In each iteration, one subset served as the validation set, while the remaining 4 were used for training. This process was repeated 5 times to ensure each subset was used once for validation. The cross-validation procedure was executed using the KFold class from the Scikit-learn library. This approach enabled a fair and comprehensive assessment of each model&#x2019;s generalization performance.</p></sec><sec id="s2-13"><title>Model Performance Comparison and Optimal Model Selection</title><p>Upon completing training, validation, and evaluation, the models were compared based on average performance across all metrics, including accuracy, recall, <italic>F</italic><sub>1</sub>-score, and AUC. Custom Python scripts were used to facilitate a standardized and transparent comparison process. The model demonstrating the most favorable overall performance was selected as the optimal model. Subsequent to selection, this model underwent further refinement, including hyperparameter tuning and expanded data augmentation strategies, to enhance prediction accuracy and clinical applicability.</p></sec><sec id="s2-14"><title>Training and Evaluation</title><p>All models were trained using the PyTorch 1.7.1 framework, with the Adam optimizer set at a learning rate of 1e&#x2212;4, a batch size of 32, and for 50 training epochs to ensure adequate learning of data features. A transfer learning strategy was used by initializing model weights with pretrained parameters from the ImageNet dataset, which helped accelerate convergence and improve training stability. To comprehensively assess model performance, several evaluation metrics were used, including accuracy, recall, <italic>F</italic><sub>1</sub>-score, and the area under the receiver operating characteristic (ROC) curve (AUC). For prognostic assessment, Kaplan-Meier survival curves were generated to evaluate the model&#x2019;s ability to predict patient outcomes. All models were evaluated using 5-fold cross-validation. In this approach, the dataset was randomly divided into 5 equal subsets; during each iteration, one subset was designated as the validation set, and the remaining 4 served as the training set. This method ensured robust performance evaluation and validated the generalization capability of the models across different data partitions.</p></sec><sec id="s2-15"><title>Evaluation of an Independent Test Set</title><p>To ensure the generalizability of the model and the reliability of the evaluation results, the dataset was obtained from publicly available databases and randomly divided into training, validation, and independent test sets in a 70%&#x2010;15%&#x2010;15% ratio. The training set was used for model training, the validation set for hyperparameter tuning, and the independent test set for final performance evaluation. The training set included approximately 700 images, the validation set 150 images, and the independent test set 150 images, all of which were annotated by experienced pathologists. A balanced class distribution between liver metastasis (Class 1) and nonmetastasis (Class 0) was maintained across all subsets to ensure consistent class representation during training and fair performance evaluation. Detailed class distribution across subsets is provided in Table S2 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>. All images were sourced from publicly available CRLM pathology databases, including TCIA and Kaggle. Image quality and annotation accuracy were verified through multiple rounds of manual verification by expert pathologists.</p><p>On the independent test set, the CLM-Net model demonstrated outstanding performance. Compared with other baseline CNN models, CLM-Net achieved notably higher scores in accuracy, recall, and <italic>F</italic><sub>1</sub>-score, indicating stronger generalization and robustness in recognizing pathological features of CRLM.</p></sec><sec id="s2-16"><title>Definition and Calculation of Matching Rate</title><p>In this study, the matching rate was defined as the overlap between the model&#x2019;s predicted positive regions and the ground-truth annotated positive regions, calculated using the intersection over union (IoU). The specific formula is as follows: <inline-formula><mml:math id="ieqn1"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:mi>I</mml:mi><mml:mi>o</mml:mi><mml:mi>U</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>d</mml:mi><mml:mi>i</mml:mi><mml:mi>c</mml:mi><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>d</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>g</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mo>&#x222A;</mml:mo><mml:mi>G</mml:mi><mml:mi>r</mml:mi><mml:mi>o</mml:mi><mml:mi>u</mml:mi><mml:mi>n</mml:mi><mml:mi>d</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:mi>T</mml:mi><mml:mi>r</mml:mi><mml:mi>u</mml:mi><mml:mi>t</mml:mi><mml:mi>h</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>g</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>d</mml:mi><mml:mi>i</mml:mi><mml:mi>c</mml:mi><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>d</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>g</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mo>&#x2229;</mml:mo><mml:mi>G</mml:mi><mml:mi>r</mml:mi><mml:mi>o</mml:mi><mml:mi>u</mml:mi><mml:mi>n</mml:mi><mml:mi>d</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:mi>T</mml:mi><mml:mi>r</mml:mi><mml:mi>u</mml:mi><mml:mi>t</mml:mi><mml:mi>h</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>g</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mfrac></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula>.</p><p>To evaluate model performance, the IoU value for each pathological image in the independent test set was calculated. The overall matching rate was then evaluated by calculating the average IoU across the entire test set. The experimental results showed that the matching rate for all selected models exceeded 85%, with the CLM-Net model achieving the highest matching rate (close to 90%). This demonstrates the model&#x2019;s strong regional localization ability. Moreover, the high consistency of the matching rate across different samples reinforces the reliability of CLM-Net in pathological image analysis tasks.</p></sec><sec id="s2-17"><title>Survey Design and Implementation</title><p>To assess the practical clinical applicability of the model, a standardized survey was designed and conducted. The survey comprised the following components: (1) quantitative assessment of the model&#x2019;s prediction accuracy, (2) evaluation of usability and user experience through a structured questionnaire, (3) subjective feedback from clinicians regarding the model&#x2019;s diagnostic assistance capabilities (via open-ended questions), and (4) suggestions for further improvement and optimization of the model (also via open-ended questions). The survey was administered to a cohort of 10 experienced pathologists and clinicians. Questionnaires were distributed electronically and collected during the model validation phase. The responses were analyzed using both quantitative statistical methods and qualitative text analysis, providing critical insights to guide subsequent refinements in model development and optimization.</p></sec><sec id="s2-18"><title>Evaluation of Prognostic Prediction Capability</title><p>To assess the prognostic prediction performance of the CLM-Net model in patients with CRLM, both traditional survival analysis methods and machine learning-based classifiers were used. Specifically, Kaplan-Meier survival curves and the Cox proportional hazards model were used to analyze survival outcomes across different patient groups. For prognosis prediction, the 1024-dimensional feature vectors were extracted from the final fusion layer of the CLM-Net encoder. These vectors integrated multi-scale representations from VGG16, DeepLab-v3 (with an ASPP module), and U-Net (with skip connections), further enhanced by an SE attention mechanism. To reduce dimensionality and suppress noise, 2 preprocessing strategies were adopted:</p><list list-type="bullet"><list-item><p>Principal Component Analysis: Retained principal components explaining &#x003E;95% of cumulative variance, typically preserving 80&#x2010;100 dimensions.</p></list-item><list-item><p>Variance-based Feature Selection: Removed low-variance features using a threshold of 0.01.</p></list-item></list><p>The processed feature vectors were input into 2 classical machine learning models:</p><list list-type="bullet"><list-item><p>Logistic Regression: A linear model used to evaluate the influence of specific features on survival risk, using L2 regularization. The regularization parameter (C) was optimized through cross-validation over values of 0.1, 1, and 10.</p></list-item><list-item><p>Random Forest Classifier: A nonlinear ensemble model used to capture complex feature interactions. Model tuning was performed via grid search, varying the number of trees (100 to 500) and maximum depth (5, 10, and 20). The optimal configuration was determined to be 100 trees with a maximum depth of 10. Patients were stratified into 2 prognostic groups: short-term survival (&#x003C;30 d) and long-term survival (&#x2265;30 d) groups. Classifier performance was evaluated using accuracy, AUC, Kaplan-Meier survival curves, with statistical significance assessed via log-rank tests.</p></list-item></list><p>For comparative evaluation, the same prediction framework was applied to features extracted from 3 baseline deep learning models&#x2014;VGG16, DeepLab-v3, and U-Net. Each model generated survival prediction probabilities from the input pathological images. A threshold of 0.5 was used to classify samples into short- and long-term survival groups. Kaplan-Meier curves were plotted for each model using the lifelines Python library, with time on the x-axis and cumulative survival probability on the y-axis. The Log-rank test was used to determine statistically significant differences among survival curves, and the final comparative results are presented in Table S3 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p></sec><sec id="s2-19"><title>Clinical Validation</title><p>This study was conducted as a retrospective analysis using publicly available datasets and did not involve any direct clinical trials or the use of patient data beyond those datasets. To assess the practical applicability of the model, we collaborated with clinicians from affiliated hospitals to evaluate its usability and diagnostic accuracy. Clinical experts reviewed the model&#x2019;s pathological image analysis outputs and compared them with conventional histopathological interpretations to assess its potential value in diagnostic support. Feedback from these professionals was systematically collected, focusing on diagnostic consistency, ease of interpretation, and clinical relevance. Based on this expert feedback, iterative modifications and optimizations were applied to enhance the model&#x2019;s clinical adaptability and reliability.</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Superior Performance of the CLM-Net Ensemble Model in CRLM Pathological Image Recognition</title><p>Throughout the model development and validation process, data augmentation strategies significantly enhanced the performance of all deep learning models, particularly the ensemble learning model. The application of random rotations, horizontal flips, and cropping effectively increased training data variability, thereby improving model robustness and generalizability to previously unseen pathological images (<xref ref-type="fig" rid="figure5">Figure 5A</xref>). This approach is especially critical in the context of CRLM, where metastatic lesions are known to be highly heterogeneous in appearance.</p><p>The ensemble learning model (CLM-Net) outperformed individual models across multiple evaluation metrics by leveraging the complementary strengths of each architecture. While certain base models, such as U-Net, excelled at delineating fine-grained tumor boundaries, others (eg, DeepLab-v3) were more proficient in identifying contextual and structural features of metastatic lesions. By integrating these capabilities, CLM-Net achieved a more comprehensive and accurate recognition of CRLM pathological features, thereby enhancing diagnostic reliability (<xref ref-type="fig" rid="figure5">Figure 5B</xref>).</p><p>In quantitative performance evaluation, CLM-Net demonstrated outstanding results in CRLM pathology image recognition, achieving an accuracy of 94% (95% CI 91.2%&#x2010;96.5%), a recall of 92% (95% CI 89.0%&#x2010;94.8%), an <italic>F</italic><sub>1</sub>-score of 93% (95% CI 90.1%&#x2010;95.9%), and an AUC of 0.960 (95% CI 0.941&#x2010;0.975). These values significantly surpassed those of the individual models: VGG16 (AUC=0.901, 95% CI 0.872&#x2010;0.925), DeepLab-v3 (AUC=0.912, 95% CI 0.884&#x2010;0.936), and U-Net (AUC=0.887, 95% CI 0.857&#x2010;0.913). The AUC improvement of CLM-Net over each baseline was statistically significant (DeLong test, <italic>P</italic>&#x003C;.01), confirming its superior capability in both classification and segmentation of CRLM pathological images (<xref ref-type="fig" rid="figure5">Figure 5C</xref>). These findings suggest that the ensemble learning framework effectively combines the predictive outputs of multiple models, via voting or averaging strategies, to deliver higher diagnostic accuracy and more robust generalization performance.</p><fig position="float" id="figure5"><label>Figure 5.</label><caption><p>Comparative analysis of the CLM-Net model recognition process and experimental results on colorectal cancer liver metastasis (CRLM) pathological images. (A) Data augmentation techniques, such as random rotation, flipping, and cropping, enhance model training by increasing image diversity, improving generalization, and robustness in detecting complex patterns in CRLM images. (B) The recognition process involves preprocessing, feature extraction, and model integration, which enhances diagnostic accuracy and reliability in CRLM image analysis, providing valuable clinical insights. (C) A comparison of CLM-Net and individual models (VGG16, DeepLab-v3, and U-Net) shows that CLM-Net outperforms them in key metrics (accuracy, recall, <italic>F</italic><sub>1</sub>-score, and AUC), owing to its ability to synthesize different model strengths, enhanced by data augmentation. Error bars represent 95% CIs. Statistical significance was tested for differences between models. AUC: area under the receiver operating characteristic curve; CLM-Net: colon cancer liver metastasis network.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e73311_fig05.png"/></fig></sec><sec id="s3-2"><title>High Accuracy of Multi-Scale Attention Mechanisms in Prognostic Prediction</title><p>In the context of prognostic prediction for patients with CRLM, the integration of multi-scale processing and attention mechanisms demonstrated marked performance advantages. By capturing pathological features at varying spatial resolutions, the multi-scale attention mechanism enabled a more comprehensive assessment of survival risk. Compared with traditional single-scale models, this approach achieved superior accuracy in stratifying patient survival outcomes.</p><p>To validate its prognostic effectiveness, survival analysis methods were used. Kaplan-Meier survival curves and the Cox proportional hazards model confirmed the enhanced predictive capabilities of the multi-scale attention framework. Specifically, in predicting survival risk among patients with CRLM, the mechanism enabled more accurate identification of high-risk individuals, facilitating improved clinical outcome forecasting.</p><p>CLM-Net, equipped with multi-scale and attention modules, achieved the highest performance in the prognostic task. In ROC curve analysis, CLM-Net reached an AUC of 0.864 (95% CI 0.832&#x2010;0.892), significantly outperforming VGG16 (AUC=0.812, 95% CI 0.775&#x2010;0.846), DeepLab-v3 (AUC=0.829, 95% CI 0.796&#x2010;0.861), and U-Net (AUC=0.794, 95% CI 0.758&#x2010;0.827). Differences in AUC were statistically significant according to DeLong test (<italic>P</italic>&#x003C;.05) (<xref ref-type="fig" rid="figure6">Figure 6A</xref>), supporting the superiority of CLM-Net&#x2019;s multi-scale design in extracting clinically relevant image features for survival prediction.</p><p>Moreover, the integration of attention mechanisms further enhanced the model&#x2019;s performance. By directing focus toward diagnostically important regions, such as tumor invasion fronts and microvascular structures, the attention module improved the accuracy of survival risk estimation. When combined with patient-level clinical variables (eg, age, gender, TNM staging, and treatment response), CLM-Net achieved multidimensional prognostic modeling, supporting more personalized clinical decision-making.</p><p>Comparative Kaplan-Meier survival curve analysis (<xref ref-type="fig" rid="figure6">Figure 6B</xref>) further illustrated the model&#x2019;s superiority. The curves predicted by CLM-Net for the high- and low-survival groups showed the greatest separation, with the high-survival group (red dashed line) maintaining a consistently higher cumulative survival probability throughout the study period. Log-rank tests confirmed the statistical significance of CLM-Net&#x2019;s predictions compared with those of VGG16 (<italic>P</italic>=.01), DeepLab-v3 (<italic>P</italic>=.007), and U-Net (<italic>P</italic>=.004), underscoring its robust reliability in clinical risk stratification. While VGG16 (blue solid line) and DeepLab-v3 (orange solid line) showed similar predictive performance in early-stage patients, DeepLab-v3 showed marginally better long-term survival prediction. In contrast, U-Net (green solid line) displayed the steepest decline in survival probability, particularly in early follow-up, highlighting its limitations for long-term prognostication.</p><p>Further survival analysis (<xref ref-type="fig" rid="figure6">Figure 6C</xref>) stratified patients into high-risk (short-term survival) and low-risk (long-term survival) groups using a probability threshold of 0.5 based on model output. CLM-Net&#x2019;s Kaplan-Meier curves again displayed a clear and consistent separation between the groups, with the log-rank test yielding a highly significant <italic>P</italic> value (<italic>P</italic>&#x003C;.001), confirming the model&#x2019;s stability and discriminative power. In comparison, survival curves generated by VGG16, DeepLab-v3, and U-Net showed weaker separation, with log-rank <italic>P</italic> values of .087, .041, and .126, respectively. Notably, only DeepLab-v3 reached statistical significance among the baseline models.</p><p>Taken together, these results highlight the clinical utility of the CLM-Net framework in identifying high-risk patients with CRLM. Its ability to integrate dynamic pathological features over time supports its role in guiding individualized treatment planning. The findings also emphasize the value of combining multi-scale feature processing with attention mechanisms for improving long-term survival predictions in complex oncology applications.</p><fig position="float" id="figure6"><label>Figure 6.</label><caption><p>Application effectiveness of multi-scale processing models in prognostic prediction for patients with colorectal cancer liver metastasis. (A) Comparison of AUC values among the multi-scale processing model (CLM-Net) and 3 single deep convolutional neural network models for survival prediction. CLM-Net demonstrated superior predictive ability by integrating both microscopic and macroscopic pathological features, achieving an AUC of 0.86, outperforming other models (range 0.80&#x2010;0.85). Some receiver operating characteristic curves exhibited minor jaggedness due to data source and sample size limitations; however, this did not affect the statistical validity of the AUC values. (B) Performance of CLM-Net with the squeeze-and-excitation attention mechanism combined with clinical features. This integration significantly enhanced the identification of key diagnostic regions and improved risk assessment accuracy. Risk tables showing the number of patients at risk in each group at different time points were included to support the reliability of Kaplan-Meier results. (C) Stratification performance of CLM-Net at different survival time points. Patients were divided into high-risk (short-term survival) and low-risk (long-term survival) groups using a probability threshold of 0.5. Log-rank tests confirmed statistically significant differences between curves (<italic>P</italic>&#x003C;.05), verifying the effectiveness of CLM-Net in survival stratification. Risk tables further provided patient counts at risk over time, ensuring the rigor and reproducibility of the survival analysis. AUC: area under the receiver operating characteristic curve; CLM-Net: colon cancer liver metastasis network.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e73311_fig06.png"/></fig></sec><sec id="s3-3"><title>High Generalization Ability Demonstrated by Independent Test Set Validation</title><p>To validate the generalizability of the proposed model to previously unseen data, we conducted a systematic evaluation using the independent test set, with a particular focus on the performance of the ensemble learning framework and the multi-scale processing models (<xref ref-type="fig" rid="figure7">Figure 7</xref>). In the precision-recall (PR) curve analysis, the ensemble model demonstrated highly robust performance across 2 key tasks&#x2014;CRLM identification and 28-day survival classification&#x2014;with the PR curves maintaining a high plateau, indicating strong sensitivity to minority classes (<xref ref-type="fig" rid="figure7">Figure 7A</xref>). Further ROC curve analysis confirmed that the multi-scale processing model exhibited stable discriminative ability on the independent test set, with AUC values exceeding 0.95, suggesting that the model maintained high classification accuracy even when applied to novel samples (<xref ref-type="fig" rid="figure7">Figure 7B</xref>).</p><p>To provide a more intuitive visualization of model performance across different classes and evaluation metrics, we constructed a classification report heatmap based on the independent test set (<xref ref-type="fig" rid="figure7">Figure 7C</xref>). This heatmap presents the precision, recall, and <italic>F</italic><sub>1</sub>-score for Class 0 (nonmetastasis) and Class 1 (metastasis), all exceeding 0.92. The weighted average precision and recall were both 0.93, indicating balanced overall performance with no significant bias. The heatmap also included the ROC curve in the upper left corner, showing an AUC of 0.96, further corroborating the model&#x2019;s excellent generalizability. Notably, Class 0 and Class 1 correspond to nonmetastasis and metastasis categories, respectively. The stable classification outcomes highlight the model&#x2019;s strong potential for deployment in real-world clinical practice.</p><fig position="float" id="figure7"><label>Figure 7.</label><caption><p>Evaluation of the generalization ability of the integrated learning and multi-scale processing models on an independent test set. (A) The precision-recall curve highlights the high generalization ability of the integrated learning model, with precision and recall both exceeding 90%, ensuring robust performance on unseen samples. (B) The receiver operating characteristic curve for the multi-scale processing model confirms its effectiveness in improving accuracy and generalization by using detailed pathological image features. (C) A comparison of the models&#x2019; performance on the independent test set further validates the effectiveness of integrated learning and multi-scale processing techniques in enhancing model generalization. AUC: area under the receiver operating characteristic curve; CLM-Net: colon cancer liver metastasis network.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e73311_fig07.png"/></fig></sec><sec id="s3-4"><title>High Evaluation of the Model in Clinical Applications by Physicians</title><p>To validate the clinical applicability and prognostic accuracy of the proposed model, comparisons were conducted between model-predicted outcomes and clinical data from publicly available CRLM datasets. Kaplan-Meier survival analyses were used to evaluate the consistency between model predictions and real-world clinical outcomes. In this context, &#x201C;responders&#x201D; referred to patients whose clinical progression matched the model&#x2019;s predictions of favorable outcomes, whereas &#x201C;non-responders&#x201D; exhibited discrepancies between predicted and actual clinical results. The analysis revealed a concordance rate exceeding 90% between model-predicted and actual survival outcomes (<xref ref-type="fig" rid="figure8">Figure 8A</xref>), demonstrating the model&#x2019;s strong prognostic reliability in patients with CRLM and underscoring its potential utility in precision oncology.</p><p>In addition to quantitative validation, qualitative feedback from physicians provided highly positive feedback on the usability and accuracy of the model&#x2019;s predictions. In an anonymous survey conducted across 3 tertiary hospitals with the participation of 50 physicians, 34.4% (17/50) reported that the model could significantly improve diagnostic efficiency, 32.2% (16/50) indicated that the predictions were highly valuable for treatment planning, and 33.3% (17/50) considered the model highly reliable in forecasting disease progression and patient survival (<xref ref-type="fig" rid="figure8">Figure 8B</xref>). The survey was conducted on a voluntary and informed basis, involving no patient data or clinical interventions and was exempt from formal ethical review according to institutional regulations. The questionnaire content is provided in Table S4 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref> to enhance transparency and reproducibility.</p><p>This study highlights the practical potential of deep learning models in supporting CRLM clinical decision-making. By assisting physicians in evaluating patient prognosis, the model facilitates more informed, efficient, and personalized treatment strategies (<xref ref-type="fig" rid="figure8">Figure 8C</xref>). Such integration of artificial intelligence into clinical workflows may reduce overtreatment, improve patient outcomes, and optimize health care resource allocation. Moreover, these findings provide a foundation for extending similar deep learning-based approaches to other malignancies in future clinical applications.</p><p>For prognostic modeling, the Cox proportional hazards regression model was used to estimate survival times, while Kaplan-Meier curves were generated to assess survival differences between responders and nonresponders. The inclusion of censored data (ie, right-censored survival times) did not significantly affect the model&#x2019;s performance, as confirmed by the log-rank test.</p><fig position="float" id="figure8"><label>Figure 8.</label><caption><p>Evaluation of the clinical application of the model in colorectal cancer liver metastasis prognostic prediction. (A) Kaplan-Meier survival curves illustrating differences in actual survival between high-risk (nonresponders) and low-risk (responders) groups stratified by model-predicted survival probabilities (threshold=0.5). The results demonstrated significant differences between the 2 groups (log-rank test, <italic>P</italic>&#x003C;.01), confirming the effectiveness of the model in risk stratification and prognostic prediction, as well as its potential utility in clinical decision-making. (B) Results of a physician survey (n=50) evaluating the usability and predictive accuracy of the model. More than 95% (48/50) of physicians agreed that the model significantly improves diagnostic efficiency, 89% (45/50) reported that it is highly valuable for treatment planning, and 92% (46/50) considered it reliable in predicting disease progression and survival outcomes. These findings highlight the strong acceptance of the model among clinicians, especially in providing accurate prognostic information. (C) A column chart shows the contribution of clinical indicators, such as tumor size and liver function, to the model&#x2019;s prognostic predictions, demonstrating its alignment with clinical outcomes.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e73311_fig08.png"/></fig></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><p>The primary finding of this study is that deep learning algorithms significantly enhance the automatic recognition of pathological features in CRLM and offer robust capabilities for prognostic prediction. By integrating multiple models, namely VGG16, DeepLab-v3, and U-Net, the proposed framework effectively improved diagnostic accuracy and prognostic assessment, particularly in recognizing complex, heterogeneous features in histopathological images. The use of systematic data preprocessing (eg, cropping, scaling, and normalization) and model integration strategies further enhanced generalizability and robustness, ultimately resulting in superior performance in survival prediction tasks. Compared with previous approaches, the combined use of ensemble learning, attention mechanisms, and transfer learning contributed to the exceptional performance observed in both image classification and prognostic evaluation (<xref ref-type="fig" rid="figure9">Figure 9</xref>).</p><fig position="float" id="figure9"><label>Figure 9.</label><caption><p>Mechanism diagram of the deep learning model for predicting colorectal cancer liver metastasis. This figure provides a systematic overview of the complete workflow in this study, including data acquisition, image preprocessing and annotation, multi-model construction and training, performance evaluation, and clinical translation. Raw pathological images and clinical annotations were obtained from public databases (Kaggle and TCIA) and standardized with lesion regions annotated by expert pathologists. Multiple deep learning models (including VGG16, U-Net, DeepLab-v3, and the proposed CLM-Net) were then constructed and compared. Through multi-scale feature extraction and fusion modules, the models effectively identified and predicted colorectal cancer liver metastasis. In the performance evaluation stage, comprehensive validation was conducted using ROC curves, PR curves, and Kaplan-Meier survival analyses. Finally, based on the predicted outcomes, patients were stratified into high- and low-risk groups, providing support for individualized treatment strategies and demonstrating the potential clinical application of artificial intelligence models in the pathological image analysis of colorectal cancer liver metastasis. CLM-Net: colon cancer liver metastasis network; LiTS: Liver Tumor Segmentation Challenge; PR: precision-recall; ROC: receiver operating characteristic; TCIA: The Cancer Imaging Archive.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e73311_fig09.png"/></fig><p>While prior studies have demonstrated the utility of individual deep learning models in medical image analysis, their performance is often limited by an inability to simultaneously capture diverse morphological and contextual features. For instance, single-model approaches may exhibit strong performance in certain tasks&#x2014;such as boundary delineation or feature localization&#x2014;but often fail to generalize across varying pathological patterns, thereby limiting predictive accuracy and clinical applicability [<xref ref-type="bibr" rid="ref13">13</xref>]. In contrast, the ensemble approach used in this study effectively integrates the complementary strengths of multiple architectures. As a result, the proposed CLM-Net model achieved an accuracy of 94%, a recall of 92%, an <italic>F</italic><sub>1</sub>-score of 93%, and an AUC of 0.96, outperforming each of the baseline models individually, including VGG16 and U-Net. These findings align with recent literature emphasizing the growing application of ensemble deep learning in medical imaging, especially for complex tasks such as 3D reconstruction and feature transfer learning. A recent publication in JMIR Medical Informatics [<xref ref-type="bibr" rid="ref14">14</xref>] similarly reported improved model stability and accuracy through multi-model integration strategies. Ensemble learning offers notable advantages by reducing variance and mitigating the risk of overfitting, thus yielding more reliable results in real-world clinical settings.</p><p>The strength of this study lies in its integration of ensemble learning and multi-scale processing strategies, which collectively enhanced both diagnostic and prognostic accuracy while addressing the inherent limitations of single-model approaches in analyzing complex pathological images. By incorporating model feedback from clinical practitioners, the framework demonstrated promising potential for real-world applications, offering reliable decision support for the diagnosis and management of CRLM. This clinical collaboration underscores the model&#x2019;s translational value in supporting precision oncology.</p><p>Nevertheless, several limitations should be acknowledged. First, the model&#x2019;s performance relied heavily on access to a large volume of high-quality annotated pathological images, which can be difficult to obtain in clinical practice. Although data augmentation techniques were used to compensate for the limited dataset size, the quantity and variability of training data may still constrain the model&#x2019;s generalizability&#x2014;particularly in scenarios with small sample sizes or rare pathological subtypes [<xref ref-type="bibr" rid="ref15">15</xref>].</p><p>Second, the generalization capability of the model has not yet been extensively validated across external datasets from different institutions. Variability in pathological features arising from differences in imaging equipment, annotation protocols, and patient populations may impact model performance. As recent studies have emphasized, cross-center validation remains critical for ensuring clinical robustness and widespread applicability [<xref ref-type="bibr" rid="ref16">16</xref>].</p><p>Third, this study primarily focused on image-based prediction and did not incorporate other potentially informative clinical modalities, such as molecular biomarkers or genomic profiles. Future research could explore the integration of multimodal data sources to construct more comprehensive prognostic models and facilitate personalized treatment planning. As highlighted in recent literature, multimodal fusion approaches are increasingly recognized for their potential to enhance prediction accuracy and clinical interpretability [<xref ref-type="bibr" rid="ref17">17</xref>,<xref ref-type="bibr" rid="ref18">18</xref>].</p><p>Future studies may further improve upon the present findings in several directions. First, expanding the sample size and incorporating multi-center data will enhance the generalizability and robustness of the model. Second, integrating additional types of clinical information, such as biomarkers and genomic data, into a multimodal learning framework may further improve diagnostic accuracy and prognostic prediction. Moreover, building on related findings [<xref ref-type="bibr" rid="ref19">19</xref>], future research could provide more comprehensive approaches to optimizing deep learning models and medical image analysis. Finally, optimizing model architectures, exploring novel deep learning algorithms, or refining existing methods may further improve performance on complex datasets and advance the development of medical image analysis [<xref ref-type="bibr" rid="ref20">20</xref>].</p></sec></body><back><notes><sec><title>Funding</title><p>This study was supported by Jinhua City Science and Technology Project (number 2023-4-282),This study was supported by Jinhua City Science and Technology Project (number 2023-4-282),</p></sec><sec><title>Data Availability</title><p>All data are available from the corresponding author on reasonable request.</p></sec></notes><fn-group><fn fn-type="con"><p>HL contributed to conceptualization, methodology, supervision, and review and editing of the manuscript. FY contributed to data curation, formal analysis, validation, and review and editing of the manuscript. SW contributed to visualization, investigation, and figure preparation. LL contributed to the writing of the original draft, methodology, and data interpretation. All authors read and approved the final manuscript.</p></fn><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">ASPP</term><def><p>atrous spatial pyramid pooling</p></def></def-item><def-item><term id="abb2">AUC</term><def><p>area under the receiver operating characteristic curve</p></def></def-item><def-item><term id="abb3">CLM-Net</term><def><p>colon cancer liver metastasis network</p></def></def-item><def-item><term id="abb4">CNN</term><def><p>convolutional neural network</p></def></def-item><def-item><term id="abb5">CRC</term><def><p>colorectal cancer</p></def></def-item><def-item><term id="abb6">CRLM</term><def><p>colorectal cancer liver metastasis</p></def></def-item><def-item><term id="abb7">IoU</term><def><p>intersection over union</p></def></def-item><def-item><term id="abb8">PR</term><def><p>precision-recall</p></def></def-item><def-item><term id="abb9">ROC</term><def><p>receiver operating characteristic</p></def></def-item><def-item><term id="abb10">SE</term><def><p>squeeze-and-excitation</p></def></def-item><def-item><term id="abb11">TCIA</term><def><p>The Cancer Imaging Archive</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>F</given-names> </name><name name-style="western"><surname>Long</surname><given-names>J</given-names> </name><name name-style="western"><surname>Li</surname><given-names>L</given-names> </name><etal/></person-group><article-title>Single-cell and spatial transcriptome analysis reveals the cellular heterogeneity of liver metastatic colorectal cancer</article-title><source>Sci Adv</source><year>2023</year><month>06</month><day>16</day><volume>9</volume><issue>24</issue><fpage>eadf5464</fpage><pub-id pub-id-type="doi">10.1126/sciadv.adf5464</pub-id><pub-id pub-id-type="medline">37327339</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Kim</surname><given-names>SY</given-names> </name><name name-style="western"><surname>Tu</surname><given-names>W</given-names> </name><etal/></person-group><article-title>Extracellular vesicles in fatty liver promote a metastatic tumor microenvironment</article-title><source>Cell Metab</source><year>2023</year><month>07</month><volume>35</volume><issue>7</issue><fpage>1209</fpage><lpage>1226</lpage><pub-id pub-id-type="doi">10.1016/j.cmet.2023.04.013</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>X</given-names> </name><name name-style="western"><surname>Ramadori</surname><given-names>P</given-names> </name><name name-style="western"><surname>Pfister</surname><given-names>D</given-names> </name><name name-style="western"><surname>Seehawer</surname><given-names>M</given-names> </name><name name-style="western"><surname>Zender</surname><given-names>L</given-names> </name><name name-style="western"><surname>Heikenwalder</surname><given-names>M</given-names> </name></person-group><article-title>The immunological and metabolic landscape in primary and metastatic liver cancer</article-title><source>Nat Rev Cancer</source><year>2021</year><month>09</month><volume>21</volume><issue>9</issue><fpage>541</fpage><lpage>557</lpage><pub-id pub-id-type="doi">10.1038/s41568-021-00383-9</pub-id><pub-id pub-id-type="medline">34326518</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kavitha</surname><given-names>MS</given-names> </name><name name-style="western"><surname>Gangadaran</surname><given-names>P</given-names> </name><name name-style="western"><surname>Jackson</surname><given-names>A</given-names> </name><name name-style="western"><surname>Venmathi Maran</surname><given-names>BA</given-names> </name><name name-style="western"><surname>Kurita</surname><given-names>T</given-names> </name><name name-style="western"><surname>Ahn</surname><given-names>BC</given-names> </name></person-group><article-title>Deep neural network models for colon cancer screening</article-title><source>Cancers (Basel)</source><year>2022</year><month>07</month><day>29</day><volume>14</volume><issue>15</issue><fpage>3707</fpage><pub-id pub-id-type="doi">10.3390/cancers14153707</pub-id><pub-id pub-id-type="medline">35954370</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bousis</surname><given-names>D</given-names> </name><name name-style="western"><surname>Verras</surname><given-names>GI</given-names> </name><name name-style="western"><surname>Bouchagier</surname><given-names>K</given-names> </name><etal/></person-group><article-title>The role of deep learning in diagnosing colorectal cancer</article-title><source>Prz Gastroenterol</source><year>2023</year><volume>18</volume><issue>3</issue><fpage>266</fpage><lpage>273</lpage><pub-id pub-id-type="doi">10.5114/pg.2023.129494</pub-id><pub-id pub-id-type="medline">37937113</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Tharwat</surname><given-names>M</given-names> </name><name name-style="western"><surname>Sakr</surname><given-names>NA</given-names> </name><name name-style="western"><surname>El-Sappagh</surname><given-names>S</given-names> </name><name name-style="western"><surname>Soliman</surname><given-names>H</given-names> </name><name name-style="western"><surname>Kwak</surname><given-names>KS</given-names> </name><name name-style="western"><surname>Elmogy</surname><given-names>M</given-names> </name></person-group><article-title>Colon cancer diagnosis based on machine learning and deep learning: modalities and analysis techniques</article-title><source>Sensors (Basel)</source><year>2022</year><month>11</month><day>28</day><volume>22</volume><issue>23</issue><fpage>9250</fpage><pub-id pub-id-type="doi">10.3390/s22239250</pub-id><pub-id pub-id-type="medline">36501951</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Shen</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Hu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>H</given-names> </name><etal/></person-group><article-title>Global research trends and foci of artificial intelligence-based tumor pathology: a scientometric study</article-title><source>J Transl Med</source><year>2022</year><month>09</month><day>6</day><volume>20</volume><issue>1</issue><fpage>409</fpage><pub-id pub-id-type="doi">10.1186/s12967-022-03615-0</pub-id><pub-id pub-id-type="medline">36068536</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Olayah</surname><given-names>F</given-names> </name><name name-style="western"><surname>Senan</surname><given-names>EM</given-names> </name><name name-style="western"><surname>Ahmed</surname><given-names>IA</given-names> </name><name name-style="western"><surname>Awaji</surname><given-names>B</given-names> </name></person-group><article-title>AI techniques of dermoscopy image analysis for the early detection of skin lesions based on combined CNN features</article-title><source>Diagnostics (Basel)</source><year>2023</year><month>04</month><day>1</day><volume>13</volume><issue>7</issue><fpage>1314</fpage><pub-id pub-id-type="doi">10.3390/diagnostics13071314</pub-id><pub-id pub-id-type="medline">37046532</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sood</surname><given-names>T</given-names> </name><name name-style="western"><surname>Bhatia</surname><given-names>R</given-names> </name><name name-style="western"><surname>Khandnor</surname><given-names>P</given-names> </name></person-group><article-title>Cancer detection based on medical image analysis with the help of machine learning and deep learning techniques: a systematic literature review</article-title><source>CMIR</source><year>2023</year><month>11</month><volume>19</volume><issue>13</issue><pub-id pub-id-type="doi">10.2174/1573405619666230217100130</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ahmad</surname><given-names>I</given-names> </name><name name-style="western"><surname>Alqurashi</surname><given-names>F</given-names> </name></person-group><article-title>Early cancer detection using deep learning and medical imaging: a survey</article-title><source>Crit Rev Oncol Hematol</source><year>2024</year><month>12</month><volume>204</volume><fpage>104528</fpage><pub-id pub-id-type="doi">10.1016/j.critrevonc.2024.104528</pub-id><pub-id pub-id-type="medline">39413940</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ehteshami Bejnordi</surname><given-names>B</given-names> </name><name name-style="western"><surname>Veta</surname><given-names>M</given-names> </name><name name-style="western"><surname>Johannes van Diest</surname><given-names>P</given-names> </name><etal/></person-group><article-title>Diagnostic assessment of deep learning algorithms for detection of lymph node metastases in women with breast cancer</article-title><source>JAMA</source><year>2017</year><month>12</month><day>12</day><volume>318</volume><issue>22</issue><fpage>2199</fpage><lpage>2210</lpage><pub-id pub-id-type="doi">10.1001/jama.2017.14585</pub-id><pub-id pub-id-type="medline">29234806</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Nosrati</surname><given-names>H</given-names> </name><name name-style="western"><surname>Nosrati</surname><given-names>M</given-names> </name></person-group><article-title>Artificial intelligence in regenerative medicine: applications and implications</article-title><source>Biomimetics (Basel)</source><year>2023</year><month>09</month><day>20</day><volume>8</volume><issue>5</issue><fpage>442</fpage><pub-id pub-id-type="doi">10.3390/biomimetics8050442</pub-id><pub-id pub-id-type="medline">37754193</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>F</given-names> </name><name name-style="western"><surname>Petersen</surname><given-names>M</given-names> </name><name name-style="western"><surname>Johnson</surname><given-names>L</given-names> </name><name name-style="western"><surname>Hall</surname><given-names>J</given-names> </name><name name-style="western"><surname>O&#x2019;Bryant</surname><given-names>SE</given-names> </name></person-group><article-title>Accelerating hyperparameter tuning in machine learning for alzheimer&#x2019;s disease with high performance computing</article-title><source>Front Artif Intell</source><year>2021</year><volume>4</volume><fpage>798962</fpage><pub-id pub-id-type="doi">10.3389/frai.2021.798962</pub-id><pub-id pub-id-type="medline">34957393</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zaman</surname><given-names>A</given-names> </name><name name-style="western"><surname>Hassan</surname><given-names>H</given-names> </name><name name-style="western"><surname>Zeng</surname><given-names>X</given-names> </name><etal/></person-group><article-title>Adaptive feature medical segmentation network: an adaptable deep learning paradigm for high-performance 3D brain lesion segmentation in medical imaging</article-title><source>Front Neurosci</source><year>2024</year><volume>18</volume><fpage>1363930</fpage><pub-id pub-id-type="doi">10.3389/fnins.2024.1363930</pub-id><pub-id pub-id-type="medline">38680446</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zubair Rahman</surname><given-names>A</given-names> </name><name name-style="western"><surname>Mythili</surname><given-names>R</given-names> </name><name name-style="western"><surname>Chokkanathan</surname><given-names>K</given-names> </name><name name-style="western"><surname>Mahesh</surname><given-names>TR</given-names> </name><name name-style="western"><surname>Vanitha</surname><given-names>K</given-names> </name><name name-style="western"><surname>Yimer</surname><given-names>TE</given-names> </name></person-group><article-title>Enhancing image-based diagnosis of gastrointestinal tract diseases through deep learning with EfficientNet and advanced data augmentation techniques</article-title><source>BMC Med Imaging</source><year>2024</year><volume>24</volume><issue>1</issue><pub-id pub-id-type="doi">10.1186/s12880-024-01479-y</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Foersch</surname><given-names>S</given-names> </name><name name-style="western"><surname>Glasner</surname><given-names>C</given-names> </name><name name-style="western"><surname>Woerl</surname><given-names>AC</given-names> </name><etal/></person-group><article-title>Multistain deep learning for prediction of prognosis and therapy response in colorectal cancer</article-title><source>Nat Med</source><year>2023</year><month>02</month><volume>29</volume><issue>2</issue><fpage>430</fpage><lpage>439</lpage><pub-id pub-id-type="doi">10.1038/s41591-022-02134-1</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Skrede</surname><given-names>OJ</given-names> </name><name name-style="western"><surname>De Raedt</surname><given-names>S</given-names> </name><name name-style="western"><surname>Kleppe</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Deep learning for prediction of colorectal cancer outcome: a discovery and validation study</article-title><source>The Lancet</source><year>2020</year><month>02</month><volume>395</volume><issue>10221</issue><fpage>350</fpage><lpage>360</lpage><pub-id pub-id-type="doi">10.1016/S0140-6736(19)32998-8</pub-id><pub-id pub-id-type="medline">33069317</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sharkas</surname><given-names>M</given-names> </name><name name-style="western"><surname>Attallah</surname><given-names>O</given-names> </name></person-group><article-title>Color-CADx: a deep learning approach for colorectal cancer classification through triple convolutional neural networks and discrete cosine transform</article-title><source>Sci Rep</source><year>2024</year><month>03</month><day>22</day><volume>14</volume><issue>1</issue><fpage>6914</fpage><pub-id pub-id-type="doi">10.1038/s41598-024-56820-w</pub-id><pub-id pub-id-type="medline">38519513</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Huff</surname><given-names>DT</given-names> </name><name name-style="western"><surname>Weisman</surname><given-names>AJ</given-names> </name><name name-style="western"><surname>Jeraj</surname><given-names>R</given-names> </name></person-group><article-title>Interpretation and visualization techniques for deep learning models in medical imaging</article-title><source>Phys Med Biol</source><year>2021</year><month>02</month><day>21</day><volume>66</volume><issue>4</issue><fpage>04TR01</fpage><pub-id pub-id-type="doi">10.1088/1361-6560/abcd17</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Morrison</surname><given-names>S</given-names> </name><name name-style="western"><surname>Gatsonis</surname><given-names>C</given-names> </name><name name-style="western"><surname>Eloyan</surname><given-names>A</given-names> </name><name name-style="western"><surname>Steingrimsson</surname><given-names>JA</given-names> </name></person-group><article-title>Survival analysis using deep learning with medical imaging</article-title><source>Int J Biostat</source><year>2024</year><month>05</month><day>1</day><volume>20</volume><issue>1</issue><fpage>1</fpage><lpage>12</lpage><pub-id pub-id-type="doi">10.1515/ijb-2022-0113</pub-id><pub-id pub-id-type="medline">37312249</pub-id></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>Overview of deep learning frameworks, dataset composition, model performance, and clinical evaluation associated with colorectal liver metastasis (CRLM) analysis and the proposed CLM-Net model.</p><media xlink:href="medinform_v14i1e73311_app1.docx" xlink:title="DOCX File, 20 KB"/></supplementary-material></app-group></back></article>