<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Med Inform</journal-id><journal-id journal-id-type="publisher-id">medinform</journal-id><journal-id journal-id-type="index">7</journal-id><journal-title>JMIR Medical Informatics</journal-title><abbrev-journal-title>JMIR Med Inform</abbrev-journal-title><issn pub-type="epub">2291-9694</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v13i1e75665</article-id><article-id pub-id-type="doi">10.2196/75665</article-id><article-categories><subj-group subj-group-type="heading"><subject>Original Paper</subject></subj-group></article-categories><title-group><article-title>Deep Learning Radiomics Model Based on Computed Tomography Image for Predicting the Classification of Osteoporotic Vertebral Fractures: Algorithm Development and Validation</article-title></title-group><contrib-group><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Liu</surname><given-names>Jiayi</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Zhang</surname><given-names>Lincen</given-names></name><degrees>MS</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Yuan</surname><given-names>Yousheng</given-names></name><degrees>MS</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Tang</surname><given-names>Jun</given-names></name><degrees>MS</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Liu</surname><given-names>Yongkang</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff3">3</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Xia</surname><given-names>Liang</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Zhang</surname><given-names>Jun</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib></contrib-group><aff id="aff1"><institution>Department of Radiology, Sir Run Run Hospital, Nanjing Medical University</institution><addr-line>109 Longmian Road</addr-line><addr-line>Nanjing</addr-line><country>China</country></aff><aff id="aff2"><institution>Department of Radiology, The Affiliated Taizhou People's Hospital of Nanjing Medical University</institution><addr-line>Taizhou</addr-line><country>China</country></aff><aff id="aff3"><institution>Department of Radiology, The Affiliated Hospital of Nanjing University of Chinese Medicine</institution><addr-line>Nanjing</addr-line><country>China</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Coristine</surname><given-names>Andrew</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Tang</surname><given-names>Guangyu</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Zeng</surname><given-names>Qingshi</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Jun Zhang, PhD, Department of Radiology, Sir Run Run Hospital, Nanjing Medical University, 109 Longmian Road, Nanjing, 211100, China, 86 18851667275; <email>zhangjun21416@Hotmail.com</email></corresp><fn fn-type="equal" id="equal-contrib1"><label>*</label><p>these authors contributed equally</p></fn></author-notes><pub-date pub-type="collection"><year>2025</year></pub-date><pub-date pub-type="epub"><day>29</day><month>8</month><year>2025</year></pub-date><volume>13</volume><elocation-id>e75665</elocation-id><history><date date-type="received"><day>10</day><month>04</month><year>2025</year></date><date date-type="rev-recd"><day>29</day><month>07</month><year>2025</year></date><date date-type="accepted"><day>30</day><month>07</month><year>2025</year></date></history><copyright-statement>&#x00A9; Jiayi Liu, Lincen Zhang, Yousheng Yuan, Jun Tang, Yongkang Liu, Liang Xia, Jun Zhang. Originally published in JMIR Medical Informatics (<ext-link ext-link-type="uri" xlink:href="https://medinform.jmir.org">https://medinform.jmir.org</ext-link>), 29.8.2025. </copyright-statement><copyright-year>2025</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://medinform.jmir.org/">https://medinform.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://medinform.jmir.org/2025/1/e75665"/><abstract><sec><title>Background</title><p>Osteoporotic vertebral fractures (OVFs) are common in older adults and often lead to disability if not properly diagnosed and classified. With the increased use of computed tomography (CT) imaging and the development of radiomics and deep learning technologies, there is potential to improve the classification accuracy of OVFs.</p></sec><sec><title>Objective</title><p>This study aims to evaluate the efficacy of a deep learning radiomics model, derived from CT imaging, in accurately classifying OVFs.</p></sec><sec sec-type="methods"><title>Methods</title><p>The study analyzed 981 patients (aged 50&#x2010;95 years; 687 women, 294 men), involving 1098 vertebrae, from 3 medical centers who underwent both CT and magnetic resonance imaging examinations. The Assessment System of Thoracolumbar Osteoporotic Fractures (ASTLOF) classified OVFs into Classes 0, 1, and 2. The data were categorized into 4 cohorts: training (n=750), internal validation (n=187), external validation (n=110), and prospective validation (n=51). Deep transfer learning used the ResNet-50 architecture, pretrained on RadImageNet and ImageNet, to extract imaging features. Deep transfer learning&#x2013;based features were combined with radiomics features and refined using Least Absolute Shrinkage and Selection Operator (LASSO) regression. The performance of 8 machine learning classifiers for OVF classification was assessed using receiver operating characteristic metrics and the &#x201C;One-vs-Rest&#x201D; approach. Performance comparisons between RadImageNet- and ImageNet-based models were performed using the DeLong test. Shapley Additive Explanations (SHAP) analysis was used to interpret feature importance and the predictive rationale of the optimal fusion model.</p></sec><sec sec-type="results"><title>Results</title><p>Feature selection and fusion yielded 33 and 54 fused features for the RadImageNet- and ImageNet-based models, respectively, following pretraining on the training set. The best-performing machine learning algorithms for these 2 deep learning radiomics models were the multilayer perceptron and Light Gradient Boosting Machine (LightGBM). The macro-average area under the curve (AUC) values for the fused models based on RadImageNet and ImageNet were 0.934 and 0.996, respectively, with DeLong test showing no statistically significant difference (<italic>P</italic>=2.34). The RadImageNet-based model significantly surpassed the ImageNet-based model across internal, external, and prospective validation sets, with macro-average AUCs of 0.837 versus 0.648, 0.773 versus 0.633, and 0.852 versus 0.648, respectively (<italic>P</italic>&#x003C;.05). Using the binary &#x201C;One-vs-Rest&#x201D; approach, the RadImageNet-based fused model achieved superior predictive performance for Class 2 (AUC=0.907, 95% CI 0.805&#x2010;0.999), with Classes 0 and 1 following (AUC/accuracy=0.829/0.803 and 0.794/0.768, respectively). SHAP analysis provided a visualization of feature importance in the RadImageNet-based fused model, highlighting the top 3 most influential features: cluster shade, mean, and large area low gray level emphasis, and their respective impacts on predictions.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>The RadImageNet-based fused model using CT imaging data exhibited superior predictive performance compared to the ImageNet-based model, demonstrating significant utility in OVF classification and aiding clinical decision-making for treatment planning. Among the 3 classes, the model performed best in identifying Class 2, followed by Class 0 and Class 1.</p></sec></abstract><kwd-group><kwd>deep learning</kwd><kwd>radiomics</kwd><kwd>osteoporotic vertebral fractures</kwd><kwd>tomography</kwd><kwd>classification</kwd><kwd>model interpretability</kwd><kwd>x-ray computed</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>Osteoporotic vertebral fractures (OVFs) have a subtle onset and complex progression, affecting about 40% of postmenopausal women and 25%&#x2010;33% of older men. In China, a new OVF case is reported approximately every 17.4 seconds [<xref ref-type="bibr" rid="ref1">1</xref>]. OVFs are linked to high disability and mortality rates [<xref ref-type="bibr" rid="ref2">2</xref>]. Accurate classification of OVFs is widely recognized as critical for early diagnosis, treatment planning, and prognosis evaluation [<xref ref-type="bibr" rid="ref3">3</xref>]. Current classification systems, including the Genant semi-quantitative method [<xref ref-type="bibr" rid="ref4">4</xref>], Heini classification [<xref ref-type="bibr" rid="ref5">5</xref>], osteoporotic fractures classification [<xref ref-type="bibr" rid="ref6">6</xref>], and the Assessment System of Thoracolumbar Osteoporotic Fractures (ASTLOF) [<xref ref-type="bibr" rid="ref7">7</xref>], differ in methodology but lack global agreement [<xref ref-type="bibr" rid="ref8">8</xref>]. Among these, the ASTLOF system has demonstrated good reproducibility and clinical utility, integrating vertebral morphology, magnetic resonance imaging (MRI) signal characteristics, bone mineral density (BMD), and pain severity into a preoperative scoring framework. This system supports targeted treatment selection and provides valuable clinical guidance [<xref ref-type="bibr" rid="ref9">9</xref>]. Accordingly, the ASTLOF classification was adopted as the standard in this study.</p><p>Computed tomography (CT) imaging, with its high spatial resolution, allows for detailed observation of subtle changes in vertebral endplates, cortical bone, and cancellous bone, providing a more reliable basis for OVF classification and clinical guidance [<xref ref-type="bibr" rid="ref10">10</xref>]. Currently, CT equipment is widely available in secondary and tertiary hospitals, and some community hospitals have also introduced CT systems [<xref ref-type="bibr" rid="ref11">11</xref>]. CT imaging is crucial for accurately classifying OVFs and has substantial clinical importance. Radiomics aids in analyzing trabecular bone microstructure [<xref ref-type="bibr" rid="ref12">12</xref>], assessing BMD [<xref ref-type="bibr" rid="ref13">13</xref>], differentiating acute from chronic OVFs [<xref ref-type="bibr" rid="ref14">14</xref>], and predicting residual back pain in these patients [<xref ref-type="bibr" rid="ref15">15</xref>]. Deep learning radiomics (DLR) uses network architectures such as ResNet, pretrained on ImageNet, to extract deep imaging features from images, a widely adopted approach. RadImageNet, as an open-access, public medical imaging dataset, theoretically provides better performance for deep transfer learning (DTL) in medical imaging tasks compared to ImageNet [<xref ref-type="bibr" rid="ref16">16</xref>]. Our research team has preliminarily validated this hypothesis [<xref ref-type="bibr" rid="ref17">17</xref>]. However, whether it can improve the performance of 3-class prediction models still requires further exploration and verification.</p><p>In this study, we used thoracolumbar medical imaging data from multiple health care centers. We applied DTL on both RadImageNet and ImageNet datasets to extract DTL features and also used the open-source PyRadiomics package (developed by the Computational Imaging &#x0026; Bioinformatics Lab, Brigham and Women&#x2019;s Hospital/Harvard Medical School; lead developer: Joost J. M. van Griethuysen) to extract traditional radiomics features. We developed a predictive model for OVF classification using CT imaging by integrating and selecting DTL and radiomics features within the ASTLOF system. The model was validated, tested, and compared for its predictive performance using multicenter data.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Patient Selection</title><p>This study used medical imaging data from 3 Chinese hospitals, with ethics committee approval from each institution. The retrospective study design negated the need for informed patient consent. CT and MRI data from patients diagnosed with OVFs at Center I (Taizhou People&#x2019;s Hospital, Nanjing Medical University) and Center II (Affiliated Hospital of Nanjing University of Chinese Medicine) between December 2018 and December 2024 were used to create the training, internal validation, and external validation datasets. <xref ref-type="other" rid="box1">Textbox 1</xref> shows the inclusion criteria and exclusion criteria. An independent test dataset was acquired from Center III (Sir Run Run Hospital, Nanjing Medical University) spanning January to December 2024. The dataset&#x2019;s inclusion and exclusion criteria matched those of the training and validation datasets.</p><p><xref ref-type="fig" rid="figure1">Figures 1</xref> and <xref ref-type="fig" rid="figure2">2</xref> provide detailed information on case collection, grouping, image preprocessing, feature extraction, analysis, and model development through flowcharts and the DLR workflow.</p><boxed-text id="box1"><p><bold>Inclusion criteria</bold></p><list list-type="bullet"><list-item><p>Patients aged &#x2265;50 years, meeting osteoporotic vertebral fracture diagnostic criteria [<xref ref-type="bibr" rid="ref18">18</xref>], with no trauma history or only minor trauma.</p></list-item><list-item><p>Complete computed tomography and magnetic resonance imaging DICOM data, with no more than a 2-week interval between the 2 scans.</p></list-item><list-item><p>Comprehensive clinical records encompassing gender, age, and dual-energy X-ray absorptiometry outcomes.</p></list-item><list-item><p>Clinical presentations, such as absence of significant pain, back pain triggered by posture, persistent pain, or neurological symptoms.</p></list-item></list><p><bold>Exclusion criteria</bold></p><list list-type="bullet"><list-item><p>Suspected infections or pathological fractures related to tumors.</p></list-item><list-item><p>Poor image quality or artifacts caused by foreign objects.</p></list-item><list-item><p>Uncertain osteoporotic vertebral fracture classification.</p></list-item></list></boxed-text><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>The flowchart in this study outlines the key steps and processes involved in the research workflow.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v13i1e75665_fig01.png"/></fig><fig position="float" id="figure2"><label>Figure 2.</label><caption><p>The workflow of the deep learning radiomics process illustrates the systematic steps involved in data processing, feature extraction, model development, and validation. ROC: receiver operating characteristic; ROI: region of interest.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v13i1e75665_fig02.png"/></fig></sec><sec id="s2-2"><title>CT and MRI Acquisition Protocol</title><sec id="s2-2-1"><title>CT Scans</title><p>CT scans were performed across 3 centers using multidetector or dual-source CT systems, including: GE Lightspeed Ultra (16-slice, USA), Siemens Somatom Definition (128-slice and 256-slice, Germany), Siemens Sensation64 (64-slice, Germany), Philips Brilliance iCT (256-slice, Netherlands), GE Optima CT670 (64-slice, USA). Key scan parameters were: tube voltage 120 kVp (with or without automated current modulation), tube current 118&#x2010;320 mA (with or without automated current modulation), and image matrix 512&#x00D7;512, and layer thickness and interval 1 mm.</p></sec><sec id="s2-2-2"><title>MRI Scans</title><p>MRI examinations were conducted on 3.0T scanners from Siemens (Verio, Skyra, and Prisma), Philips (Ingenia CX and Achieva TX), across the 3 centers. Image sequences included short tau inversion recovery (STIR) and T2-weighted fat-suppressed images. All patients underwent STIR or T2-weighted fat-suppressed MRI scans. Additional details regarding the imaging devices and parameters for both CT and MRI are available in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p></sec></sec><sec id="s2-3"><title>Classification</title><p>The ASTLOF system classifies OVFs by assigning scores based on vertebral morphology, MRI findings, BMD, and clinical symptoms. Changes in morphology seen in CT or MRI scans are rated as normal (0 points), compression (1 point), or burst fracture (2 points). MRI assessments use sagittal T2-weighted fat-suppressed or STIR sequences, assigning scores based on normal appearance (0 points), high signal alterations (1 point), or the presence of vacuum or fluid signs within vertebrae (2 points). BMD is assessed via T-scores, with values &#x003E;&#x2212;2.5 scoring 0, between &#x2212;2.5 and &#x2212;3.5 scoring 1, and &#x2264;&#x2212;3.5 scoring 2. Clinical symptoms are categorized as no significant pain (0 points), positional low back pain (1 point), or persistent pain or neurological symptoms (2 points). No significant pain refers to an absence of discomfort during daily activities, while positional low back pain is triggered by specific postures such as prolonged standing, sitting, or bending. Persistent pain is continuous and unrelieved by rest or posture changes, while neurological symptoms indicate nerve involvement, manifesting as numbness, tingling, or muscle weakness in the lower limbs. OVFs are classified based on total scores: Class 0 (&#x2264;3) for conservative treatment, Class 1 (=4) for either conservative or surgical treatment, and Class 2 (&#x2265;5) for surgical intervention. Evaluation scores were independently determined by 2 musculoskeletal radiologists (Doctor A with 6 years of experience and Doctor B with 10 years of experience), with disagreements resolved through discussion and consensus.</p></sec><sec id="s2-4"><title>Clinical and CT Images Evaluation</title><p>Patient information, including age, gender, dual-energy X-ray absorptiometry (DXA)-measured T-scores, and treatment details, was obtained from the clinical case management system. CT images were obtained using a bone window setting (width 1500 and level 500) and reconstructed with a 1-mm slice thickness for subsequent processing and analysis.</p></sec><sec id="s2-5"><title>Region of Interest Segmentation</title><p>Radiologists manually segmented fractured vertebrae using ITK-SNAP software (version 3.8.0; developed by the Penn Image Computing and Science Laboratory, University of Pennsylvania; lead developer: Paul A. Yushkevich) in this study. Radiologist A outlined and filled the edges of the fractured vertebrae on the CT images to create regions of interest, carefully excluding adjacent intervertebral discs, pedicles, and surrounding tissues to ensure precise delineation. The segmented masks were then saved as &#x201C;nii&#x201D; files for further analysis (<xref ref-type="fig" rid="figure3">Figure 3</xref>). Interobserver agreement was evaluated by having Radiologist A and Radiologist B independently resegment a random subset of 30 patients from the training dataset after 1 month, using the intraclass correlation coefficient (ICC) for assessment.</p><fig position="float" id="figure3"><label>Figure 3.</label><caption><p>Segmentation of a fractured vertebral body for radiomic analysis in an 82-year-old woman with an acute osteoporotic vertebral fracture is illustrated.(<bold>A</bold>) Sagittal non&#x2013;contrast-enhanced spine computed tomography images show an osteoporotic vertebral fracture of L4. (<bold>B</bold>) Sagittal T2-weighted fat-suppressed imaging reveals hyperintensity associated with the acute vertebral fracture. (<bold>C</bold>) The region of interest is delineated on sagittal computed tomography images. (<bold>D</bold>) Three-dimensional volume meshes are reconstructed to visualize the segmentation.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v13i1e75665_fig03.png"/></fig></sec><sec id="s2-6"><title>Radiomics and DTL Features Extraction</title><p>All images were resampled using B-spline interpolation and standardized with Z-score normalization to reduce variability across centers. Feature extraction algorithms were standardized in accordance with the Image Biomarker Standardization Initiative [<xref ref-type="bibr" rid="ref19">19</xref>]. Radiomic features, encompassing first-order, shape, and texture characteristics, were extracted using the open-source Python package PyRadiomics (developed by the Computational Imaging &#x0026; Bioinformatics Lab, Brigham and Women&#x2019;s Hospital/Harvard Medical School; lead developer: Joost J. M. van Griethuysen) [<xref ref-type="bibr" rid="ref20">20</xref>]. These texture features include the gray level co-occurrence matrix, gray level size zone matrix, gray level run length matrix, neighboring gray tone difference matrix, and gray level dependence matrix. For comprehensive details on the extracted features, refer to the PyRadiomics documentation [<xref ref-type="bibr" rid="ref21">21</xref>]. To minimize variations across centers, the Combat method was applied for feature harmonization [<xref ref-type="bibr" rid="ref22">22</xref>]. To mitigate bias and minimize overfitting risks from excessive features, a 2-step feature selection process was implemented: initially, features demonstrating strong reliability were retained using ICC evaluation, followed by further selection through the Least Absolute Shrinkage and Selection Operator (LASSO) algorithm.</p><p>Transfer learning is used because retraining a convolutional neural network for a specific task demands extensive image data and intricate parameter configurations, which are challenging to obtain in this study. Transfer learning involves fine-tuning a pretrained deep learning network to adapt it for a new task, allowing deep learning to be applied effectively on smaller datasets. Images were resampled to 64&#x00D7;64 (as is common practice in deep learning pipelines) and pixel intensities normalized to a mean of 0 and SD of 1. We acknowledge that resampling to 64&#x00D7;64 may lead to some loss of spatial detail. However, we chose this size after preliminary experiments demonstrated that it retained sufficient image features for accurate classification, while balancing computational efficiency and memory requirements. The DTL approach, akin to previous studies [<xref ref-type="bibr" rid="ref23">23</xref>], was implemented using the Python 3.6-based deep learning library (Guido van Rossum), PyTorch. The study used ResNet50 as the foundational model (<xref ref-type="supplementary-material" rid="app2">Multimedia Appendices 2</xref> and <xref ref-type="supplementary-material" rid="app3">3</xref>).</p><p>To execute transfer learning effectively, the learning rate was carefully configured. Features were extracted from the model&#x2019;s penultimate layer (AveragePooling), with model parameters divided into backbone and task-specific components. The backbone component used pretrained parameters from RadImageNet [<xref ref-type="bibr" rid="ref24">24</xref>] or ImageNet for initialization, whereas the task-specific component was initialized randomly. Drawing inspiration from the cosine annealing learning rate decay algorithm, optimizations were implemented by fine-tuning the backbone component with pretrained weights only when essential to maintain transfer learning quality. Concurrently, task-specific parameters were modified according to task demands, enabling the model to effectively adapt to the target data.</p></sec><sec id="s2-7"><title>Data Dimensionality Reduction</title><p>To identify reproducible and nonredundant radiomic features, a systematic process was implemented. First, features with ICC &#x2265;0.8 from 2 independent evaluations were retained for reproducibility [<xref ref-type="bibr" rid="ref25">25</xref>]. Redundancy was minimized by computing the Spearman rank correlation coefficient between features. Features with a correlation above 0.9 were subjected to a greedy recursive elimination strategy to remove the most redundant ones, ensuring overall representation was maintained. Stable features were then selected using the LASSO algorithm, which applies a penalty parameter (&#x03BB;) to shrink regression coefficients, retaining only relevant features. The optimal &#x03BB; value was determined through 10-fold cross-validation, and features with nonzero coefficients were selected for the final set. To further reduce redundancy, correlated features with a coefficient greater than 0.5 were excluded, resulting in a refined subset of independent features. For DTL features (initially with a dimension of 2048), principal component analysis was applied to reduce dimensionality, balancing deep learning and radiomic features while mitigating overfitting. The chosen radiomic and deep learning features were combined through early fusion to create a unified feature set, with all features standardized using Z-score normalization for compatibility. Finally, after fusion, LASSO-Cox regression was applied to select the most robust features, which were further refined through dimensionality reduction to define an optimal subset. This carefully curated feature set represented the most relevant combination of radiomic and deep learning features, facilitating reliable model development.</p></sec><sec id="s2-8"><title>Model Development</title><p>To prevent data leakage, all features used for building the predictive model were exclusively derived from the training set. Machine learning models were implemented using the scikit-learn library following feature selection and fusion. The models comprised logistic regression (LR), support vector machine (SVM), k-nearest neighbor (KNN), decision tree (DT), random forest (RF), extremely randomized trees (ExtraTrees), eXtreme gradient boosting (XGBoost), Light Gradient Boosting Machine (LightGBM), and multilayer perceptron (MLP). We observed an imbalance in the distribution of samples among the ASTLOF classification categories, with notably fewer cases in Class 2. To reduce the risk of biased model performance, we apply strategies such as class weighting during model training. Model training was conducted using the training set and optimized through grid search with adjustable parameters specific to each algorithm. Model performance was assessed using 5-fold cross-validation on the training data, selecting the best parameters to construct the optimal fused-feature model. Using a larger k (such as 10 or more) would have increased computational cost and training time substantially, without necessarily providing a significant improvement in model assessment, especially given the size of our dataset. Therefore, 5-fold cross-validation was appropriate for our study while maintaining a reasonable balance between thoroughness and practicality. The receiver operating characteristic (ROC) curve was plotted, and model accuracy was validated through 1000 iterations of bootstrap resampling. Performance metrics such as area under the curve (AUC), accuracy (ACC), sensitivity (SEN), and specificity (SPE) were evaluated. Finally, statistically significant clinical baseline characteristics were integrated with the best fused feature model to develop a combined model, which was visualized through a nomogram.</p><p>This study used the &#x201C;One-vs-Rest&#x201D; (OvR) strategy for multiclass tasks by decomposing the problem into several binary classification tasks. When Class 0 was labeled positive, Classes 1 and 2 were negative; similarly, labeling Class 1 or 2 as positive made the others negative, creating 3 OvR classification models. Model performance was assessed by plotting ROC curves and calculating metrics including AUC, ACC, SEN, and SPE. The generalization ability was evaluated using internal, external validation, and test datasets. Macro- and micro-average AUC were used for a thorough assessment of multiclass tasks [<xref ref-type="bibr" rid="ref26">26</xref>]. Macro-average AUC computes the AUC for each class and averages them equally, which can be less representative in cases of significant class imbalance. Conversely, micro-average AUC aggregates predictions from all classes into a single confusion matrix, emphasizing the influence of larger sample sizes and providing a better reflection of overall performance on imbalanced datasets.</p></sec><sec id="s2-9"><title>Data Analysis</title><p>Statistical analyses were conducted using R software (R Core Team; version 4.0.3), and radiomics and deep learning models were developed and implemented on Python 3.7 (Python Software Foundation). Continuous variables are presented as mean (SD), while categorical variables are shown as counts or percentages. Independent samples <italic>t</italic> tests were used to assess differences in continuous variables, while chi-square tests were applied for comparisons of categorical variables. The DeLong test was used to compare ROC curves and assess the predictive models&#x2019; overall performance. In addition, bootstrap validation with 1000 resamples was conducted to ensure robust evaluation of model accuracy. <italic>P</italic>&#x003C;.05 was considered statistically significant, serving as a benchmark for assessing the reliability of observed differences and associations. These comprehensive statistical methods ensured the rigor of model evaluation and the clarity of results interpretation.</p></sec><sec id="s2-10"><title>Explainable Artificial Intelligence</title><p>The Shapley Additive Explanations (SHAP) method (GitHub, Inc [<xref ref-type="bibr" rid="ref27">27</xref>]) was used to assess the importance of various features by calculating their contributions to prediction outcomes, offering a clear explanation of their significance [<xref ref-type="bibr" rid="ref28">28</xref>]. Using SHAP values, the predictive output for each sample is decomposed into individual feature contributions, providing a quantifiable measure of feature influence. The magnitude of a SHAP value indicates the extent of a feature&#x2019;s influence on the model&#x2019;s prediction, where positive values signify a positive impact and negative values signify a negative impact. For example, in a disease prediction model, a feature with a SHAP value greater than 0 suggests it increases the predicted likelihood of disease occurrence, whereas a value below 0 implies a reduced likelihood. Beyond individual predictions, SHAP also ranks features by their overall importance across the model and reveals the relationships between features and prediction outcomes. This integration of quantitative contribution, directional influence, and feature importance ranking facilitates a comprehensive understanding of the model&#x2019;s decision-making, revealing how particular features influence predictions and their significance.</p></sec><sec id="s2-11"><title>Ethical Considerations</title><p>This study comprised a retrospective component and a prospective validation cohort. For the retrospective component, the local Ethics Committees of the Affiliated Hospital of Nanjing University of Chinese Medicine and the Affiliated Taizhou People&#x2019;s Hospital of Nanjing Medical University waived the requirement for ethical approval and informed consent because the analysis involved existing data collected during routine clinical care and posed minimal risk to participants. The prospective validation cohort was approved by the Institutional Ethics Committee of Sir Run Run Hospital, Nanjing Medical University, on November 25, 2023 (approval no. 2023-SR-055). All participants in the prospective cohort provided written informed consent before enrollment. To protect privacy, all images and relevant data were deidentified prior to analysis and reporting. No individually identifiable information was used. Participants did not receive financial or other material compensation for participation. The study was conducted in accordance with the principles of the Declaration of Helsinki and relevant institutional guidelines and regulations.</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Clinical Features of the Studied Patients</title><p>The study enrolled 981 patients aged 50 to 95 years, with an average age of 69.56 (9.88) years. Of these, 687 were females (70%) and 294 were males (30%). Based on T-scores, 30 patients (3.1%) were classified as having normal bone mass, 257 (26.2%) as having low bone mass, and 694 (70.7%) as having osteoporosis. Among the participants, 87 patients presented with 2 OVFs, and 15 patients had 3 OVFs, resulting in a total of 1098 fractured vertebrae included in the analysis. The dataset was partitioned into a training set (750 cases, 68.4%), an internal validation set (187 cases, 17%), an external validation set (110 cases, 10%), and a prospective validation set (51 cases, 4.6%). <xref ref-type="table" rid="table1">Table 1</xref> summarizes the demographic and clinical characteristics of each dataset, and <xref ref-type="table" rid="table2">Table 2</xref> details the treatment conditions across the 3 classifications. <xref ref-type="fig" rid="figure4">Figure 4</xref> illustrates the case selection process, emphasizing the random 8:2 allocation of cases into the training and internal validation sets.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Baseline characteristics of patients with osteoporotic vertebral fracture in the training, internal and external validation, and prospective validation cohorts.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Characteristics</td><td align="left" valign="bottom">Training set<break/>(n=750)</td><td align="left" valign="bottom">Interval validation set<break/>(n=187)</td><td align="left" valign="bottom">External validation set<break/>(n=110)</td><td align="left" valign="bottom">Prospective validation set<break/>(n=51)</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="5">Sex, n (%)</td></tr><tr><td align="left" valign="top">&#x2003;Female</td><td align="left" valign="top">&#x2003;541 (72.1)</td><td align="left" valign="top">&#x2003;138 (73.8)</td><td align="left" valign="top">&#x2003;78 (70.9)</td><td align="left" valign="top">&#x2003;36 (70.6)</td></tr><tr><td align="left" valign="top">&#x2003;Male</td><td align="left" valign="top">&#x2003;209 (27.9)</td><td align="left" valign="top">&#x2003;49 (26.2)</td><td align="left" valign="top">&#x2003;32 (29.1)</td><td align="left" valign="top">&#x2003;15 (29.4)</td></tr><tr><td align="left" valign="top" colspan="5">Age (years)</td></tr><tr><td align="left" valign="top">&#x2003;Mean (SD)</td><td align="left" valign="top">&#x2003;68.25 (11.18)</td><td align="left" valign="top">&#x2003;70.19 (10.56)</td><td align="left" valign="top">&#x2003;69.56 (10.23)</td><td align="left" valign="top">&#x2003;69.51 (10.32)</td></tr><tr><td align="left" valign="top" colspan="5">DXA<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup> T-score</td></tr><tr><td align="left" valign="top">&#x2003;Mean (SD)</td><td align="left" valign="top">&#x2003;&#x2212;2.82 (0.82)</td><td align="left" valign="top">&#x2003;&#x2212;2.79 (0.81)</td><td align="left" valign="top">&#x2003;&#x2212;2.85 (0.75)</td><td align="left" valign="top">&#x2003;&#x2212;2.83 (0.77)</td></tr><tr><td align="left" valign="top" colspan="5">Fracture location, n (%)</td></tr><tr><td align="left" valign="top">&#x2003;Thoracic</td><td align="left" valign="top">&#x2003;247 (32.9)</td><td align="left" valign="top">&#x2003;52 (27.8)</td><td align="left" valign="top">&#x2003;32 (29.1)</td><td align="left" valign="top">&#x2003;17 (33.3)</td></tr><tr><td align="left" valign="top">&#x2003;Lumbar</td><td align="left" valign="top">&#x2003;503 (67.1)</td><td align="left" valign="top">&#x2003;135 (72.2)</td><td align="left" valign="top">&#x2003;78 (70.9)</td><td align="left" valign="top">&#x2003;34 (66.7)</td></tr><tr><td align="left" valign="top" colspan="5">Fracture staging, n (%)</td></tr><tr><td align="left" valign="top">&#x2003;Acute</td><td align="left" valign="top">&#x2003;492 (65.6)</td><td align="left" valign="top">&#x2003;112 (59.9)</td><td align="left" valign="top">&#x2003;69 (62.7)</td><td align="left" valign="top">&#x2003;33 (64.7)</td></tr><tr><td align="left" valign="top">&#x2003;Chronic</td><td align="left" valign="top">&#x2003;258 (34.4)</td><td align="left" valign="top">&#x2003;75 (40.1)</td><td align="left" valign="top">&#x2003;41 (37.3)</td><td align="left" valign="top">&#x2003;18 (35.3)</td></tr><tr><td align="left" valign="top" colspan="5">ASTLOF<sup><xref ref-type="table-fn" rid="table1fn2">b</xref></sup> score, n (%)</td></tr><tr><td align="left" valign="top">&#x2003;1&#x2010;3 points</td><td align="left" valign="top">&#x2003;345 (46.0)</td><td align="left" valign="top">&#x2003;88 (47.1)</td><td align="left" valign="top">&#x2003;51 (46.4)</td><td align="left" valign="top">&#x2003;24 (47.1)</td></tr><tr><td align="left" valign="top">&#x2003;4 points</td><td align="left" valign="top">&#x2003;338 (45.1)</td><td align="left" valign="top">&#x2003;76 (40.6)</td><td align="left" valign="top">&#x2003;46 (41.8)</td><td align="left" valign="top">&#x2003;22 (43.1)</td></tr><tr><td align="left" valign="top">&#x2003;5&#x2010;8 points</td><td align="left" valign="top">&#x2003;67 (8.9)</td><td align="left" valign="top">&#x2003;23 (12.3)</td><td align="left" valign="top">&#x2003;13 (11.8)</td><td align="left" valign="top">&#x2003;5 (9.8)</td></tr><tr><td align="left" valign="top" colspan="5">Therapeutic method, n (%)</td></tr><tr><td align="left" valign="top">&#x2003;Conservative treatment</td><td align="left" valign="top">&#x2003;435 (58)</td><td align="left" valign="top">&#x2003;106 (56.7)</td><td align="left" valign="top">&#x2003;62 (56.3)</td><td align="left" valign="top">&#x2003;28 (54.9)</td></tr><tr><td align="left" valign="top">&#x2003;PVA<sup><xref ref-type="table-fn" rid="table1fn3">c</xref></sup></td><td align="left" valign="top">&#x2003;258 (34.4)</td><td align="left" valign="top">&#x2003;73 (39)</td><td align="left" valign="top">&#x2003;40 (36.4)</td><td align="left" valign="top">&#x2003;19 (37.3)</td></tr><tr><td align="left" valign="top">&#x2003;Open surgery</td><td align="left" valign="top">&#x2003;57 (7.6)</td><td align="left" valign="top">&#x2003;8 (4.3)</td><td align="left" valign="top">&#x2003;8 (7.3)</td><td align="left" valign="top">&#x2003;4 (7.8)</td></tr></tbody></table><table-wrap-foot><fn id="table1fn1"><p><sup>a</sup>DXA: dual-energy X-ray absorptiometry.</p></fn><fn id="table1fn2"><p><sup>b</sup>ASTLOF: Assessment System of Thoracolumbar Osteoporotic Fracture.</p></fn><fn id="table1fn3"><p><sup>c</sup>PVA: percutaneous vertebral augmentation.</p></fn></table-wrap-foot></table-wrap><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>Distribution of osteoporotic vertebral fractures based on the Assessment System of Thoracolumbar Osteoporotic Fractures classification and their association with different therapeutic methods.</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Classification</td><td align="left" valign="bottom">Conservative treatment (n=631, %)</td><td align="left" valign="bottom">PVA<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup> (n=390, %)</td><td align="left" valign="bottom">Open surgery (n=77, %)</td></tr></thead><tbody><tr><td align="left" valign="top">Class 0 (1-3 points)</td><td align="left" valign="top">411 (80.9)</td><td align="left" valign="top">71 (14)</td><td align="left" valign="top">26 (5.1)</td></tr><tr><td align="left" valign="top">Class 1 (4 points)</td><td align="left" valign="top">162 (33.6)</td><td align="left" valign="top">289 (60)</td><td align="left" valign="top">31 (6.4)</td></tr><tr><td align="left" valign="top">Class 2 (5-8 points)</td><td align="left" valign="top">58 (53.7)</td><td align="left" valign="top">30 (27.8)</td><td align="left" valign="top">20 (18.5)</td></tr></tbody></table><table-wrap-foot><fn id="table2fn1"><p><sup>a</sup>PVA, percutaneous vertebral augmentation.</p></fn></table-wrap-foot></table-wrap><fig position="float" id="figure4"><label>Figure 4.</label><caption><p>The flowchart summarizes patient selection and allocation to the training set, internal and external validation set, and prospective validation set of the multicenter study. CT: computed tomography; MRI: magnetic resonance imaging; OVFs: osteoporotic vertebral fractures.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v13i1e75665_fig04.png"/></fig></sec><sec id="s3-2"><title>Radiomics Feature Selection (RadImageNet-Based)</title><p>The LASSO-Cox regression analysis model was used to perform dimensionality reduction on the fused features. The penalty coefficient (&#x03BB;=0.0031) was chosen to optimize feature selection, with <xref ref-type="supplementary-material" rid="app4">Multimedia Appendix 4</xref> depicting the changes in feature coefficients as &#x03BB; varied. Following the final feature selection, 17 radiomics features and 16 DTL features were retained. The DTL_Radscore was constructed using the fused features and their regression coefficients, as shown in <xref ref-type="supplementary-material" rid="app5">Multimedia Appendix 5</xref>.</p></sec><sec id="s3-3"><title>Radiomics Feature Selection (ImageNet-Based)</title><p>The LASSO-Cox regression model, with a penalty coefficient of &#x03BB;=0.0295, was used to optimally select features by reducing the dimensionality of the fused dataset. <xref ref-type="supplementary-material" rid="app6">Multimedia Appendix 6</xref> displays the feature selection process and the curve showing the change in feature coefficients with &#x03BB;. Following the final selection, 17 radiomics features and 37 DTL features were retained. Using these fused features and their associated regression coefficients, the DTL_Radscore was constructed, as detailed in <xref ref-type="supplementary-material" rid="app7">Multimedia Appendix 7</xref>.</p></sec><sec id="s3-4"><title>Overall Validation of Different Radiomics Models</title><p>The optimal machine learning algorithms for fused feature models trained on RadImageNet and ImageNet datasets, based on macro-average AUC, ACC, and <italic>F</italic><sub>1</sub>-score, were identified as MLP and LightGBM, respectively. <xref ref-type="table" rid="table3">Table 3</xref> summarizes the validation results for the 2 fused feature models in the 3-class classification task. In the training set, the DeLong test indicated no statistically significant difference between the 2 fused feature models (0.934 vs 0.996, <italic>P</italic>=2.34). In the internal, external, and prospective validation sets, the RadImageNet-based fused feature model demonstrated significantly higher macro-average AUC values than the ImageNet-based model (0.837 vs 0.648, 0.773 vs 0.633, and 0.852 vs 0.648, respectively), as confirmed by the DeLong test (<italic>P</italic>&#x003C;.05). <xref ref-type="fig" rid="figure5">Figure 5</xref> displays the ROC curves for both models predicting OVF classifications in the prospective validation set. The RadImageNet-based fused feature model, using the binary OvR strategy, excelled in predicting classification 2 with an AUC of 0.907 and an ACC of 0.857. For classifications 0 and 1, the model achieved AUCs and ACCs of 0.829, 0.803 and 0.794, 0.768, respectively. <xref ref-type="fig" rid="figure6">Figure 6</xref> highlights instances where the ImageNet-based fused feature model made incorrect predictions, while the RadImageNet-based model successfully identified the correct classifications.</p><table-wrap id="t3" position="float"><label>Table 3.</label><caption><p>The performance of the models across the training set, internal and external validation sets, and the prospective validation set.</p></caption><table id="table3" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom" rowspan="2">Model</td><td align="left" valign="bottom" colspan="2">Training set</td><td align="left" valign="bottom" colspan="2">Interval validation set</td><td align="left" valign="bottom" colspan="2">External validation set</td><td align="left" valign="bottom" colspan="2">Prospective validation set</td></tr><tr><td align="left" valign="bottom">Accuracy</td><td align="left" valign="bottom">AUC<sup><xref ref-type="table-fn" rid="table3fn1">a</xref></sup></td><td align="left" valign="bottom">Accuracy</td><td align="left" valign="bottom">AUC<sup><xref ref-type="table-fn" rid="table3fn1">a</xref></sup></td><td align="left" valign="bottom">Accuracy</td><td align="left" valign="bottom">AUC<sup><xref ref-type="table-fn" rid="table3fn1">a</xref></sup></td><td align="left" valign="bottom">Accuracy</td><td align="left" valign="bottom">AUC<sup><xref ref-type="table-fn" rid="table3fn1">a</xref></sup></td></tr></thead><tbody><tr><td align="left" valign="top" colspan="9">RadImageNet-based</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Class 0</td><td align="left" valign="top">0.867</td><td align="left" valign="top">0.939<break/>(0.924&#x2010;0.955)</td><td align="left" valign="top">0.777</td><td align="left" valign="top">0.834<break/>(0.779&#x2010;0.889)</td><td align="left" valign="top">0.715</td><td align="left" valign="top">0.794<break/>(0.714&#x2010;0.875)</td><td align="left" valign="top">0.803</td><td align="left" valign="top">0.829<break/>(0.719&#x2010;0.938)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Class 1</td><td align="left" valign="top">0.825</td><td align="left" valign="top">0.905<break/>(0.884&#x2010;0.925)</td><td align="left" valign="top">0.726</td><td align="left" valign="top">0.768<break/>(0.700&#x2010;0.836)</td><td align="left" valign="top">0.681</td><td align="left" valign="top">0.747<break/>(0.658&#x2010;0.836)</td><td align="left" valign="top">0.768</td><td align="left" valign="top">0.794<break/>(0.673&#x2010;0.915)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Class 2</td><td align="left" valign="top">0.886</td><td align="left" valign="top">0.953<break/>(0.934&#x2010;0.973)</td><td align="left" valign="top">0.746</td><td align="left" valign="top">0.898<break/>(0.839&#x2010;0.957)</td><td align="left" valign="top">0.767</td><td align="left" valign="top">0.756<break/>(0.593&#x2010;0.920)</td><td align="left" valign="top">0.857</td><td align="left" valign="top">0.907<break/>(0.805&#x2010;0.999)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Three classifications<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td><td align="left" valign="top">0.793</td><td align="left" valign="top">0.934<break/>(0.914&#x2010;0.951)</td><td align="left" valign="top">0.660</td><td align="left" valign="top">0.837<break/>(0.773&#x2010;0.894)</td><td align="left" valign="top">0.647</td><td align="left" valign="top">0.773<break/>(0.655&#x2010;0.877)</td><td align="left" valign="top">0.732</td><td align="left" valign="top">0.852<break/>(0.732&#x2010;0.951)</td></tr><tr><td align="left" valign="top" colspan="9">ImageNet-based</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Class 0</td><td align="left" valign="top">0.969</td><td align="left" valign="top">0.995<break/>(0.992&#x2010;0.997)</td><td align="left" valign="top">0.619</td><td align="left" valign="top">0.619<break/>(0.540&#x2010;0.698)</td><td align="left" valign="top">0.655</td><td align="left" valign="top">0.675<break/>(0.576&#x2010;0.774)</td><td align="left" valign="top">0.625</td><td align="left" valign="top">0.586<break/>(0.433&#x2010;0.739)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Class 1</td><td align="left" valign="top">0.964</td><td align="left" valign="top">0.996<break/>(0.994&#x2010;0.999)</td><td align="left" valign="top">0.624</td><td align="left" valign="top">0.576<break/>(0.488&#x2010;0.664)</td><td align="left" valign="top">0.560</td><td align="left" valign="top">0.551<break/>(0.445&#x2010;0.656)</td><td align="left" valign="top">0.607</td><td align="left" valign="top">0.545<break/>(0.385&#x2010;0.705)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Class 2</td><td align="left" valign="top">0.952</td><td align="left" valign="top">0.995<break/>(0.993&#x2010;0.999)</td><td align="left" valign="top">0.756</td><td align="left" valign="top">0.737<break/>(0.631&#x2010;0.843)</td><td align="left" valign="top">0.621</td><td align="left" valign="top">0.654<break/>(0.480&#x2010;0.827)</td><td align="left" valign="top">0.803</td><td align="left" valign="top">0.767<break/>(0.580&#x2010;0.953)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Three classifications<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td><td align="left" valign="top">0.916</td><td align="left" valign="top">0.996<break/>(0.993&#x2010;0.998)</td><td align="left" valign="top">0.533</td><td align="left" valign="top">0.648<break/>(0.553&#x2010;0.735)</td><td align="left" valign="top">0.551</td><td align="left" valign="top">0.633<break/>(0.501&#x2010;0.753)</td><td align="left" valign="top">0.429</td><td align="left" valign="top">0.648<break/>(0.466&#x2010;0.799)</td></tr></tbody></table><table-wrap-foot><fn id="table3fn1"><p><sup>a</sup>Data in parentheses are 95% CIs.</p></fn><fn id="table3fn2"><p><sup>b</sup>Date are macro-average. </p></fn></table-wrap-foot></table-wrap><fig position="float" id="figure5"><label>Figure 5.</label><caption><p>The receiver operating characteristic curves for the predictive performance of the 2 models (A. RadImageNet, B. ImageNet). AUC: area under the curve; MLP: multilayer perceptron; LightGBM: Light Gradient Boosting Machine.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v13i1e75665_fig05.png"/></fig><fig position="float" id="figure6"><label>Figure 6.</label><caption><p>A case from the prospective validation cohort involves a 60-year-old female patient with osteoporotic vertebral fractures and an Assessment System of Thoracolumbar Osteoporotic Fractures score of 6. It was misclassified by the ImageNet model, but was correctly classified by the RadImageNet model. (<bold>A</bold>) Computed tomography imaging; (<bold>B</bold>) MRI imaging; (<bold>C</bold>) Postpercutaneous vertebroplasty showing bone cement leakage.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v13i1e75665_fig06.png"/></fig></sec><sec id="s3-5"><title>Feature Contribution and Model Interpretation</title><p>The SHAP value for each feature was calculated. <xref ref-type="fig" rid="figure7">Figure 7</xref> presents the global SHAP values for both the overall 3-class classification and each specific class, evaluating their impact on the model&#x2019;s predictions. The highest-ranked features were cluster shade, mean, and large area low gray level emphasis. <xref ref-type="fig" rid="figure8">Figure 8</xref> presents the SHAP decision plot, illustrating the prediction model&#x2019;s workflow in classifying Class 0 (male, 65 years; ASTLOF 2 points), Class 1 (female, 72 years; ASTLOF 4 points), and Class 2 (female, 68 years; ASTLOF 6 points) within the prospective validation set.</p><fig position="float" id="figure7"><label>Figure 7.</label><caption><p>The feature contributions of the optimal fusion model are visualized as follows: The y-axis displays features arranged in descending order based on their mean absolute impact on the predictive model. The Shapley Additive Explanations (SHAP) value of a specific feature is represented by its distance from x=0, where a greater distance signifies a stronger impact&#x2014;either positive or negative&#x2014;on the model&#x2019;s output. Each point&#x2019;s color corresponds to the original value of that feature, transitioning from low (blue) to high (magenta) on the color scale. (<bold>A</bold>) The global feature contribution bar chart illustrates the contributions for the 3-class classification, with blue, red, and dark green bars indicating classifications 0, 1, and 2, respectively. (<bold>B, C, D</bold>) Beehive summary plots depict the decreasing feature contributions for predictions corresponding to classifications 0, 1, and 2, respectively.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v13i1e75665_fig07.png"/></fig><fig position="float" id="figure8"><label>Figure 8.</label><caption><p>Shapley Additive Explanation (SHAP) decision plots for the 3 classifications (A: Class 0, B: Class 1, C: Class 2) are presented. The x-axis represents the model output, while the y-axis lists the feature names. The gray vertical line at the center shows the baseline value. Each line traces the prediction process, starting from the baseline value and incorporating the contributions of various features, both positive and negative, to arrive at the final model output. For classification 0, the baseline value is 0.468, with a final model output of 0.090. For classification 1, the baseline value is 0.423, with a final model output of 0.138. For classification 2, the baseline value is 0.108, with a final model output of 0.062.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v13i1e75665_fig08.png"/></fig></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Principal Findings</title><p>This study developed and validated a DLR model based on CT imaging data from multiple medical centers for the classification of OVFs according to the ASTLOF system. By integrating radiomics and DTL features extracted from both RadImageNet and ImageNet datasets, the fused model&#x2014;especially when using RadImageNet pretraining&#x2014;demonstrated superior predictive performance across internal, external, and prospective validation cohorts. The model achieved the highest accuracy in identifying Class 2, with SHAP analysis indicating that features such as cluster shade, mean, and large area low gray level emphasis played the most significant roles in prediction. These findings highlight the model&#x2019;s robustness and generalizability, supporting its potential utility in guiding clinical decision-making for OVF classification and treatment planning.</p></sec><sec id="s4-2"><title>Study Implications</title><p>Compared with obvious traumatic vertebral fractures, OVFs are an insidious condition and are often misdiagnosed. Improper treatment can affect spinal stability and balance, and in severe cases, lead to neurological dysfunction and increased risk of mortality [<xref ref-type="bibr" rid="ref29">29</xref>]. A scientific classification of OVFs is the prerequisite for appropriate treatment. However, existing classification methods are primarily based on classification systems for early thoracolumbar fractures (which do not differentiate between traumatic and osteoporotic vertebral fractures), resulting in confusion in the treatment of OVFs [<xref ref-type="bibr" rid="ref30">30</xref>]. These methods fail to adequately consider the characteristics of osteoporotic vertebrae, are overly complex, and lack widely accepted unified standards. Some even overly emphasize surgical treatment. A systematic classification of OVFs is crucial for assessing fracture risk, guiding treatment decisions, and forecasting patient outcomes [<xref ref-type="bibr" rid="ref31">31</xref>]. An ideal classification system for OVFs should encompass several essential features to ensure comprehensive and practical utility. First, it should integrate imaging parameters from X-rays, CT, and MRI, enabling a thorough multiperspective assessment of the fractures. Second, it must incorporate patients&#x2019; clinical presentations, such as lower back pain and neurological symptoms, to provide a holistic understanding of the condition. Third, the system should offer treatment guidance tailored to each classification type, facilitating targeted clinical interventions. Fourth, it is essential that the system demonstrates high reliability and reproducibility, ensuring consistent application across different clinical settings. Finally, it should effectively evaluate the severity of the condition and provide prognostic insights based on classification outcomes. The ASTLOF classification system provides a thorough framework for evaluation by integrating vertebral morphology, MRI signal characteristics, BMD, and clinical symptoms. Through its scoring system, it enables clinicians to select targeted treatment plans, streamlining clinical workflows while delivering significant clinical guidance. Furthermore, existing studies have validated the system&#x2019;s high consistency and reproducibility, reinforcing its effectiveness in guiding clinical treatment decisions.</p><p>Recent advancements in artificial intelligence have shifted OVF classification research toward detection, with studies showing that deep learning and radiomics methods significantly surpass traditional visual analysis approaches [<xref ref-type="bibr" rid="ref32">32</xref>]. Most current research on OVF classification detection primarily uses single-center data split into training and validation sets for internal validation. This method is constrained by significant variability in radiomics analysis results due to differences in imaging techniques, postprocessing, reconstruction workflows, and scanning parameters across devices from various manufacturers [<xref ref-type="bibr" rid="ref33">33</xref>]. In addition, single-center studies often lack data heterogeneity, increasing the risk of overfitting and reducing the generalizability of the findings. By contrast, multicenter studies leverage diverse imaging data, and predictive models validated using independent external datasets better account for the heterogeneity of OVFs, offering results that are more aligned with the principles of precision medicine. Our study&#x2019;s strength is the use of CT imaging data from various hospitals combined with the ASTLOF classification system. A fused predictive model integrating radiomics and DTL features was developed using datasets such as RadImageNet and ImageNet. The model was thoroughly evaluated for its predictive performance in OVF classifications, offering a robust and generalizable framework for clinical application.</p><p>Studies indicate that the RadImageNet dataset notably improves DTL performance in medical applications, offering superior generalization over conventional datasets [<xref ref-type="bibr" rid="ref34">34</xref>]. Our study&#x2019;s findings confirmed that prediction models using the RadImageNet dataset surpassed those using the ImageNet dataset. In situations where sample sizes are imbalanced, the &#x201C;OvR&#x201D; strategy is commonly used for 3-class classification tasks [<xref ref-type="bibr" rid="ref35">35</xref>]. In this study, the use of the &#x201C;OvR&#x201D; strategy in prediction models for CT images proved to be highly effective. Notably, classification 2, despite having a smaller sample size, was identified with the highest accuracy. The enhanced spatial and density resolution of CT images enables prediction models to more effectively identify radiomic and DTL features. These results highlight the potential of leveraging high-resolution imaging data and advanced datasets such as RadImageNet to achieve robust and accurate predictions, even under the challenge of imbalanced samples.</p></sec><sec id="s4-3"><title>Comparison to Prior Work</title><p>Finally, our study used SHAP values to evaluate the importance of features. SHAP values indicate the positive or negative contributions of each predictive variable to the target variable [<xref ref-type="bibr" rid="ref36">36</xref>]. Based on game theory, SHAP is a classical post hoc explanation framework used to analyze typically incomprehensible black box models. Aggregating SHAP values across features offers a comprehensive view of each feature&#x2019;s impact on the model&#x2019;s predictions, clearly explaining the decision-making process. In this study, the feature with the highest contribution in CT images was cluster shade, which measures the skewness and asymmetry of the intensity distribution of grayscale in an image. It is inversely proportional to the number of asymmetric densities in the image. High skewness in the co-occurrence matrix results in lower cluster shade values, while smaller cluster shade values suggest greater homogeneity in the distribution of lesions [<xref ref-type="bibr" rid="ref37">37</xref>]. We hypothesize that a smaller cluster shade value indicates a more homogeneous distribution of lesions within the vertebral body. While our model incorporates SHAP analysis to provide post hoc interpretability, we recognize that the deep learning component remains largely a &#x201C;black box,&#x201D; which may limit clinician trust and acceptance. We acknowledge the importance of further enhancing the model&#x2019;s explainability&#x2014;particularly in elucidating how deep features correspond to specific anatomical or pathological findings relevant to ASTLOF classification. In future work, we intend to explore and integrate advanced interpretability techniques such as attention maps, layer-wise relevance propagation, and feature visualization. These methods have the potential to provide more granular and intuitive explanations for the model&#x2019;s predictions, thereby facilitating broader clinical adoption and understanding.</p></sec><sec id="s4-4"><title>Limitations</title><p>Although this study has achieved certain results, it still has some limitations. First, a key limitation of our study is the unequal distribution of cases among the ASTLOF classifications, particularly the small sample size for Class 2. This class imbalance may affect the statistical power and generalizability of the model for underrepresented classes. Moving forward, we intend to increase the sample size for each class and explore robust solutions such as synthetic oversampling, class weighting, or other augmentation strategies to enhance model performance and clinical applicability across all categories. Second, the lack of interpretability of deep learning features limits its widespread clinical adoption and trust to some extent. Moving forward, strengthening research on the interpretability of deep learning features will be crucial. This will improve model transparency, foster clinician trust, and guide clinical decision-making, thus enhancing the practical application of deep learning in medical imaging diagnosis and treatment planning. Third, our study&#x2019;s dataset was exclusively sourced from Chinese hospitals, which may introduce geographic or ethnic bias, potentially limiting the generalizability of the findings to other regions and populations. Imaging protocols, equipment, and patient demographics may differ significantly across health care systems worldwide. Future work will focus on expanding our dataset with more diverse, multinational samples and performing external validation in independent international cohorts. Such steps are essential for demonstrating the model&#x2019;s robustness and ensuring clinical applicability on a global scale.</p></sec><sec id="s4-5"><title>Conclusions</title><p>Compared to the fusion feature model (ImageNet), the fusion feature model based on CT images (RadImageNet) demonstrated higher predictive performance. Notably, it achieved the best performance in identifying classification 2, followed by classifications 0 and 1. This may have significant clinical value for predicting OVF classifications and guiding the formulation of treatment plans.</p></sec></sec></body><back><ack><p>We express our gratitude to the editor and anonymous reviewers for their insightful comments and suggestions, which have greatly enhanced the quality of our paper. We also extend our thanks to American Journal Experts for their assistance in editing the language of an earlier draft of this manuscript. We appreciate the guidance and support from the PixelmedAI platform and its developers concerning the code used in this revised manuscript. Finally, we thank Dr. Zihan Chen (School of Mathematics and Physics, Xi&#x2019;an Jiaotong-Liverpool University) for his professional statistical advice. This work was supported by the Medical Scientific Research Project of the Jiangsu Provincial Health Commission (Grant No. M2024055) and the Medical Imaging Artificial Intelligence Special Research Fund Project, Nanjing Medical Association Radiology Branch (Grant No. 16).</p></ack><notes><sec><title>Data Availability</title><p>The datasets generated or analyzed during this study are not publicly available due to privacy or ethical restrictions, but are available from the corresponding author on reasonable request.</p></sec></notes><fn-group><fn fn-type="con"><p>LX and JZ contributed equally as co-corresponding authors.</p><p>JL contributed to conceptualization, investigation, resource provision, original draft writing, review and editing of the manuscript, and visualization. LZ contributed to conceptualization, investigation, resource provision, and review and editing of the manuscript. YY contributed to conceptualization, investigation, original draft writing, review and editing of the manuscript, and visualization. JT contributed to conceptualization, methodology, review and editing of the manuscript, and project administration. YL contributed to conceptualization, review and editing of the manuscript, visualization, and supervision. LX contributed to conceptualization, methodology, software development, review and editing of the manuscript, and supervision. JZ contributed to conceptualization, investigation, resource provision, review and editing of the manuscript, visualization, and funding acquisition.</p></fn><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">ACC</term><def><p>accuracy</p></def></def-item><def-item><term id="abb2">ASTLOF</term><def><p>Assessment System of Thoracolumbar Osteoporotic Fractures</p></def></def-item><def-item><term id="abb3">AUC</term><def><p>area under the curve</p></def></def-item><def-item><term id="abb4">BMD</term><def><p>bone mineral density</p></def></def-item><def-item><term id="abb5">CT</term><def><p>computed tomography</p></def></def-item><def-item><term id="abb6">DLR</term><def><p>deep learning radiomics</p></def></def-item><def-item><term id="abb7">DT</term><def><p>decision tree</p></def></def-item><def-item><term id="abb8">DTL</term><def><p>deep transfer learning</p></def></def-item><def-item><term id="abb9">DXA</term><def><p>dual-energy X-ray absorptiometry</p></def></def-item><def-item><term id="abb10">ExtraTrees</term><def><p>extremely randomized trees</p></def></def-item><def-item><term id="abb11">ICC</term><def><p>intraclass correlation coefficients</p></def></def-item><def-item><term id="abb12">KNN</term><def><p>k-nearest neighbor</p></def></def-item><def-item><term id="abb13">LASSO</term><def><p>Least Absolute Shrinkage and Selection Operator</p></def></def-item><def-item><term id="abb14">LightGBM</term><def><p>Light Gradient Boosting Machine</p></def></def-item><def-item><term id="abb15">LR</term><def><p>logistic regression</p></def></def-item><def-item><term id="abb16">MLP</term><def><p>multilayer perceptron</p></def></def-item><def-item><term id="abb17">MRI</term><def><p>magnetic resonance imaging</p></def></def-item><def-item><term id="abb18">OVFs</term><def><p>osteoporotic vertebral fractures</p></def></def-item><def-item><term id="abb19">OvR</term><def><p>One-vs-Rest</p></def></def-item><def-item><term id="abb20">RF</term><def><p>random forest</p></def></def-item><def-item><term id="abb21">ROC</term><def><p>receiver operating characteristic</p></def></def-item><def-item><term id="abb22">SEN</term><def><p>sensitivity</p></def></def-item><def-item><term id="abb23">SHAP</term><def><p>SHapley Additive exPlanations</p></def></def-item><def-item><term id="abb24">SPE</term><def><p>specificity</p></def></def-item><def-item><term id="abb25">STIR</term><def><p>short tau inversion recovery</p></def></def-item><def-item><term id="abb26">SVM</term><def><p>support vector machine</p></def></def-item><def-item><term id="abb27">XGBoost</term><def><p>Extreme gradient boosting</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zeng</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Li</surname><given-names>N</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>Q</given-names> </name><etal/></person-group><article-title>The prevalence of osteoporosis in China, a nationwide, multicenter DXA survey</article-title><source>J Bone Miner Res</source><year>2019</year><month>10</month><volume>34</volume><issue>10</issue><fpage>1789</fpage><lpage>1797</lpage><pub-id pub-id-type="doi">10.1002/jbmr.3757</pub-id><pub-id pub-id-type="medline">31067339</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Skj&#x00F8;dt</surname><given-names>MK</given-names> </name><name name-style="western"><surname>Abrahamsen</surname><given-names>B</given-names> </name></person-group><article-title>New insights in the pathophysiology, epidemiology, and response to treatment of osteoporotic vertebral fractures</article-title><source>J Clin Endocrinol Metab</source><year>2023</year><month>10</month><day>18</day><volume>108</volume><issue>11</issue><fpage>e1175</fpage><lpage>e1185</lpage><pub-id pub-id-type="doi">10.1210/clinem/dgad256</pub-id><pub-id pub-id-type="medline">37186550</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Schnake</surname><given-names>KJ</given-names> </name><name name-style="western"><surname>Blattert</surname><given-names>TR</given-names> </name><name name-style="western"><surname>Hahn</surname><given-names>P</given-names> </name><etal/></person-group><article-title>Classification of osteoporotic thoracolumbar spine fractures: recommendations of the spine section of the German Society for Orthopaedics and Trauma (DGOU)</article-title><source>Global Spine J</source><year>2018</year><month>09</month><volume>8</volume><issue>2 Suppl</issue><fpage>46S</fpage><lpage>49S</lpage><pub-id pub-id-type="doi">10.1177/2192568217717972</pub-id><pub-id pub-id-type="medline">30210960</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Dong</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Luo</surname><given-names>G</given-names> </name><name name-style="western"><surname>Lane</surname><given-names>NE</given-names> </name><etal/></person-group><article-title>Deep learning classification of spinal osteoporotic compression fractures on radiographs using an adaptation of the genant semiquantitative criteria</article-title><source>Acad Radiol</source><year>2022</year><month>12</month><volume>29</volume><issue>12</issue><fpage>1819</fpage><lpage>1832</lpage><pub-id pub-id-type="doi">10.1016/j.acra.2022.02.020</pub-id><pub-id pub-id-type="medline">35351363</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Heini</surname><given-names>PF</given-names> </name></person-group><article-title>The current treatment--a survey of osteoporotic fracture treatment. Osteoporotic spine fractures: the spine surgeon&#x2019;s perspective</article-title><source>Osteoporos Int</source><year>2005</year><month>03</month><volume>16 Suppl 2</volume><issue>S02</issue><fpage>S85</fpage><lpage>92</lpage><pub-id pub-id-type="doi">10.1007/s00198-004-1723-1</pub-id><pub-id pub-id-type="medline">15365699</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sch&#x00F6;nrogge</surname><given-names>M</given-names> </name><name name-style="western"><surname>Lahodski</surname><given-names>V</given-names> </name><name name-style="western"><surname>Otto</surname><given-names>R</given-names> </name><etal/></person-group><article-title>Inter- and intraobserver reliabilities and critical analysis of the osteoporotic fracture classification of osteoporotic vertebral body fractures</article-title><source>Eur Spine J</source><year>2022</year><month>09</month><volume>31</volume><issue>9</issue><fpage>2431</fpage><lpage>2438</lpage><pub-id pub-id-type="doi">10.1007/s00586-022-07201-2</pub-id><pub-id pub-id-type="medline">35378632</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Du</surname><given-names>JP</given-names> </name><name name-style="western"><surname>Fan</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>JJ</given-names> </name><etal/></person-group><article-title>The analysis of MSTMOVCF (Multi-segment thoracolumbar mild osteoporotic fractures surgery or conservative treatment) based on ASTLOF (the assessment system of thoracolumbar osteoporotic fracture)</article-title><source>Sci Rep</source><year>2018</year><month>05</month><day>29</day><volume>8</volume><issue>1</issue><fpage>8185</fpage><pub-id pub-id-type="doi">10.1038/s41598-018-26562-7</pub-id><pub-id pub-id-type="medline">29844542</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hao</surname><given-names>DJ</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>JS</given-names> </name><name name-style="western"><surname>Tuo</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>Reliability and application of the new morphological classification system for chronic symptomatic osteoporotic thoracolumbar fracture</article-title><source>J Orthop Surg Res</source><year>2020</year><month>08</month><day>24</day><volume>15</volume><issue>1</issue><fpage>348</fpage><pub-id pub-id-type="doi">10.1186/s13018-020-01882-5</pub-id><pub-id pub-id-type="medline">32831125</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Du</surname><given-names>JP</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>JJ</given-names> </name><name name-style="western"><surname>Fan</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>Surgery for multisegment thoracolumbar mild osteoporotic fractures: revised assessment system of thoracolumbar osteoporotic fracture</article-title><source>World Neurosurg</source><year>2018</year><month>06</month><volume>114</volume><fpage>e969</fpage><lpage>e975</lpage><pub-id pub-id-type="doi">10.1016/j.wneu.2018.03.122</pub-id><pub-id pub-id-type="medline">29588238</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Rosenberg</surname><given-names>GS</given-names> </name><name name-style="western"><surname>Cina</surname><given-names>A</given-names> </name><name name-style="western"><surname>Schir&#x00F3;</surname><given-names>GR</given-names> </name><etal/></person-group><article-title>Artificial intelligence accurately detects traumatic thoracolumbar fractures on sagittal radiographs</article-title><source>Med Bogota Colomb</source><year>2022</year><volume>58</volume><issue>8</issue><fpage>998</fpage><pub-id pub-id-type="doi">10.3390/medicina58080998</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>He</surname><given-names>L</given-names> </name><name name-style="western"><surname>Yu</surname><given-names>H</given-names> </name><name name-style="western"><surname>Shi</surname><given-names>L</given-names> </name><etal/></person-group><article-title>Equity assessment of the distribution of CT and MRI scanners in China: a panel data analysis</article-title><source>Int J Equity Health</source><year>2018</year><month>10</month><day>5</day><volume>17</volume><issue>1</issue><fpage>157</fpage><pub-id pub-id-type="doi">10.1186/s12939-018-0869-y</pub-id><pub-id pub-id-type="medline">30290807</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Muehlematter</surname><given-names>UJ</given-names> </name><name name-style="western"><surname>Mannil</surname><given-names>M</given-names> </name><name name-style="western"><surname>Becker</surname><given-names>AS</given-names> </name><etal/></person-group><article-title>Vertebral body insufficiency fractures: detection of vertebrae at risk on standard CT images using texture analysis and machine learning</article-title><source>Eur Radiol</source><year>2019</year><month>05</month><volume>29</volume><issue>5</issue><fpage>2207</fpage><lpage>2217</lpage><pub-id pub-id-type="doi">10.1007/s00330-018-5846-8</pub-id><pub-id pub-id-type="medline">30519934</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>L</given-names> </name><name name-style="western"><surname>Si</surname><given-names>M</given-names> </name><name name-style="western"><surname>Ma</surname><given-names>H</given-names> </name><etal/></person-group><article-title>A hierarchical opportunistic screening model for osteoporosis using machine learning applied to clinical data and CT images</article-title><source>BMC Bioinformatics</source><year>2022</year><month>02</month><day>10</day><volume>23</volume><issue>1</issue><fpage>63</fpage><pub-id pub-id-type="doi">10.1186/s12859-022-04596-z</pub-id><pub-id pub-id-type="medline">35144529</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Liang</surname><given-names>Z</given-names> </name><etal/></person-group><article-title>Differentiation of acute and chronic vertebral compression fractures using conventional CT based on deep transfer learning features and hand-crafted radiomics features</article-title><source>BMC Musculoskelet Disord</source><year>2023</year><volume>24</volume><issue>1</issue><fpage>165</fpage><pub-id pub-id-type="doi">10.1186/s12891-023-06281-5</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ge</surname><given-names>C</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Lin</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Zheng</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Cao</surname><given-names>P</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>X</given-names> </name></person-group><article-title>Preoperative prediction of residual back pain after vertebral augmentation for osteoporotic vertebral compression fractures: initial application of a radiomics score based nomogram</article-title><source>Front Endocrinol (Lausanne)</source><year>2022</year><volume>13</volume><fpage>1093508</fpage><pub-id pub-id-type="doi">10.3389/fendo.2022.1093508</pub-id><pub-id pub-id-type="medline">36619583</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Parakh</surname><given-names>A</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>H</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>JH</given-names> </name><name name-style="western"><surname>Eisner</surname><given-names>BH</given-names> </name><name name-style="western"><surname>Sahani</surname><given-names>DV</given-names> </name><name name-style="western"><surname>Do</surname><given-names>S</given-names> </name></person-group><article-title>Urinary stone detection on CT images using deep convolutional neural networks: evaluation of model performance and generalization</article-title><source>Radiol Artif Intell</source><year>2019</year><month>07</month><volume>1</volume><issue>4</issue><fpage>e180066</fpage><pub-id pub-id-type="doi">10.1148/ryai.2019180066</pub-id><pub-id pub-id-type="medline">33937795</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Xia</surname><given-names>L</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Exploring deep learning radiomics for classifying osteoporotic vertebral fractures in X-ray images</article-title><source>Front Endocrinol</source><year>2023</year><volume>15</volume><fpage>1370838</fpage><pub-id pub-id-type="doi">10.3389/fendo.2024.1370838</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Spiegl</surname><given-names>U</given-names> </name><name name-style="western"><surname>Bork</surname><given-names>H</given-names> </name><name name-style="western"><surname>Gr&#x00FC;ninger</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Osteoporotic fractures of the thoracic and lumbar vertebrae: diagnosis and conservative treatment</article-title><source>Dtsch Arztebl Int</source><year>2021</year><month>10</month><day>8</day><volume>118</volume><issue>40</issue><fpage>670</fpage><lpage>677</lpage><pub-id pub-id-type="doi">10.3238/arztebl.m2021.0295</pub-id><pub-id pub-id-type="medline">34342263</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lei</surname><given-names>M</given-names> </name><name name-style="western"><surname>Varghese</surname><given-names>B</given-names> </name><name name-style="western"><surname>Hwang</surname><given-names>D</given-names> </name><etal/></person-group><article-title>Benchmarking various radiomic toolkit features while applying the image biomarker standardization initiative toward clinical translation of radiomic analysis</article-title><source>J Digit Imaging</source><year>2021</year><month>10</month><volume>34</volume><issue>5</issue><fpage>1156</fpage><lpage>1170</lpage><pub-id pub-id-type="doi">10.1007/s10278-021-00506-6</pub-id><pub-id pub-id-type="medline">34545475</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="web"><article-title>Pyradiomics</article-title><source>Python Software Foundation</source><access-date>2023-05-17</access-date><comment><ext-link ext-link-type="uri" xlink:href="http://pypi.org/project/pyradiomics">http://pypi.org/project/pyradiomics</ext-link></comment></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="web"><article-title>PyRadiomics documentation</article-title><source>pyradiomics</source><access-date>2016-12-11</access-date><comment><ext-link ext-link-type="uri" xlink:href="http://pyradiomics.readthedocs.io">http://pyradiomics.readthedocs.io</ext-link></comment></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Orlhac</surname><given-names>F</given-names> </name><name name-style="western"><surname>Frouin</surname><given-names>F</given-names> </name><name name-style="western"><surname>Nioche</surname><given-names>C</given-names> </name><name name-style="western"><surname>Ayache</surname><given-names>N</given-names> </name><name name-style="western"><surname>Buvat</surname><given-names>I</given-names> </name></person-group><article-title>Validation of a method to compensate multicenter effects affecting CT radiomics</article-title><source>Radiology</source><year>2019</year><month>04</month><volume>291</volume><issue>1</issue><fpage>53</fpage><lpage>59</lpage><pub-id pub-id-type="doi">10.1148/radiol.2019182023</pub-id><pub-id pub-id-type="medline">30694160</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Xu</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Hao</surname><given-names>D</given-names> </name><name name-style="western"><surname>He</surname><given-names>L</given-names> </name><etal/></person-group><article-title>An assessment system for evaluating the severity of thoracolumbar osteoporotic fracture and its clinical application: a retrospective study of 381 cases</article-title><source>Clin Neurol Neurosurg</source><year>2015</year><month>12</month><volume>139</volume><fpage>70</fpage><lpage>75</lpage><pub-id pub-id-type="doi">10.1016/j.clineuro.2015.09.006</pub-id><pub-id pub-id-type="medline">26383865</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Okazaki</surname><given-names>S</given-names> </name><name name-style="western"><surname>Mine</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Yoshimi</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>RadImageNet and ImageNet as datasets for transfer learning in the assessment of dental radiographs: a comparative study</article-title><source>J Digit Imaging Inform med</source><year>2025</year><volume>38</volume><issue>1</issue><fpage>534</fpage><lpage>544</lpage><pub-id pub-id-type="doi">10.1007/s10278-024-01204-9</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fang</surname><given-names>C</given-names> </name><name name-style="western"><surname>Ji</surname><given-names>X</given-names> </name><name name-style="western"><surname>Pan</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>Combining clinical-radiomics features with machine learning methods for building models to predict postoperative recurrence in patients with chronic subdural hematoma: retrospective cohort study</article-title><source>J Med Internet Res</source><year>2024</year><month>08</month><day>28</day><volume>26</volume><fpage>e54944</fpage><pub-id pub-id-type="doi">10.2196/54944</pub-id><pub-id pub-id-type="medline">39197165</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>L</given-names> </name><name name-style="western"><surname>Lin</surname><given-names>N</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>W</given-names> </name><name name-style="western"><surname>Xiao</surname><given-names>H</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Sha</surname><given-names>Y</given-names> </name></person-group><article-title>Deep learning models for differentiating three sinonasal malignancies using multi-sequence MRI</article-title><source>BMC Med Imaging</source><year>2023</year><volume>25</volume><issue>1</issue><fpage>56</fpage><pub-id pub-id-type="doi">10.1186/s12880-024-01517-9</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="web"><article-title>SHAP</article-title><source>GitHub</source><access-date>2018-04-20</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://github.com/slundberg/shap">https://github.com/slundberg/shap</ext-link></comment></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ejiyi</surname><given-names>CJ</given-names> </name><name name-style="western"><surname>Cai</surname><given-names>D</given-names> </name><name name-style="western"><surname>Ejiyi</surname><given-names>MB</given-names> </name><etal/></person-group><article-title>Polynomial-SHAP analysis of liver disease markers for capturing of complex feature interactions in machine learning models</article-title><source>Comput Biol Med</source><year>2024</year><month>11</month><volume>182</volume><fpage>109168</fpage><pub-id pub-id-type="doi">10.1016/j.compbiomed.2024.109168</pub-id><pub-id pub-id-type="medline">39342675</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Suseki</surname><given-names>K</given-names> </name><name name-style="western"><surname>Yamashita</surname><given-names>M</given-names> </name><name name-style="western"><surname>Kojima</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Minegishi</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Komiya</surname><given-names>K</given-names> </name><name name-style="western"><surname>Takaso</surname><given-names>M</given-names> </name></person-group><article-title>Lower SMI is a risk factor for dysphagia in Japanese hospitalized patients with osteoporotic vertebral and hip fracture: a retrospective study</article-title><source>Osteoporos Sarcopenia</source><year>2022</year><month>12</month><volume>8</volume><issue>4</issue><fpage>152</fpage><lpage>157</lpage><pub-id pub-id-type="doi">10.1016/j.afos.2022.11.001</pub-id><pub-id pub-id-type="medline">36605170</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Xia</surname><given-names>L</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>X</given-names> </name><etal/></person-group><article-title>Development and validation of a predictive model for vertebral fracture risk in osteoporosis patients</article-title><source>Eur Spine J</source><year>2024</year><month>08</month><volume>33</volume><issue>8</issue><fpage>3242</fpage><lpage>3260</lpage><pub-id pub-id-type="doi">10.1007/s00586-024-08235-4</pub-id><pub-id pub-id-type="medline">38955868</pub-id></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Palmowski</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Balmer</surname><given-names>S</given-names> </name><name name-style="western"><surname>Hu</surname><given-names>Z</given-names> </name><etal/></person-group><article-title>Relationship between the OF classification and radiological outcome OF osteoporotic vertebral fractures after kyphoplasty</article-title><source>Global Spine J</source><year>2022</year><month>05</month><volume>12</volume><issue>4</issue><fpage>646</fpage><lpage>653</lpage><pub-id pub-id-type="doi">10.1177/2192568220964051</pub-id><pub-id pub-id-type="medline">33131331</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>YC</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>HH</given-names> </name><name name-style="western"><surname>Horng-Shing Lu</surname><given-names>H</given-names> </name><name name-style="western"><surname>Hondar Wu</surname><given-names>HT</given-names> </name><name name-style="western"><surname>Chang</surname><given-names>MC</given-names> </name><name name-style="western"><surname>Chou</surname><given-names>PH</given-names> </name></person-group><article-title>Can a deep-learning model for the automated detection of vertebral fractures approach the performance level of human subspecialists?</article-title><source>Clin Orthop Relat Res</source><year>2021</year><month>07</month><day>1</day><volume>479</volume><issue>7</issue><fpage>1598</fpage><lpage>1612</lpage><pub-id pub-id-type="doi">10.1097/CORR.0000000000001685</pub-id><pub-id pub-id-type="medline">33651768</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Song</surname><given-names>K</given-names> </name><name name-style="western"><surname>Ko</surname><given-names>T</given-names> </name><name name-style="western"><surname>Chae</surname><given-names>HW</given-names> </name><etal/></person-group><article-title>Development and validation of a prediction model using sella magnetic resonance imaging&#x2013;based radiomics and clinical parameters for the diagnosis of growth hormone deficiency and idiopathic short stature: cross-sectional, multicenter study</article-title><source>J Med Internet Res</source><year>2024</year><month>11</month><day>27</day><volume>26</volume><fpage>e54641</fpage><pub-id pub-id-type="doi">10.2196/54641</pub-id><pub-id pub-id-type="medline">39602803</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mei</surname><given-names>X</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Singh</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Interstitial lung disease diagnosis and prognosis using an AI system integrating longitudinal data</article-title><source>Nat Commun</source><year>2023</year><month>04</month><day>20</day><volume>14</volume><issue>1</issue><fpage>2272</fpage><pub-id pub-id-type="doi">10.1038/s41467-023-37720-5</pub-id><pub-id pub-id-type="medline">37080956</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chen</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>L</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>C</given-names> </name><etal/></person-group><article-title>CT radiomics model for discriminating the risk stratification of gastrointestinal stromal tumors: a multi-class classification and multi-center study</article-title><source>Front Oncol</source><year>2021</year><volume>11</volume><fpage>11</fpage><pub-id pub-id-type="doi">10.3389/fonc.2021.654114</pub-id></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kim</surname><given-names>JK</given-names> </name><name name-style="western"><surname>Mun</surname><given-names>S</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>S</given-names> </name></person-group><article-title>Detection and analysis of circadian biomarkers for metabolic syndrome using wearable data: cross-sectional study</article-title><source>JMIR Med Inform</source><year>2025</year><month>07</month><day>16</day><volume>13</volume><fpage>e69328</fpage><pub-id pub-id-type="doi">10.2196/69328</pub-id><pub-id pub-id-type="medline">40669055</pub-id></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Xiang</surname><given-names>F</given-names> </name><name name-style="western"><surname>Meng</surname><given-names>QT</given-names> </name><name name-style="western"><surname>Deng</surname><given-names>JJ</given-names> </name><etal/></person-group><article-title>A deep learning model based on contrast-enhanced computed tomography for differential diagnosis of gallbladder carcinoma</article-title><source>Hepatobiliary Pabcreat Dis Int</source><year>2024</year><month>08</month><volume>23</volume><issue>4</issue><fpage>376</fpage><lpage>384</lpage><pub-id pub-id-type="doi">10.1016/j.hbpd.2023.04.001</pub-id><pub-id pub-id-type="medline">37080813</pub-id></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>The computed tomography and magnetic resonance imaging acquisition parameters of the 3 centers.</p><media xlink:href="medinform_v13i1e75665_app1.docx" xlink:title="DOCX File, 19 KB"/></supplementary-material><supplementary-material id="app2"><label>Multimedia Appendix 2</label><p>A basic architecture of a convolutional neural network.</p><media xlink:href="medinform_v13i1e75665_app2.png" xlink:title="PNG File, 2205 KB"/></supplementary-material><supplementary-material id="app3"><label>Multimedia Appendix 3</label><p>Schematic diagram of the deep convolutional neural network pretraining and fine-tuning network structure.</p><media xlink:href="medinform_v13i1e75665_app3.png" xlink:title="PNG File, 807 KB"/></supplementary-material><supplementary-material id="app4"><label>Multimedia Appendix 4</label><p>Feature selection was performed using the Least Absolute Shrinkage and Selection Operator (LASSO). Left: Histogram of feature importance scores based on fused features. Right: Curve of feature coefficients varying with &#x03BB;, with the optimal &#x03BB; value being 0.0031.</p><media xlink:href="medinform_v13i1e75665_app4.png" xlink:title="PNG File, 364 KB"/></supplementary-material><supplementary-material id="app5"><label>Multimedia Appendix 5</label><p>DTL_Radscore: The y-axis represents the selected fused features, while the x-axis represents the correlation coefficients.</p><media xlink:href="medinform_v13i1e75665_app5.png" xlink:title="PNG File, 483 KB"/></supplementary-material><supplementary-material id="app6"><label>Multimedia Appendix 6</label><p>Feature selection was performed using the Least Absolute Shrinkage and Selection Operator. Left: Histogram of feature importance scores based on fused features. Right: Curve of feature coefficients varying with &#x03BB;, with the optimal &#x03BB; value being 0.0295.</p><media xlink:href="medinform_v13i1e75665_app6.png" xlink:title="PNG File, 633 KB"/></supplementary-material><supplementary-material id="app7"><label>Multimedia Appendix 7</label><p>DTL_Radscore: The y-axis represents the selected fused features, while the x-axis represents the correlation coefficients.</p><media xlink:href="medinform_v13i1e75665_app7.png" xlink:title="PNG File, 951 KB"/></supplementary-material></app-group></back></article>