Abstract
Background: Early diagnosis and intervention in glottic carcinoma (GC) can significantly improve long-term prognosis. However, the accurate diagnosis of early GC is challenging due to its morphological similarity to vocal cord dysplasia, with the difficulty further exacerbated in medically underserved areas.
Objective: This study aims to address the limitations of existing technologies by designing a vision-language multimodal model, providing a more efficient and accurate early diagnostic method for GC.
Methods: The data used in this study were sourced from the information system of the First Affiliated Hospital of Sun Yat-sen University, comprising laryngoscopy reports and 5796 laryngoscopic images from 404 patients with glottic lesions. We propose a vision-language–guided multimodal fusion network (VLMF-Net) based on a large vision-language model for the early automated diagnosis of GC. The text processing module of this model uses the pretrained Large Language Model Meta AI (LLaMa) to generate text vector representations, while the image processing module uses a pretrained vision transformer to extract features from laryngoscopic images, achieving cross-modal alignment through the Q-Former module. By leveraging a feature fusion module, deep integration of text and image features is achieved, ultimately enabling classification diagnosis. To validate the model’s performance, the study selected contrastive language-image pretraining (CLIP), bootstrapping language-image pretraining with frozen image encoders and large language models (BLIP-2), a large-scale image and noisy-text embedding (ALIGN), and vision-and-language transformer (VILT) as baseline methods for experimental evaluation on the same dataset, with comprehensive performance assessment conducted using accuracy, recall, precision, F1-score, and area under the curve.
Results: We found that on the internal test set, the VLMF-Net model significantly outperformed existing methods with an accuracy of 77.6% (CLIP: 70.5%; BLIP-2: 71.5%; ALIGN: 67.3%; and VILT: 64.3%), achieving a 6.1-percentage point improvement over the best baseline model (BLIP-2). On the external test set, our method also demonstrated robust performance, achieving an accuracy of 73.9%, which is 4.6 percentage points higher than the second-best model (BLIP-2: 69.3%). This indicates that our model surpasses these methods in the early diagnosis of GC and exhibits strong generalization ability and robustness.
Conclusions: The proposed VLMF-Net model can be effectively used for the early diagnosis of GC, helping to address the challenges in its early detection.
doi:10.2196/74902
Keywords
Introduction
Glottic carcinoma (GC) is a common malignant tumor of the head and neck [
]. According to GLOBOCAN, China reported 29,500 new cases and 16,900 deaths from laryngeal cancer in 2022 [ ], posing heavy burdens on health care systems. Approximately 60% of patients are diagnosed at an advanced stage [ ], leading to significant impairment of vital physiological functions and compromising both physical and mental health. Early diagnosis of malignant tumors has been increasingly emphasized in clinical practice due to its potential to improve cure rates and organ function preservation [ ]. Therefore, optimizing diagnostic methods for GC and enhancing early detection capability are urgent tasks for otolaryngologists.Laryngoscopy is the primary diagnostic tool for GC [
], offering direct visualization of lesion shape, extent, and surface texture. When combined with narrow band imaging, it enhances early tumor detection by identifying neovascularization [ ], making it a valuable tool for early diagnosis. However, vocal cord dysplasia (VCD), a precancerous condition situated between normal epithelium and squamous cell carcinoma, is characterized by a small lesion with clinical and laryngoscopic features similar to early GC [ ]. Therefore, it is challenging for the human eye to distinguish between them. In addition, the lesion is often covered by “leukoplakia-like” substance, which interferes with the ability of narrow band imaging to reveal submucosal vasculature [ ], increasing the risk of misdiagnosis. Laryngoscopy reports provide textual descriptions of lesion morphology observed dynamically during the examination, supplementing static images and assisting in diagnostic decision-making. Studies have demonstrated a correlation between morphological grading and malignancy risk [ - ], underscoring the diagnostic value of textual reports. Furthermore, reports authored by experienced clinicians serve as valuable references, facilitating more accurate diagnoses by less experienced clinicians. Histopathology examination remains the gold standard for diagnosis [ ]. However, biopsy is invasive, painful, and carries procedural risks [ ], impeding its widespread application in large-scale clinical screening. To address this issue, it is necessary to develop an efficient and noninvasive method to improve the diagnostic accuracy of early GC.Recently, significant progress has been made in deep learning techniques for tackling real-world classification tasks in computer vision and natural language processing [
- ]. Many researchers have sought to apply these models to the detection of laryngeal cancer, yielding promising outcomes. However, most existing methods only use laryngoscopic images as input, including UC-DenseNet [ ], MTANet [ ], Dlgnet [ ], RedFormer [ ], and SAM-FNet [ ]. Although these methods have demonstrated promising performance in laryngeal cancer detection and other tasks, they neglect certain latent information present in other modalities. Such information is typically inaccessible through unimodal approaches, thereby highlighting the advantages of multimodal methodologies [ , ].Therefore, this study proposes a novel method named vision-language–guided multimodal fusion network (VLMF-Net) for early diagnosis of GC. Addressing the limitations of traditional single-modality diagnostic methods, our approach integrates the images and reports text of laryngoscopy to provide a more comprehensive representation of lesion characteristics. Using a pretrained vision transformer (ViT) [
] model for image feature extraction and a LlaMa3 [ , ] model fine-tuned for text processing, we achieve effective multimodal feature fusion. Compared to single-modality methods, our approach significantly improves the diagnostic accuracy and robustness, achieving an accuracy of 0.776 on real clinical datasets. This study highlights the potential of multi-modal fusion in clinical auxiliary diagnosis and provides new insights for reducing misdiagnosis rates and improving patient treatment outcomes.Methods
Dataset
In our study, we constructed 2 datasets for model development and validation. First, we built an internal dataset for model training, validation, and testing. For the internal dataset, we collected data from 404 patients with glottic lesions at the First Affiliated Hospital of Sun Yat-sen University in Guangzhou, China. This dataset consists of 5799 professionally annotated image-text pairs, covering two types of lesions: VCD and GC. Each sample includes a laryngoscopic diagnostic report written by an experienced otolaryngologist and its corresponding laryngoscopic image.
In addition, to assess the model’s generalization ability and robustness, we constructed an external dataset. The external dataset was collected from the First People’s Hospital of Zhaoqing, consisting of data from 47 patients with glottic lesions between January 1, 2018, and August 31, 2024. This dataset includes 308 image-text pairs and strictly follows the principle of isolation from the training data, serving only for final performance evaluation. For detailed information on internal and external datasets, please refer to
and .
Datasets | Internal dataset | External dataset |
Number of laryngoscopy reports | 404 | 47 |
Number of VCD | 206 | 22 |
Number of GC | 198 | 25 |
aVCD: vocal cord dysplasia.
bGC: glottic carcinoma.
Model Architecture
Overview
illustrates the overall architecture of our proposed VLMF-Net model, which consists of 3 main modules: the laryngoscopic image encoder, the clinical report encoder, and the laryngeal feature fusion module. Specifically, the laryngoscopic image encoder is responsible for extracting visual feature representations from laryngoscopic images, while the clinical report encoder captures textual feature representations from the patient’s laryngoscopic examination findings. These features are then fused through the laryngeal feature fusion module. Finally, the fused features are passed through a fully connected layer to complete the classification task. The detailed implementation of each module in our proposed VLMF-Net model is as follows:

Laryngoscopic Image Encoder
The laryngoscopic image encoder is an adapted version of a pretrained ViT [
] model. Considering that a single ViT model may introduce cross-modal discrepancies in multimodal tasks, potentially affecting model accuracy, we incorporated an additional Q-Former [ ] module into the ViT model to mitigate the impact of modality differences and bridge the gap between image and text features. Q-Former is a trainable module based on Transformer that extracts and condenses visual features through alternating stacking of self-attention and cross-attention. Specifically, we first use ViT to extract features from the image, and then we input the extracted image features into a Q-Former module with frozen parameters to reduce modal differences.Formally, let denote the image encoder and denote the Q-Former module. Given an image , first resize the image to 224×224×3, then the image feature is obtained as follows:
(1)
where and are the weight parameters of the image encoder and the Q-Former module, the shape of is 1×1024. It is worth noting that in terms of the weights of the pretrained model, we adopt the transfer learning strategy: we directly use the pretrained weights of the BLIP-2 [
] model as the weight parameters of the laryngoscope image encoder, because a large number of studies have demonstrated the effectiveness of BLIP-2 in downstream tasks [ - ]. Such operations can ensure that the extracted image features and text features are in the same scale space, thus effectively avoiding catastrophic problems caused by modal differences.Clinical Report Encoder
In the clinical report encoder module, we use LLaMA3 [
], an advanced large language model, as the text feature extractor to obtain textual feature representations from the patient’s laryngoscopic examination findings. This model is renowned for its exceptional ability to understand long-form text.Formally, let denote the clinical report encoder function. Given a clinical report the process of obtaining the text feature can be formulated as follows:
(2)
where represents the weight parameters of the clinical report encoder, the shape of is 1×4096. Notably, we adopt the same transfer learning strategy for the clinical report encoder’s weight parameters as we did for the laryngoscopic image encoder. Furthermore, considering that the text in our dataset is in Chinese, we use a fine-tuned LLaMA3 model trained on a Chinese dataset as the encoder’s weight source [
]. This approach effectively addresses the original LLaMA3 model’s limitations in Chinese language understanding.Laryngeal Feature Fusion
In the 2 modules described above, we obtained image and text features. To effectively fuse these 2 modalities, we introduced the laryngeal feature fusion module, which aligns, maps, and integrates the features from both modalities. Specifically, we first use the vision projector and the text projector to map the image feature and text feature into a unified feature space:
(3)
(4)
Where and represent the mapped image and text features (their shapes are 1×512 and 1×2048), respectively, and and are the learnable parameters of the projectors and .
Next, we apply L2 normalization to both features to ensure consistency, obtaining:
and . Then and text feature along the feature dimension to form a vision-language joint representation, denoted as:. This joint feature is then fed into a classifier, which consists of multiple fully connected layers, dropout layers, and ReLU activation functions. The classification process is formulated as:
(5)
where represents the learnable parameters of the classifier.
Model Training
The internal dataset is divided at the patient level into training, validation, and test sets in a ratio of 8:1:1, strictly following the principle of data isolation. Regarding the loss function during training, we use the cross-entropy loss function and the contrastive loss function which are formulated as follows:
(6)
(7)
In , C represents the total number of classes, denotes the ground-truth class label in one-hot encoding, and is the predicted probability for class I. This loss function aims to minimize the difference between the model’s predicted distribution and the true distribution, thereby guiding parameter updates and optimization. In , represents cosine similarity, and are positive sample pairs, is the temperature parameter, and N is the number of negative samples. The final loss function is formulated as follows:
(8)
where c=0.1. All training processes and experiments are conducted on a dedicated server equipped with 4 NVIDIA A6000 GPUs with a total of 196GB of VRAM. The system runs on Ubuntu 20.04.5 LTS, and the model is implemented using PyTorch 3.9.0 and Scikit-learn 1.3.1. In this study, we use the AdamW optimizer to optimize VLMF-Net, with an initial learning rate set to 0.00001. A warm-up strategy and cosine learning rate scheduling are adopted to dynamically adjust the learning rate. VLMF-Net is trained for a total of 80 epochs.
Ethical Considerations
This study was approved by the Ethics Committee of the First Affiliated Hospital of Sun Yat-sen University (approval number [2023]755‐1). Informed consent was waived by the institutional review boards of all participating hospitals due to the study’s retrospective design. We implemented stringent measures to protect the privacy of all participants by anonymizing all collected data to remove any personally identifiable information. Throughout the manuscript preparation, we diligently avoided disclosing any details that could reveal the identity of participants. Furthermore, no compensation or indemnity was required as the study did not cause any losses to participants beyond the necessary clinical diagnostic and therapeutic measures.
Results
Result of Our Model
To demonstrate the effectiveness and advantages of our proposed VLMF-Net, we selected 4 classic models as baseline models and performed comparisons on 2 datasets. The 4 models include vision-and-language transformer (VILT) [
], contrastive language-image pretraining (CLIP) [ ], bootstrapping language-image pretraining with frozen image encoders and large language models (BLIP-2) [ ], and a large-scale image and noisy-text embedding (ALIGN) [ ], and the 2 datasets refer to the internal and external datasets mentioned earlier. and present the average results of 5 trials for the 4 baseline models and our proposed model on the internal and external datasets, respectively. shows the receiver operating characteristic curves and corresponding area under the curve values of different models on internal and external datasets. As shown in , on the internal dataset, our method demonstrates significant advantages, achieving the following evaluation metrics: accuracy (0.776), precision (0.820), and F1-score (0.776). Notably, compared to the second-best model, our method achieves significant improvements of 0.061, 0.032, and 0.046 in accuracy, precision, and F1-score, respectively. In terms of class-wise recall, VLMF-Net achieves recall rates of 0.754 and 0.803 for VCD and GC, respectively. Compared to other models, VLMF-Net shows significant improvements across multiple evaluation metrics, indicating its superior ability in recognizing glottic cancer. In addition, as shown in , on the external dataset, our method also demonstrates significant advantages, achieving accuracy (0.739), precision (0.828), and F1-score (0.737), with improvements of 0.046, 0.053, and 0.046 over the second-best model, respectively. In terms of class-wise recall, our model achieves recall rates of 0.701 and 0.793 for VCD and GC, respectively. Compared to other models, there is also a significant improvement, which further demonstrates that VLMF-Net possesses excellent generalization ability and robustness. Moreover, as shown in , our model demonstrates significant advantages in area under the curve metrics on both internal and external datasets. On the internal dataset, it outperforms the second-best model by 0.026, and on the external dataset by 0.012.Methods | Overall results | Recall of different classes | ||||
Accuracy, mean (SD) | Precision, mean (SD) | F1-score, mean (SD) | VCD | , mean (SD)GC | , mean (SD)||
VILT | 0.643 (0.02 | )0.677 (0.02 | )0.643 (0.01 | )0.632 (0.02 | )0.656 (0.01 | )|
CLIP | 0.705 (0.01 | )0.788 (0.02 | )0.703 (0.01 | )0.672 (0.02 | )0.750 (0.01 | )|
BLIP-2 | 0.715 (0.02 | )0.770 (0.03 | )0.714 (0.02 | )0.694 (0.04 | )0.742 (0.03 | )|
ALIGN | 0.673 (0.03 | )0.707 (0.02 | )0.673 (0.04 | )0.660 (0.02 | )0.688 (0.02 | )|
VLMF-Net | 0.776 (0.01) | 0.820 (0.02) | 0.776 (0.01) | 0.754 (0.02) | 0.803 (0.01) |
aVCD: vocal cord dysplasia.
bGC: glottic carcinoma.
cVILT: vision-and-language transformer.
dP<.001.
eCLIP: contrastive language-image pretraining.
fBLIP-2: bootstrapping language-image pretraining with frozen image encoders and large language models.
gALIGN: a large-scale image and noisy-text embedding.
hVLMF-Net: vision-language guided multimodal fusion network.
Methods | Overall results | Recall of different classes | ||||
Accuracy, mean (SD) | Precision, mean (SD) | F1-score, mean (SD) | VCD | , mean (SD)GC | , mean (SD)||
VILT | 0.631 (0.02 | )0.663 (0.02 | )0.630 (0.01 | )0.633 (0.01 | )0.628 (0.01 | )|
CLIP | 0.686 (0.01 | )0.748 (0.02 | )0.685 (0.01 | )0.670 (0.02 | )0.708 (0.02 | )|
BLIP-2 | 0.693 (0.03 | )0.775 (0.02 | )0.691 (0.03 | )0.669 (0.03 | )0.726 (0.02 | )|
ALIGN | 0.647 (0.02 | )0.680 (0.02 | )0.647 (0.03 | )0.642 (0.02 | )0.653 (0.03 | )|
VLMF-Net | 0.739 (0.02) | 0.828 (0.02) | 0.737 (0.02) | 0.701 (0.03) | 0.793 (0.02) |
aVCD: vocal cord dysplasia.
bGC: glottic carcinoma.
cVILT: vision-and-language transformer.
dP value <.001.
eCLIP: contrastive language-image pretraining.
fBLIP-2: bootstrapping language-image pretraining with frozen image encoders and large language models.
gALIGN: a large-scale image and noisy-text embedding.
hVLMF-Net: vision-language guided multimodal fusion network.

Ablation Studies
To validate the effectiveness of our multimodal approach for early diagnosis of GC, we designed an ablation study to systematically evaluate the performance differences between models using single-modal and multimodal inputs: (1) M1: a single-modal model using only laryngoscopy images; (2) M2: a single-modal model using only laryngoscopy diagnostic reports; (3) M3: a multimodal model combining images and text, but without Q-Former. (4) M4: a multimodal model combining images and text with Q-Former. The experimental results, as shown in
, indicate that the multimodal model M4 significantly outperforms both single-modal models across all evaluation metrics. Specifically, compared to the best single-modal model, M4 achieves improvements of 0.098 in accuracy (0.776), 0.106 in precision (0.820), 0.100 in recall (0.779), and 0.098 in F1-score (0.776). This demonstrates that the multimodal data fusion strategy effectively integrates visual features and textual semantic information, significantly enhancing the diagnostic performance of the model.Variants | Image | Report | Q-Former | Accuracy, mean (SD) | Precision, mean (SD) | Recall, mean (SD) | F1-score, mean (SD) |
M1 | √ | 0.678 (0.02) | 0.714 (0.03) | 0.679 (0.03) | 0.678 (0.03) | ||
M2 | √ | 0.673 (0.01) | 0.711 (0.02) | 0.675 (0.01) | 0.673 (0.01) | ||
M3 | √ | √ | 0.722 (0.02) | 0.708 (0.01) | 0.723 (0.01) | 0.722 (0.03) | |
M4 | √ | √ | √ | 0.776 (0.01) | 0.820 (0.02) | 0.779 (0.02) | 0.776 (0.01) |
In addition, to verify the effectiveness of the Q-Former module, we designed ablation experiments related to the Q-Former module. The experimental results are shown in
. M4 outperforms M3 in all indicators. Specifically, M4’s accuracy (0.776) increased by 0.054, precision (0.820) increased by 0.112, recall (0.779) increased by 0.056, and F1-score (0.776) increased by 0.054. This indicates that the Q-Former module effectively reduces the impact of cross-modal differences.Visualization for Model Prediction
To enhance the interpretability of the VLMF-Net model’s predictions, we used the Grad-CAM (Gradient-Weighted Class Activation Mapping) [
] algorithm to generate class activation heatmaps. The visualization results are shown in . In the heatmap, we can see that VLMF-Net focuses on the lesion area of the patient when analyzing the laryngoscopy image. This indicates that the model is able to correctly classify based on the features of the lesion area, enabling early diagnosis of GC.
Discussion
Main Findings
In our study, we developed a novel model for the early diagnosis of GC named VLMF-Net, which leverages multimodal fusion technology guided by vision-language information. When tested on an internal dataset, VLMF-Net achieved accuracy, precision, recall, and F1-score of 0.776, 0.820, 0.779, and 0.776, respectively, outperforming all baseline methods. These results demonstrate the feasibility and effectiveness of VLMF-Net in the early diagnosis of GC. Ablation studies further reveal that VLMF-Net significantly outperforms unimodal models. By integrating both laryngoscopic images and clinical text reports, VLMF-Net captures complementary diagnostic information, thus mitigating the risk of information loss inherent in single-modality systems. Notably, the textual modality meaningfully guides the extraction of image features; for instance, textual cues explicitly describing the lesion’s location (eg, “anterior-middle segment of the left vocal cord”) can help the model focus on the relevant visual regions. This form of cross-modal interaction may enhance the model’s ability to detect subtle pathological patterns that might otherwise be overlooked.
Moreover, Grad-CAM–based visualizations demonstrate that VLMF-Net consistently attends to clinically significant lesion areas. These attention heatmaps show strong alignment between the model’s focus and expert-defined pathological regions, enhancing interpretability. Such interpretability is crucial in clinical contexts, where transparency in decision-making processes fosters trust and facilitates integration of AI systems into diagnostic workflows.
Finally, these findings validate the design philosophy of VLMF-Net and highlight the broader potential of multimodal fusion strategies in medical AI. In particular, the ability to synthesize visual and contextual clinical information allows for more robust and informed diagnostic decisions. This work not only advances the state of the art in GC diagnosis but also lays a foundation for extending vision-language multimodal techniques to other complex diagnostic tasks where rich multimodal data is available.
Deep Learning Challenges in Early GC Diagnosis
Early GC lesions are usually small and exhibit complex morphological characteristics, making it challenging for deep learning models to capture fine-grained lesion features. As a result, important details related to the lesions may be missed, significantly affecting the model’s diagnostic accuracy. Furthermore, most existing mainstream models rely solely on laryngoscopic images as input. While these models have achieved some progress in a unimodal setting, they overlook information from other modalities, which is often beyond the reach of image-based models [
- ]. This limitation constrains the model’s comprehensive understanding of lesion characteristics and ultimately affects diagnostic accuracy.Strengths and Limitations
To the best of our knowledge, this is the first study to apply multimodal techniques to the early diagnosis of GC. In our model, we fully utilize patients’ laryngoscopy reports and laryngoscopic images to extract relevant information about GC, enabling the model to comprehensively understand the patient’s condition and thereby improve its performance. However, this study still has some limitations. First, the VLMF-Net, based on a pretrained large language model, relies heavily on powerful computational resources, particularly high-performance GPUs, during training, which may result in slower training and inference speeds. Second, the data used in this study underwent quality checks, but in real-world scenarios, more complex situations may arise, such as poor image clarity due to imaging devices or challenging shooting angles, which could affect the model’s diagnostic accuracy.
Conclusions
In this paper, we propose a VLMF-Net for the early diagnosis of GC. Extensive experiments on 2 datasets demonstrate that VLMF-Net achieves superior accuracy and robustness, effectively addressing the challenges of early GC diagnosis.
Acknowledgments
We note that a shorter conference version of this paper appeared in the 2024 China Health Information Processing Conference (CHIP 2024). In our initial conference paper, we did not investigate whether the vision-language guided multimodal fusion network (VLMF-Net) model maintains its effectiveness across different datasets and consistently outperforms baseline models. This manuscript addresses these issues while further optimizing the originally proposed model through approaches such as hybrid loss function training. In addition, we use the Grad-CAM (Gradient-Weighted Class Activation Mapping) methodology to analyze the interpretability of VLMF-Net. This work is partially supported by the National Natural Science Foundation of China (62473267), the Basic and Applied Basic Research Project of Guangdong Province (2022B1515130009, 2025A1515011614), the Guangzhou Science and Technology Planning Project (No.2025B03J0019), the Special Subject on Agriculture and Social Development, Key Research and Development Plan in Guangzhou (2023B03J0172), and the Natural Science Foundation of Top Talent of SZTU (GDRC202318).
Conflicts of Interest
None declared.
References
- Steuer CE, El-Deiry M, Parks JR, Higgins KA, Saba NF. An update on larynx cancer. CA Cancer J Clin. Jan 2017;67(1):31-50. [CrossRef] [Medline]
- Bray F, Laversanne M, Sung H, et al. Global cancer statistics 2022: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J Clin. 2024;74(3):229-263. [CrossRef] [Medline]
- Chen SA, Muller S, Chen AY, et al. Patterns of extralaryngeal spread of laryngeal cancer: thyroid cartilage penetration occurs in a minority of patients with extralaryngeal spread of laryngeal squamous cell cancers. Cancer. Nov 15, 2011;117(22):5047-5051. [CrossRef] [Medline]
- LeBlanc BJ, Shi R, Mehta V, Mills G, Ampil F, Nathan CAO. Improvements in survival and disparities for advanced-stage laryngeal cancer. JAMA Otolaryngol Head Neck Surg. Feb 2015;141(2):169-173. [CrossRef] [Medline]
- Cosway B, Drinnan M, Paleri V. Narrow band imaging for the diagnosis of head and neck squamous cell carcinoma: a systematic review. Head Neck. Apr 2016;38 Suppl 1:E2358-E2367. [CrossRef] [Medline]
- You Z, Han B, Shi Z, et al. Vocal cord leukoplakia classification using deep learning models in white light and narrow band imaging endoscopy images. Head Neck. Dec 2023;45(12):3129-3145. [CrossRef] [Medline]
- Kraft M, Fostiropoulos K, Gürtler N, Arnoux A, Davaris N, Arens C. Value of narrow band imaging in the early diagnosis of laryngeal cancer. Head Neck. Jan 2016;38(1):15-20. [CrossRef] [Medline]
- Zhao Q, He Y, Wu Y, et al. Vocal cord lesions classification based on deep convolutional neural network and transfer learning. Med Phys. Jan 2022;49(1):432-442. [CrossRef] [Medline]
- van Hulst AM, Kroon W, van der Linden ES, et al. Grade of dysplasia and malignant transformation in adults with premalignant laryngeal lesions. Head Neck. Apr 2016;38 Suppl 1(S1):E2284-E2290. [CrossRef] [Medline]
- Ni XG, Zhu JQ, Zhang QQ, Zhang BG, Wang GQ. Diagnosis of vocal cord leukoplakia: the role of a novel narrow band imaging endoscopic classification. Laryngoscope. Feb 2019;129(2):429-434. [CrossRef] [Medline]
- Mannelli G, Cecconi L, Gallo O. Laryngeal preneoplastic lesions and cancer: challenging diagnosis. Qualitative literature review and meta-analysis. Crit Rev Oncol Hematol. Oct 2016;106:64-90. [CrossRef] [Medline]
- Mao C, Li J, Pang PCI, Zhu Q, Chen R. Identifying kidney stone risk factors through patient experiences with a large language model: text analysis and empirical study. J Med Internet Res. May 22, 2025;27:e66365. [CrossRef] [Medline]
- Sahoo PK, Mishra S, Panigrahi R, Bhoi AK, Barsocchi P. An improvised deep-learning-based mask R-CNN model for laryngeal cancer detection using CT images. Sensors (Basel). Nov 15, 2022;22(22):8834. [CrossRef] [Medline]
- Bensoussan Y, Vanstrum EB, Johns MM III, Rameau A. Artificial intelligence and laryngeal cancer: from screening to prognosis: a state of the art review. Otolaryngol Head Neck Surg. Mar 2023;168(3):319-329. [CrossRef] [Medline]
- Luo X, Zhang J, Li Z, Yang R. Diagnosis of ulcerative colitis from endoscopic images based on deep learning. Biomed Signal Process Control. Mar 2022;73:103443. [CrossRef]
- Zhou W, Dong S, Lei J, Yu L. MTANet: multitask-aware network with hierarchical multimodal fusion for RGB-T urban scene understanding. IEEE Trans Intell Veh. May 2022;8(1):48-58. [CrossRef]
- Wang KN, Zhuang S, Ran QY, et al. DLGNet: a dual-branch lesion-aware network with the supervised gaussian mixture model for colon lesions classification in colonoscopy images. Med Image Anal. Jul 2023;87:102832. [CrossRef] [Medline]
- Cui C, Ma Y, Lu J, Wang Z. REDFormer: radar enlightens the darkness of camera perception with transformers. IEEE Trans Intell Veh. Nov 6, 2023;9(1):1358-1368. [CrossRef]
- Wei J, Li Y, Qiu M, Chen H, Fan X, Lei W. SAM-FNet: SAM-guided fusion network for laryngo-pharyngeal tumor detection. Presented at: 2024 IEEE International Conference on Bioinformatics and Biomedicine (BIBM); Dec 3-6, 2024; Lisbon, Portugal. [CrossRef]
- Kang EYC, Yeung L, Lee YL, et al. A multimodal imaging-based deep learning model for detecting treatment-requiring retinal vascular diseases: model development and validation study. JMIR Med Inform. May 31, 2021;9(5):e28868. [CrossRef] [Medline]
- Noda M, Yoshimura H, Okubo T, et al. Feasibility of multimodal artificial intelligence using GPT-4 vision for the classification of middle ear disease: qualitative study and validation. JMIR AI. May 31, 2024;3(1):e58342. [CrossRef] [Medline]
- Khan S, Naseer M, Hayat M, Zamir SW, Khan FS, Shah M. Transformers in vision: a survey. ACM Comput Surv. Jan 31, 2022;54(10s):1-41. [CrossRef]
- Grattafiori A, Dubey A, Jauhri A, Pandey A, Kadian A, Al-Dahle A, et al. The Llama 3 herd of models. arXiv. Preprint posted online on Jul 31, 2024. [CrossRef]
- Cui Y, Yang Z, Yao X. Efficient and effective text encoding for Chinese LLaMA and Alpaca. arXiv. Preprint posted online on Apr 17, 2023. [CrossRef]
- Li J, Li D, Savarese S, Hoi S. BLIP-2: bootstrapping language-image pre-training with frozen image encoders and large language models. arXiv. Preprint posted online on Jan 23, 2023. [CrossRef]
- Lee C, Jang J, Lee J. Personalizing text-to-image generation with visual prompts using BLIP-2. Presented at: Eleventh International Conference on Learning Representations (ICLR 2023); May 1-5, 2023; Kigali, Rwanda. URL: https://openreview.net/pdf?id=7wxCMgdj5F [Accessed 2025-09-17]
- Nguyen T, Gadre SY, Ilharco G, Oh S, Schmidt L. Improving multimodal datasets with image captioning. arXiv. Preprint posted online on Jul 19, 2023. [CrossRef]
- Zhu D, Chen J, Haydarov K, Shen X, Zhang W, Elhoseiny M. ChatGPT asks, BLIP-2 answers: automatic questioning towards enriched visual descriptions. arXiv. Preprint posted online on Mar 12, 2023. [CrossRef]
- Kim W, Son B, Kim I. ViLT: vision-and-language transformer without convolution or region supervision. arXiv. Preprint posted online on Feb 5, 2021. [CrossRef]
- Radford A, Kim JW, Hallacy C, Ramesh A, Goh G, Agarwal S, et al. Learning transferable visual models from natural language supervision. Presented at: 38th International Conference on Machine Learning (ICML 2021); Jul 18-24, 2021. URL: https://proceedings.mlr.press/v139/radford21a/radford21a.pdf [Accessed 2025-09-17]
- Jia C, Yang Y, Xia Y, et al. Scaling up visual and vision-language representation learning with noisy text supervision. arXiv. Preprint posted online on Feb 11, 2021. [CrossRef]
- Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D. Grad-CAM: visual explanations from deep networks via gradient-based localization. Presented at: International Conference on Computer Vision (ICCV); Oct 22-29, 2017; Venice, Italy. [CrossRef]
Abbreviations
ALIGN: a large-scale image and noisy-text embedding |
BLIP-2: bootstrapping language-image pretraining with frozen image encoders and large language models |
CLIP: contrastive language-image pretraining |
GC: glottic carcinoma |
Grad-CAM: Gradient-Weighted Class Activation Mapping |
LLaMa: Large Language Model Meta AI |
VCD: vocal cord dysplasia |
ViT: vision transformer |
VLIT: vision-and-language transformer |
VLMF-Net: vision-language guided multimodal fusion network |
Edited by Qiao Jin; submitted 27.Mar.2025; peer-reviewed by Ran Gu, Zhikai Yang; final revised version received 04.Aug.2025; accepted 06.Aug.2025; published 08.Oct.2025.
Copyright© Zhaohui Jin, Yi Shuai, Yun Li, Mianmian Chen, Yumeng Liu, Wenbin Lei, Xiaomao Fan. Originally published in JMIR Medical Informatics (https://medinform.jmir.org), 8.Oct.2025.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on https://medinform.jmir.org/, as well as this copyright and license information must be included.