<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Med Inform</journal-id><journal-id journal-id-type="publisher-id">medinform</journal-id><journal-title>JMIR Medical Informatics</journal-title><abbrev-journal-title>JMIR Med Inform</abbrev-journal-title><issn pub-type="epub">2291-9694</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v13i1e76148</article-id><article-id pub-id-type="doi">10.2196/76148</article-id><article-categories><subj-group subj-group-type="heading"><subject>Original Paper</subject></subj-group></article-categories><title-group><article-title>Enhancing Oral Health Diagnostics With Hyperspectral Imaging and Computer Vision: Clinical Dataset Study</article-title></title-group><contrib-group><contrib contrib-type="author" corresp="yes" equal-contrib="yes"><name name-style="western"><surname>R&#x00F6;mer</surname><given-names>Paul</given-names></name><degrees>Dr med dent, DMD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Ponciano</surname><given-names>Jean-Jacques</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Kloster</surname><given-names>Katharina</given-names></name><degrees>DMD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Siegberg</surname><given-names>Fabia</given-names></name><degrees>Dr med dent, DMD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Pla&#x00DF;</surname><given-names>Bastian</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Vinayahalingam</surname><given-names>Shankeeth</given-names></name><degrees>MD, DMD</degrees><xref ref-type="aff" rid="aff3">3</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Al-Nawas</surname><given-names>Bilal</given-names></name><degrees>MD, Dr med habil, Dr med dent, DMD, PhD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>K&#x00E4;mmerer</surname><given-names>Peer W</given-names></name><degrees>MA, MD, Dr med habil, Dr med dent, DMD, PhD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Klauer</surname><given-names>Thomas</given-names></name><degrees>DrIng</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Thiem</surname><given-names>Daniel</given-names></name><degrees>MD, MHBA, Dr med habil, Dr med, DMD, PhD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib></contrib-group><aff id="aff1"><institution>Department of Oral and Maxillofacial Surgery, University Medical Center of the Johannes Gutenberg University Mainz</institution><addr-line>Augustusplatz 2</addr-line><addr-line>Mainz</addr-line><country>Germany</country></aff><aff id="aff2"><institution>Institute for Spatial Information and Surveying Technology, University of Applied Sciences</institution><addr-line>Mainz</addr-line><country>Germany</country></aff><aff id="aff3"><institution>Department of Oral and Maxillofacial Surgery, Radboud University Nijmegen</institution><addr-line>Nijmegen</addr-line><country>The Netherlands</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Benis</surname><given-names>Arriel</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Afolabi</surname><given-names>Boluwatife</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Alabi</surname><given-names>Deborah Oluwatobi</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Singh</surname><given-names>Prabhishek</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Paul R&#x00F6;mer, Dr med dent, DMD, Department of Oral and Maxillofacial Surgery, University Medical Center of the Johannes Gutenberg University Mainz, Augustusplatz 2, Mainz, 55131, Germany, 49 1747978980; <email>p.roemer@icloud.com</email></corresp><fn fn-type="equal" id="equal-contrib1"><label>*</label><p>these authors contributed equally</p></fn></author-notes><pub-date pub-type="collection"><year>2025</year></pub-date><pub-date pub-type="epub"><day>11</day><month>9</month><year>2025</year></pub-date><volume>13</volume><elocation-id>e76148</elocation-id><history><date date-type="received"><day>17</day><month>04</month><year>2025</year></date><date date-type="rev-recd"><day>05</day><month>07</month><year>2025</year></date><date date-type="accepted"><day>06</day><month>07</month><year>2025</year></date></history><copyright-statement>&#x00A9; Paul R&#x00F6;mer, Jean-Jacques Ponciano, Katharina Kloster, Fabia Siegberg, Bastian Pla&#x00DF;, Shankeeth Vinayahalingam, Bilal Al-Nawas, Peer W K&#x00E4;mmerer, Thomas Klauer, Daniel Thiem. Originally published in JMIR Medical Informatics (<ext-link ext-link-type="uri" xlink:href="https://medinform.jmir.org">https://medinform.jmir.org</ext-link>), 11.9.2025. </copyright-statement><copyright-year>2025</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://medinform.jmir.org/">https://medinform.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://medinform.jmir.org/2025/1/e76148"/><abstract><sec><title>Background</title><p>Diseases of the oral cavity, including oral squamous cell carcinoma, pose major challenges to health care worldwide due to their late diagnosis and complicated differentiation of oral tissues. The combination of endoscopic hyperspectral imaging (HSI) and deep learning (DL) models offers a promising approach to the demand for modern, noninvasive tissue diagnostics. This study presents a large-scale in vivo dataset designed to support DL-based segmentation and classification of healthy oral tissues.</p></sec><sec><title>Objective</title><p>This study aimed to develop a comprehensive, annotated endoscopic HSI dataset of the oral cavity and to demonstrate automated, reliable differentiation of intraoral tissue structures by integrating endoscopic HSI with advanced machine learning methods.</p></sec><sec sec-type="methods"><title>Methods</title><p>A total of 226 participants (166 women [73.5%], 60 men [26.5%], aged 24&#x2010;87 years) were examined using an endoscopic HSI system, capturing spectral data in the range of 500 to 1000 nm. Oral structures in red, green, and blue and HSI scans were annotated using RectLabel Pro (by Ryo Kawamura). DeepLabv3 (Google Research) with a ResNet-50 backbone was adapted for endoscopic HSI segmentation. The model was trained for 50 epochs on 70% of the dataset, with 30% for evaluation. Performance metrics (precision, recall, and <italic>F</italic><sub>1</sub>-score) confirmed its efficacy in distinguishing oral tissue types.</p></sec><sec sec-type="results"><title>Results</title><p>DeepLabv3 (ResNet-101) and U-Net (EfficientNet-B0/ResNet-50) achieved the highest overall <italic>F</italic><sub>1</sub>-scores of 0.857 and 0.84, respectively, particularly excelling in segmenting the mucosa (0.915), retractor (0.94), tooth (0.90), and palate (0.90). Variability analysis confirmed high spectral diversity across tissue classes, supporting the dataset&#x2019;s complexity and authenticity for realistic clinical conditions.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>The presented dataset addresses a key gap in oral health imaging by developing and validating robust DL algorithms for endoscopic HSI data. It enables accurate classification of oral tissue and paves the way for future applications in individualized noninvasive pathological tissue analysis, early cancer detection, and intraoperative diagnostics of oral diseases.</p></sec></abstract><kwd-group><kwd>hyperspectral Imaging</kwd><kwd>endoscopic HSI</kwd><kwd>eHSI</kwd><kwd>oral mucosa perfusion</kwd><kwd>ischemia</kwd><kwd>oral health diagnostics</kwd><kwd>deep learning</kwd><kwd>oral squamous cell carcinoma</kwd><kwd>medical imaging dataset</kwd><kwd>tissue classification</kwd><kwd>spectral analysis</kwd><kwd>oral pathology</kwd><kwd>artificial intelligence</kwd><kwd>dentistry</kwd><kwd>non-invasive diagnostics</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>Oral diseases, including malignant and premalignant lesions, often occur on pre-existing chronic tissue alterations that are difficult to discern through conventional visual examination. This process heavily relies on the clinician&#x2019;s expertise and subjective interpretation, frequently necessitating additional diagnostic measures. Available methods include incisional and excisional biopsies, brush biopsies, cytological techniques, and optical approaches. Among these, scalpel biopsy remains the gold standard for diagnosing potentially malignant lesions, offering a diagnostic accuracy of up to 88.9% [<xref ref-type="bibr" rid="ref1">1</xref>]. However, it involves the partial removal of tissue from a suspicious area for histopathological examination, making it invasive, costly, and potentially inadequate for multilocular lesions. Excisional biopsies, while more reliable due to larger sample sizes, bear the risk of incomplete removal of malignancies and overtreatment in cases of benign lesions [<xref ref-type="bibr" rid="ref2">2</xref>].</p><p>In recent years, less invasive methods, such as brush biopsies, tissue autofluorescence, and chemiluminescence (eg, toluidine blue staining) have been introduced as cost-effective alternatives. Meta-analyses report high sensitivities for these techniques, ranging from 30% to 100% for tissue autofluorescence, 77% for toluidine blue staining, and 91%&#x2010;100% for brush biopsies [<xref ref-type="bibr" rid="ref3">3</xref>-<xref ref-type="bibr" rid="ref5">5</xref>]. However, these methods have not significantly improved the early detection of oral squamous cell carcinoma. The examiner-dependent variability of less invasive methods can, at worst, delay the accurate diagnosis of oral squamous cell carcinoma, one of the most prevalent malignant tumors globally, accounting for 90%&#x2010;95% of all malignant oral cavity pathologies [<xref ref-type="bibr" rid="ref6">6</xref>-<xref ref-type="bibr" rid="ref8">8</xref>]. Consequently, the surgical scalpel biopsy remains the diagnostic gold standard [<xref ref-type="bibr" rid="ref9">9</xref>]. Evidence shows that patients undergoing routine clinical evaluations, including visual inspection and digital palpation, achieve significantly higher 5-year survival rates [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref11">11</xref>].</p><p>The integration of advanced imaging technologies into clinical diagnostics is revolutionizing oral health care. Among these innovations, hyperspectral imaging (HSI) stands out as a noninvasive, highly sensitive modality that captures detailed spectral information across hundreds of wavelengths beyond the visible light spectrum. By detecting tissue-specific spectral signatures, HSI combines imaging remission spectroscopy with conventional imaging techniques [<xref ref-type="bibr" rid="ref12">12</xref>]. HSI acquires spatial and spectral information as a 3D hyperspectral cube in a noncontact, noninvasive, and radiation-free manner. This approach provides vast datasets spanning wide wavelength spectra, enabling the immediate extraction of diagnostically relevant information [<xref ref-type="bibr" rid="ref13">13</xref>-<xref ref-type="bibr" rid="ref18">18</xref>]. The efficacy of HSI relies on the distinct spectral signatures of tissues, which result from their absorption, reflection, and refraction properties when illuminated. Each tissue type generates a unique light spectrum that HSI systems can capture, offering the potential for detecting pathological changes with high precision and sensitivity [<xref ref-type="bibr" rid="ref19">19</xref>].</p><p>Endoscopic HSI enhances access to the oral cavity and improves illumination, making it particularly valuable in addressing the growing demand for noninvasive diagnostic methods in oral medicine [<xref ref-type="bibr" rid="ref20">20</xref>-<xref ref-type="bibr" rid="ref23">23</xref>]. In this context, the combination of endoscopic HSI with deep learning (DL)&#x2013;based computer vision techniques provides the potential to significantly enhance diagnostic accuracy, allowing for more comprehensive mapping of the oral cavity and accurate distinction between healthy and pathologically altered mucosa. This offers potential opportunities to avoid unnecessary biopsies and to improve patient-centered therapy by determining individual resection margins. The advanced capabilities of endoscopic HSI, while promising, pose challenges in managing its vast and intricate data output. The spectral data acquired by endoscopic HSI, characterized by tissue-specific signatures, surpass the processing abilities of human observers and traditional analytical tools. This complexity necessitates advanced data analysis, where DL plays a pivotal role. Specialized in interpreting multidimensional datasets, DL uses neural networks to process extensive endoscopic HSI data efficiently, identifying subtle patterns and deviations indicative of pathology that might be missed by human assessment or conventional algorithms [<xref ref-type="bibr" rid="ref24">24</xref>]. The integration of DL into endoscopic HSI enhances the precision, consistency, and speed of data interpretation, significantly accelerating the diagnostic process. This is especially critical in clinical environments, where timely, accurate real-time analysis can profoundly impact patient outcomes [<xref ref-type="bibr" rid="ref25">25</xref>-<xref ref-type="bibr" rid="ref28">28</xref>]. However, implementing DL in endoscopic HSI analysis presents challenges, primarily the limited availability of annotated endoscopic HSI datasets reflecting the complexity and diversity of real-world clinical cases. Existing datasets, often derived from controlled laboratory environments, inadequately prepare DL models for the variability of clinical settings, limiting their practical utility. A review by Cui et al [<xref ref-type="bibr" rid="ref24">24</xref>] highlights this data scarcity as a major barrier to effective DL analysis in medical imaging, emphasizing the need for specialized datasets. In recent years, DL has also shown great promise in various other domains of medical imaging, including tumor classification in magnetic resonance imaging scans [<xref ref-type="bibr" rid="ref29">29</xref>] and noise reduction in medical images [<xref ref-type="bibr" rid="ref30">30</xref>-<xref ref-type="bibr" rid="ref33">33</xref>]. These studies highlight the versatility and effectiveness of neural networks in processing complex medical images. Building on these advances, the present work applies DL and endoscopic HSI to the oral cavity&#x2014;an area where such integration remains underexplored despite its clinical potential of enhancing current standards to advance precision and efficacy in oral diagnostics.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Study Cohort</title><p>A total of 226 participants, including 166 (73.5%) women and 60 (26.5%) men aged between 24 and 87 years, were included in this prospective, clinical study. Endoscopic HSI data were acquired at the Department of Oral and Maxillofacial Surgery, Facial Plastic Surgery, University Medical Center. The examined participants represented a heterogeneous group of patients from the oral and maxillofacial surgery outpatient clinic. Patients with macroscopic abnormalities of the oral mucosa, premalignant lesions, or tumors were excluded from the study. All participants provided informed consent prior to the procedure and data collection. The individual pictured in <xref ref-type="fig" rid="figure1">Figure 1</xref> has provided permission for their image to be used in this publication. This study was approved by the local ethics committee of Rhineland-Palatinate (registration number 2021&#x2010;16158) and was conducted in accordance with the code of ethics of the World Medical Association (Declaration of Helsinki).</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Experimental setup for endoscopic hyperspectral measurements of the oral cavity.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v13i1e76148_fig01.png"/></fig></sec><sec id="s2-2"><title>HSI and Patient Data Acquisition</title><p>The HSI datasets were acquired using a state-of-the-art endoscopic HSI sensor system (TIVITA Mini<sup>&#x00AE;</sup> camera system, Diaspective Vision GmbH). This system detects 100 wavelengths in the range of 500&#x2010;1000 nm, offering a bandwidth of 5 nm [<xref ref-type="bibr" rid="ref34">34</xref>]. The device uses the pushbroom principle, enabling chemical component detection based on light absorption and reflection behavior. During operation, light enters the optical system of the spectrometer through the lens, where it is collimated and separated into individual wavelengths using a transmission grating. The separated light then passes through a second optical system before reaching the sensor of the connected complementary metal-oxide-semiconductor camera. The spectrometer directly detects the spatial direction and width of the object being scanned (Y-axis) while the second spatial direction and length of the object (X-axis) are determined through the continuous mechanical movement of the light entry gate within the scanning unit. This process generates a 3D data cube that includes a spectral dimension (&#x03BB;), capturing complete tissue spectra for each pixel within the wavelength range of 500&#x2010;1000 nm [<xref ref-type="bibr" rid="ref35">35</xref>].</p><p>Standardized measurement protocols were followed during image acquisition, maintaining a consistent distance of 7&#x2010;10 cm to ensure high-quality and distinct image data (<xref ref-type="fig" rid="figure1">Figure 1</xref>). Measurements were performed in an examination room under dimmed lighting conditions to ensure uniformity and comparability. A total of 226 participants aged between 24 and 87 years, including 166 (73.5%) women and 60 men (26.5%), were prospectively recruited at the outpatient department of the clinic of Oral and Maxillofacial Surgery, University Medical Center, where all individuals underwent routine clinical evaluation and inspection. Inclusion was limited to individuals without clinically apparent mucosal lesions, premalignant conditions, or malignancies, ensuring a representative sample of nonpathological oral tissue. Informed consent was obtained from all participants prior to data acquisition. The recruitment strategy was designed to capture a broad range of healthy oral tissue presentations to ensure the generalizability of HSI data across different anatomical sites and patient demographics. For each patient, 5 images were taken, including views of the right and left cheeks, the palate, the back of the tongue, and the closed row of teeth. The data were subsequently pseudonymously archived using camera-specific software tools.</p></sec><sec id="s2-3"><title>Red, Green, and Blue&#x2013;Imaging and Endoscopic HSI Data Files</title><p>The Comprehensive Oral Health Hyperspectral Dataset (comprehensive, annotated endoscopic HSI dataset of the oral cavity) primarily comprises RGB (red, green, and blue) and endoscopic HSI data with corresponding annotations. These components in the presented dataset are each important to address the fundamental challenges in DL for semantic segmentation [<xref ref-type="bibr" rid="ref36">36</xref>], providing robust and effective model training. The RGB component offers standard visual spectrum imagery of the oral cavity, serving as a foundational baseline for comparison with more sophisticated imaging modalities like raw and processed endoscopic HSI data.</p></sec><sec id="s2-4"><title>Annotation Process and Verification</title><p>Individual annotations of the oral cavity&#x2019;s anatomical sites were manually performed using an image annotation tool (RectLabel Pro version 2024.06.07, Ryo Kawamura; Tokyo, Japan). To ensure high-quality image data annotation, several key features were implemented. First, detailed annotation guidelines were developed to ensure reproducible results. Comprehensive training sessions and practical exercises for annotators were conducted, followed by feedback rounds to enhance their skills. Each image was annotated by at least 2 independent specialists to minimize subjective errors. Regular quality checks were performed through sample reviews by senior specialists to detect inconsistencies. Continuous feedback was given to annotators, and guidelines were regularly updated based on this feedback. Interannotator analysis of 2 segmentation annotation sets (&#x201C;original&#x201D; vs &#x201C;new&#x201D;), each assigning a class label (out of 20 possible objects plus background) to every pixel, was carried out additionally for internal quality assurance. Agreement was quantified using metrics that capture both overall and class-specific consistency: (1) Pixel Accuracy, (2) Cohen Kappa (including and excluding background), and (3) Dice Coefficients, which are mathematically equivalent to the <italic>F</italic><sub>1</sub>-score in the binary case. To ensure a realistic representation of everyday clinical practice, the annotated structures marked for subsequent analysis corresponded to the individual intraoral site and situation, including &#x201C;clutter (n=752),&#x201D; &#x201C;blood (n=3),&#x201D; &#x201C;brackets (n=15),&#x201D; &#x201C;floor of mouth (n=43),&#x201D; &#x201C;gingiva (n=511),&#x201D; &#x201C;implant (n=10),&#x201D; &#x201C;lip (n=603),&#x201D; &#x201C;mucosa (n=745),&#x201D; &#x201C;palate (n=539),&#x201D; &#x201C;prosthesis (n=65),&#x201D; &#x201C;reflection (n=106),&#x201D; &#x201C;retractor (n=702),&#x201D; &#x201C;suture (n=2),&#x201D; &#x201C;telescopic crown (n=28),&#x201D; &#x201C;tongue (n=587)&#x201D; and &#x201C;tooth (n=681),&#x201D; with a number of distinct classes varying from a minimal count of 2 for &#x201C;suture&#x201D; to a substantial count of 752 for &#x201C;clutter&#x201D;.</p><p>To minimize diagnostic errors, undersegmentation of anatomical structures was avoided, and experienced clinical experts reviewed each annotation to ensure that all clinically relevant features were thoroughly labeled. This conservative approach occasionally led to small unannotated gaps between each class to prevent the merging of different anatomical sites. The digital annotations in XML format provided a semantic comprehension of the image data, which are essential for the application and training of machine learning and computer vision to differentiate various oral site conditions based on the RGB and endoscopic HSI components (<xref ref-type="fig" rid="figure2">Figure 2</xref>).</p><p>Preprocessing further involved the conversion of the original hyperspectral data from .DAT format to the NumPy format (.npy) to facilitate interoperability, reproducibility, and ease of use for DL models. Regarding data augmentation, spatial transformations, such as image rotations to enhance model robustness, were used. The addition of noise and blur was intentionally avoided, as these operations could introduce spectral distortions that are not physically plausible in the HSI context.</p><p>Given the specific goal of oral structure differentiation, classes with low incidence, such as suture, blood, implant, brackets, telescopic crown, prosthesis, reflection, and floor of mouth were deemed less relevant and excluded from further analysis to prevent skewing the learning process and improve the model&#x2019;s ability to generalize to more common and diagnostically relevant structures. By focusing on higher-incidence classes, such as palate, retractor, mucosa, tongue, tooth, lip, and gingiva, the training process is optimized, enhancing the model&#x2019;s efficiency and effectiveness across diverse oral structures. To achieve representative results for the annotation, an additional interannotator analysis and validation was carried out to detect possible person-dependent deviations.</p><fig position="float" id="figure2"><label>Figure 2.</label><caption><p>Examples highlighting the detailed annotations, red, green, blue imagery, and features within endoscopic hyperspectral imaging Band 9 (wavelength of 545 nm), which are especially discernible for human eyes.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v13i1e76148_fig02.png"/></fig></sec><sec id="s2-5"><title>Dataset Structure</title><p>To ensure user-friendly analysis of the hyperspectral dataset and reliable collaboration in the research community, the dataset has been specifically optimized for the Python ecosystem (Python Software Foundation 2023, Python Language Reference, version 3.10) instead of the proprietary software required to load endoscopic HSI data. Regarding scientific computation, Python emerges as the predominant programming language due to its extensive libraries and robust community support. After acknowledging this prevalent preference and recognizing the NPY file format as the most straightforward option for loading data into NumPy, a pivotal library for numerical computations in Python [<xref ref-type="bibr" rid="ref37">37</xref>], all elements of the comprehensive, annotated endoscopic HSI dataset of the oral cavity (&#x201C;RGB,&#x201D; &#x201C;Annotations,&#x201D; &#x201C;HSI_Data_Files&#x201D;) were archived in the &#x201C;.npy&#x201D; format. The NPY format offers several advantages: It is compact and efficient for loading and saving and supports a wide range of data types. This standardization eliminates the complexity and potential incompatibilities associated with proprietary formats. In order to provide high-quality, detailed annotations of the complex intraoral structures, digital annotations were elaborated in XML format. To enhance utility and facilitate integration into researchers&#x2019; workflows, a semantic map of the oral cavity was generated directly from the XML datasets for each image. These maps essentially consisted of binary or multiclass masks, delineating the categorization of each pixel based on the detailed information provided in XML. However, the creation of these maps is an automated process that translates the complex structured information of the XML annotations into a simple but comprehensive format that can easily be integrated into machine learning models to differentiate tissue types based on the values of each pixel corresponding to the specific annotated category. The workflow in short: First, eHSI endoscopic HSI data are loaded, and semantic information is extracted from the XML file. This information is then used to generate a semantic map of the oral cavity, where each pixel value corresponds to a specific annotated category, such as different tissue types or anatomical sites. The resulting semantic map provides comprehensive, pixel-wise annotations of the frame, ready for application in segmentation algorithms.</p></sec><sec id="s2-6"><title>Statistical Analysis of the Dataset</title><p>For an initial assessment of the spectral data, a coefficient of variation (CV) was used as a statistical measure. The CV is a standardized measure for dispersion of a probability distribution or frequency distribution and defines the ratio of the SD to the mean, expressed as percentage. It is particularly useful in the context of HSI, as it allows the comparison of variation between different classes and bands despite the different mean intensities. The formula used for the coefficient of variation is:</p><disp-formula id="E2"><mml:math id="eqn1"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mi>C</mml:mi><mml:mi>V</mml:mi><mml:mo>=</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mfrac><mml:mi>&#x03C3;</mml:mi><mml:mi>&#x03BC;</mml:mi></mml:mfrac><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mn>100</mml:mn><mml:mi mathvariant="normal">%</mml:mi></mml:mstyle></mml:mrow></mml:mstyle></mml:math></disp-formula><p>where <inline-formula><mml:math id="ieqn1"><mml:mi>&#x03C3;</mml:mi></mml:math></inline-formula> is the SD of the dataset and  <inline-formula><mml:math id="ieqn2"><mml:mi>&#x03BC;</mml:mi></mml:math></inline-formula> is the mean of the dataset.</p></sec><sec id="s2-7"><title>Machine Learning and DL Techniques</title><p>This study adapted several state-of-the-art models and their performance assessing semantic segmentation in the context of oral health diagnostics. These models included DeepLabv3 [<xref ref-type="bibr" rid="ref38">38</xref>-<xref ref-type="bibr" rid="ref40">40</xref>], as well as fully convolutional network (FCN [<xref ref-type="bibr" rid="ref41">41</xref>]) and pyramid scene parsing network (PSPNet [<xref ref-type="bibr" rid="ref42">42</xref>]), each tested with ResNet-50 and ResNet-101 [<xref ref-type="bibr" rid="ref43">43</xref>], backbones. In addition, PSPNet using VGG16 [<xref ref-type="bibr" rid="ref44">44</xref>] and U-Net [<xref ref-type="bibr" rid="ref45">45</xref>] models using both EfficientNet-B0 [<xref ref-type="bibr" rid="ref46">46</xref>] and ResNet-50 backbones were evaluated. Each model was optimized for processing our dataset, which features challenging anatomical structures, enabling a comprehensive analysis of their suitability for handling the variability inherent in endoscopic HSI data for oral health diagnosis. The models were trained over 50 epochs using 70% of the comprehensive, annotated endoscopic HSI dataset of the oral cavity test set, with the remaining 30% reserved for evaluation. This split allowed for a thorough assessment of the models&#x2019; ability to generalize to unseen data. Each model&#x2019;s performance on endoscopic HSI data was evaluated using Precision, Recall, and the <italic>F</italic><sub>1</sub>-score, providing a comprehensive measure of their effectiveness in segmenting and classifying the various anatomical structures present in the dataset.</p></sec><sec id="s2-8"><title>Ethical Considerations</title><p>The study was approved by the local ethics committee of Rhineland-Palatinate (registration number: 2021-16158) and was conducted in accordance with the protocol and in compliance with the moral, ethical, and scientific principles governing clinical research as set out in the Declaration of Helsinki of 1975 as revised in 1983. Informed consent was obtained from all participants involved in the study. This study received approval from the Ethics Committee of the Medical Association of Rhineland-Palatinate (reference number 2021-15858). All procedures adhered to institutional and national ethical standards and were conducted in accordance with the Declaration of Helsinki. Participants were provided comprehensive information regarding the study&#x2019;s nature, purpose, procedures, data usage, and the potential publication of anonymized images or data related to their participation. The privacy and confidentiality of all participants were rigorously protected. No identifying information, including names or hospital IDs, has been included in the manuscript. Images containing identifiable features were excluded except for Figure 1, which depicts the corresponding author (PR), who has provided written consent for the use of their image in this publication. No financial or material compensation was offered or provided to the participants in this study.</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Variability of the Dataset</title><p>A descriptive statistical analysis of the hyperspectral dataset was performed to describe the distribution and variability of the data. Mean values and SD across the spectral bands were analyzed for each class delineated in the hyperspectral images. The objective of the statistical data evaluation was to accurately represent the central tendencies and variabilities of spectral signatures, which are indicative of various anatomical and pathological entities in oral health. Mean values and SDs for all classes across all spectral bands were presented in a consolidated diagram (<xref ref-type="fig" rid="figure3">Figure 3</xref>).</p><p>The results indicated that the CV for the majority of classes, across most spectral bands, substantially exceeded the commonly recognized threshold of 15%, which is generally considered high. In contrast, a CV below 5% would have denoted homogeneous data, while values between 5% and 15% would have indicated moderate uniformity. The consistently elevated CV values observed across all classes thus underscored significant spectral diversity within each class (<xref ref-type="fig" rid="figure4">Figure 4</xref>).</p><fig position="float" id="figure3"><label>Figure 3.</label><caption><p>Consolidated diagram of mean values and SD for each class across all measured spectral bands.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v13i1e76148_fig03.png"/></fig><fig position="float" id="figure4"><label>Figure 4.</label><caption><p>Coefficient of variation over the corresponding bandwidth of various examination classes. CV: coefficient of variation.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v13i1e76148_fig04.png"/></fig></sec><sec id="s3-2"><title>DL Model Performance</title><p>The DeepLabv3 model with the ResNet-50 backbone achieved solid overall performance, with an <italic>F</italic><sub>1</sub>-score of 0.855, performing especially well on the segmentation of mucosa, retractor, and tooth. However, it exhibited moderate performance on gingiva and lip, with <italic>F</italic><sub>1</sub>-scores of 0.753 and 0.709, respectively. Switching to the deeper ResNet-101 backbone improved the overall performance slightly, with an <italic>F</italic><sub>1</sub>-score of 0.857, particularly enhancing the model&#x2019;s ability to segment clutter and gingiva (<xref ref-type="table" rid="table1">Table 1</xref>).</p><p>Both FCN-ResNet-50 and FCN-ResNet-101 demonstrated robust segmentation performance, achieving overall <italic>F</italic><sub>1</sub>-scores of 0.862 and 0.861, respectively (<xref ref-type="table" rid="table2">Table 2</xref>). They excelled in segmenting retractor (<italic>F</italic><sub>1</sub>-score=0.942 for both), tooth (<italic>F</italic><sub>1</sub>-score=0.910), palate (<italic>F</italic><sub>1</sub>-score=0.890), and mucosa (<italic>F</italic><sub>1</sub>-score&#x2265;0.912). However, lip and gingiva posed challenges for both backbones, reflected by lower <italic>F</italic><sub>1</sub>-scores around 0.72 and 0.77, respectively. Although switching to the deeper ResNet-101 backbone yielded marginal improvements in certain classes, the overall performance remained comparable between the 2 architectures.</p><p>PSPNet models showed slightly lower performance, with the ResNet-50 backbone achieving an <italic>F</italic><sub>1</sub>-score of 0.837 and the VGG16 backbone scoring 0.808. While these models handled classes, such as retractor and tooth, relatively well, they faced difficulties in the segmentation of gingiva and lip, particularly when using VGG16. The lower overall performance of PSPNet VGG16 compared to ResNet-50 reflects the influence of the backbone on the segmentation outcomes (<xref ref-type="table" rid="table3">Table 3</xref>).</p><p>U-Net (EfficientNet-B0) and U-Net (ResNet-50) both demonstrated robust segmentation performance, attaining overall <italic>F</italic><sub>1</sub>-scores of 0.867 and 0.840, respectively (<xref ref-type="table" rid="table4">Table 4</xref>). They excelled particularly at segmenting retractor (<italic>F</italic><sub>1</sub>-score=0.941 vs 0.927), palate (<italic>F</italic><sub>1</sub>-score=0.909 vs 0.881), and mucosa (<italic>F</italic><sub>1</sub>-score=0.920 vs 0.901). Tongue and tooth segmentation also performed well, with F1-scores exceeding 0.85 for both models. However, both networks faced challenges in segmenting lip (<italic>F</italic><sub>1</sub>-score=0.755 vs 0.695) and gingiva (<italic>F</italic><sub>1</sub>-score=0.751 vs 0.721), indicating opportunities for further refinement in these classes.</p><p><xref ref-type="fig" rid="figure5">Figure 5</xref> demonstrates that DeepLabv3 (ResNet-101) and U-Net (EfficientNet-B0) outperform other models across most tissue classes, with lower <italic>F</italic><sub>1</sub>-scores consistently observed for lip and gingiva segmentation.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Results showing precision, recall, and <italic>F</italic><sub>1</sub>-score for different classes in a comparison of DeepLabv3 ResNet-50 and DeepLabv3 ResNet-101.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="top" rowspan="2">Class</td><td align="left" valign="top" colspan="2">Precision</td><td align="left" valign="top" colspan="2">Recall</td><td align="left" valign="top" colspan="2"><italic>F</italic><sub>1</sub>-score</td></tr><tr><td align="left" valign="bottom">DeepLabv3 ResNet-50</td><td align="left" valign="bottom">DeepLabv3 ResNet-101</td><td align="left" valign="bottom">DeepLabv3 ResNet-50</td><td align="left" valign="bottom">DeepLabv3 ResNet-101</td><td align="left" valign="bottom">DeepLabv3 ResNet-50</td><td align="left" valign="bottom">DeepLabv3 ResNet-101</td></tr></thead><tbody><tr><td align="left" valign="top">Overall</td><td align="left" valign="top">0.849</td><td align="left" valign="top">0.851</td><td align="left" valign="top">0.861</td><td align="left" valign="top">0.863</td><td align="left" valign="top">0.855</td><td align="left" valign="top">0.857</td></tr><tr><td align="left" valign="top">Palate</td><td align="left" valign="top">0.886</td><td align="left" valign="top">0.874</td><td align="left" valign="top">0.918</td><td align="left" valign="top">0.929</td><td align="left" valign="top">0.901</td><td align="left" valign="top">0.900</td></tr><tr><td align="left" valign="top">Retractor</td><td align="left" valign="top">0.936</td><td align="left" valign="top">0.942</td><td align="left" valign="top">0.940</td><td align="left" valign="top">0.938</td><td align="left" valign="top">0.938</td><td align="left" valign="top">0.940</td></tr><tr><td align="left" valign="top">Mucosa</td><td align="left" valign="top">0.915</td><td align="left" valign="top">0.914</td><td align="left" valign="top">0.914</td><td align="left" valign="top">0.916</td><td align="left" valign="top">0.914</td><td align="left" valign="top">0.915</td></tr><tr><td align="left" valign="top">Tongue</td><td align="left" valign="top">0.878</td><td align="left" valign="top">0.872</td><td align="left" valign="top">0.888</td><td align="left" valign="top">0.892</td><td align="left" valign="top">0.883</td><td align="left" valign="top">0.882</td></tr><tr><td align="left" valign="top">Clutter</td><td align="left" valign="top">0.840</td><td align="left" valign="top">0.844</td><td align="left" valign="top">0.848</td><td align="left" valign="top">0.849</td><td align="left" valign="top">0.844</td><td align="left" valign="top">0.846</td></tr><tr><td align="left" valign="top">Tooth</td><td align="left" valign="top">0.862</td><td align="left" valign="top">0.871</td><td align="left" valign="top">0.937</td><td align="left" valign="top">0.936</td><td align="left" valign="top">0.898</td><td align="left" valign="top">0.902</td></tr><tr><td align="left" valign="top">Lip</td><td align="left" valign="top">0.723</td><td align="left" valign="top">0.730</td><td align="left" valign="top">0.695</td><td align="left" valign="top">0.683</td><td align="left" valign="top">0.709</td><td align="left" valign="top">0.706</td></tr><tr><td align="left" valign="top">Gingiva</td><td align="left" valign="top">0.756</td><td align="left" valign="top">0.757</td><td align="left" valign="top">0.751</td><td align="left" valign="top">0.765</td><td align="left" valign="top">0.754</td><td align="left" valign="top">0.761</td></tr></tbody></table></table-wrap><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>Results showing precision, recall, and <italic>F</italic><sub>1</sub>-score for different classes in a comparison of FCN-ResNet-50 and FCN-ResNet-101.</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="top">Class</td><td align="left" valign="top" colspan="2">Precision</td><td align="left" valign="top" colspan="2">Recall</td><td align="left" valign="top" colspan="2"><italic>F</italic><sub>1</sub>-score</td></tr><tr><td align="left" valign="bottom"/><td align="left" valign="bottom">FCN<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup>-ResNet-50</td><td align="left" valign="bottom">FCN-ResNet-101</td><td align="left" valign="bottom">FCN<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup>-ResNet-50</td><td align="left" valign="bottom">FCN-ResNet-101</td><td align="left" valign="bottom">FCN<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup>-ResNet-50</td><td align="left" valign="bottom">FCN-ResNet-101</td></tr></thead><tbody><tr><td align="left" valign="top">Overall</td><td align="left" valign="top">0.850</td><td align="left" valign="top">0.850</td><td align="left" valign="top">0.874</td><td align="left" valign="top">0.873</td><td align="left" valign="top">0.862</td><td align="left" valign="top">0.861</td></tr><tr><td align="left" valign="top">Palate</td><td align="left" valign="top">0.859</td><td align="left" valign="top">0.860</td><td align="left" valign="top">0.924</td><td align="left" valign="top">0.923</td><td align="left" valign="top">0.890</td><td align="left" valign="top">0.890</td></tr><tr><td align="left" valign="top">Retractor</td><td align="left" valign="top">0.940</td><td align="left" valign="top">0.935</td><td align="left" valign="top">0.944</td><td align="left" valign="top">0.949</td><td align="left" valign="top">0.942</td><td align="left" valign="top">0.942</td></tr><tr><td align="left" valign="top">Mucosa</td><td align="left" valign="top">0.904</td><td align="left" valign="top">0.907</td><td align="left" valign="top">0.921</td><td align="left" valign="top">0.929</td><td align="left" valign="top">0.912</td><td align="left" valign="top">0.916</td></tr><tr><td align="left" valign="top">Tongue</td><td align="left" valign="top">0.881</td><td align="left" valign="top">0.882</td><td align="left" valign="top">0.909</td><td align="left" valign="top">0.893</td><td align="left" valign="top">0.895</td><td align="left" valign="top">0.887</td></tr><tr><td align="left" valign="top">Clutter</td><td align="left" valign="top">0.864</td><td align="left" valign="top">0.863</td><td align="left" valign="top">0.839</td><td align="left" valign="top">0.843</td><td align="left" valign="top">0.852</td><td align="left" valign="top">0.853</td></tr><tr><td align="left" valign="top">Tooth</td><td align="left" valign="top">0.878</td><td align="left" valign="top">0.876</td><td align="left" valign="top">0.946</td><td align="left" valign="top">0.910</td><td align="left" valign="top">0.910</td><td align="left" valign="top">0.910</td></tr><tr><td align="left" valign="top">Lip</td><td align="left" valign="top">0.699</td><td align="left" valign="top">0.718</td><td align="left" valign="top">0.744</td><td align="left" valign="top">0.723</td><td align="left" valign="top">0.721</td><td align="left" valign="top">0.721</td></tr><tr><td align="left" valign="top">Gingiva</td><td align="left" valign="top">0.776</td><td align="left" valign="top">0.765</td><td align="left" valign="top">0.767</td><td align="left" valign="top">0.777</td><td align="left" valign="top">0.772</td><td align="left" valign="top">0.771</td></tr></tbody></table><table-wrap-foot><fn id="table2fn1"><p><sup>a</sup>FCN: fully convolutional network.</p></fn></table-wrap-foot></table-wrap><table-wrap id="t3" position="float"><label>Table 3.</label><caption><p>Results showing precision, recall, and <italic>F</italic><sub>1</sub>-score for different classes in a comparison of PSPNet-ResNet-50<sup><xref ref-type="table-fn" rid="table3fn1">a</xref></sup> and PSPNet-VGG16.</p></caption><table id="table3" frame="hsides" rules="groups"><thead><tr><td align="left" valign="top">Class</td><td align="left" valign="top" colspan="2">Precision</td><td align="left" valign="top" colspan="2">Recall</td><td align="left" valign="top" colspan="2"><italic>F</italic><sub>1</sub>-score</td></tr><tr><td align="left" valign="bottom"/><td align="left" valign="bottom">PSPNet-ResNet-50</td><td align="left" valign="bottom">PSPNet-VGG16</td><td align="left" valign="bottom">PSPNet-ResNet-50</td><td align="left" valign="bottom">PSPNet-VGG16</td><td align="left" valign="bottom">PSPNet-ResNet-50</td><td align="left" valign="bottom">PSPNet-VGG16</td></tr></thead><tbody><tr><td align="left" valign="top">Overall</td><td align="left" valign="top">0.849</td><td align="left" valign="top">0.797</td><td align="left" valign="top">0.830</td><td align="left" valign="top">0.822</td><td align="left" valign="top">0.837</td><td align="left" valign="top">0.809</td></tr><tr><td align="left" valign="top">Palate</td><td align="left" valign="top">0.840</td><td align="left" valign="top">0.854</td><td align="left" valign="top">0.915</td><td align="left" valign="top">0.840</td><td align="left" valign="top">0.876</td><td align="left" valign="top">0.847</td></tr><tr><td align="left" valign="top">Retractor</td><td align="left" valign="top">0.929</td><td align="left" valign="top">0.853</td><td align="left" valign="top">0.919</td><td align="left" valign="top">0.906</td><td align="left" valign="top">0.924</td><td align="left" valign="top">0.879</td></tr><tr><td align="left" valign="top">Mucosa</td><td align="left" valign="top">0.906</td><td align="left" valign="top">0.835</td><td align="left" valign="top">0.842</td><td align="left" valign="top">0.841</td><td align="left" valign="top">0.873</td><td align="left" valign="top">0.838</td></tr><tr><td align="left" valign="top">Tongue</td><td align="left" valign="top">0.903</td><td align="left" valign="top">0.794</td><td align="left" valign="top">0.837</td><td align="left" valign="top">0.886</td><td align="left" valign="top">0.869</td><td align="left" valign="top">0.837</td></tr><tr><td align="left" valign="top">Clutter</td><td align="left" valign="top">0.793</td><td align="left" valign="top">0.807</td><td align="left" valign="top">0.873</td><td align="left" valign="top">0.779</td><td align="left" valign="top">0.831</td><td align="left" valign="top">0.793</td></tr><tr><td align="left" valign="top">Tooth</td><td align="left" valign="top">0.877</td><td align="left" valign="top">0.865</td><td align="left" valign="top">0.910</td><td align="left" valign="top">0.907</td><td align="left" valign="top">0.893</td><td align="left" valign="top">0.886</td></tr><tr><td align="left" valign="top">Lip</td><td align="left" valign="top">0.766</td><td align="left" valign="top">0.716</td><td align="left" valign="top">0.701</td><td align="left" valign="top">0.693</td><td align="left" valign="top">0.732</td><td align="left" valign="top">0.704</td></tr><tr><td align="left" valign="top">Gingiva</td><td align="left" valign="top">0.775</td><td align="left" valign="top">0.651</td><td align="left" valign="top">0.639</td><td align="left" valign="top">0.723</td><td align="left" valign="top">0.700</td><td align="left" valign="top">0.685</td></tr></tbody></table><table-wrap-foot><fn id="table3fn1"><p><sup>a</sup>PSPNet: pyramid scene parsing network.</p></fn></table-wrap-foot></table-wrap><table-wrap id="t4" position="float"><label>Table 4.</label><caption><p>Results showing precision, recall, and <italic>F</italic><sub>1</sub>-score for different classes in a comparison of U-Net-EfficientNet-B0 and U-Net-Res</p></caption><table id="table4" frame="hsides" rules="groups"><thead><tr><td align="left" valign="top">Class</td><td align="left" valign="top" colspan="2">Precision</td><td align="left" valign="top" colspan="2">Recall</td><td align="left" valign="top" colspan="2"><italic>F</italic><sub>1</sub>-score</td></tr><tr><td align="left" valign="bottom"/><td align="left" valign="bottom">U-Net-EfficientNet-B0</td><td align="left" valign="bottom">U-Net-ResNet-50</td><td align="left" valign="bottom">U-Net-EfficientNet-B0</td><td align="left" valign="bottom">U-Net-ResNet-50</td><td align="left" valign="bottom">U-Net-EfficientNet-B0</td><td align="left" valign="bottom">U-Net-ResNet-50</td></tr></thead><tbody><tr><td align="left" valign="top">Overall</td><td align="left" valign="top">0.853</td><td align="left" valign="top">0.846</td><td align="left" valign="top">0.882</td><td align="left" valign="top">0.836</td><td align="left" valign="top">0.867</td><td align="left" valign="top">0.840</td></tr><tr><td align="left" valign="top">Palate</td><td align="left" valign="top">0.892</td><td align="left" valign="top">0.857</td><td align="left" valign="top">0.927</td><td align="left" valign="top">0.907</td><td align="left" valign="top">0.909</td><td align="left" valign="top">0.881</td></tr><tr><td align="left" valign="top">Retractor</td><td align="left" valign="top">0.956</td><td align="left" valign="top">0.936</td><td align="left" valign="top">0.927</td><td align="left" valign="top">0.918</td><td align="left" valign="top">0.941</td><td align="left" valign="top">0.927</td></tr><tr><td align="left" valign="top">Mucosa</td><td align="left" valign="top">0.915</td><td align="left" valign="top">0.918</td><td align="left" valign="top">0.924</td><td align="left" valign="top">0.884</td><td align="left" valign="top">0.920</td><td align="left" valign="top">0.901</td></tr><tr><td align="left" valign="top">Tongue</td><td align="left" valign="top">0.893</td><td align="left" valign="top">0.856</td><td align="left" valign="top">0.895</td><td align="left" valign="top">0.889</td><td align="left" valign="top">0.894</td><td align="left" valign="top">0.872</td></tr><tr><td align="left" valign="top">Clutter</td><td align="left" valign="top">0.862</td><td align="left" valign="top">0.811</td><td align="left" valign="top">0.846</td><td align="left" valign="top">0.862</td><td align="left" valign="top">0.854</td><td align="left" valign="top">0.835</td></tr><tr><td align="left" valign="top">Tooth</td><td align="left" valign="top">0.876</td><td align="left" valign="top">0.875</td><td align="left" valign="top">0.949</td><td align="left" valign="top">0.894</td><td align="left" valign="top">0.911</td><td align="left" valign="top">0.885</td></tr><tr><td align="left" valign="top">Lip</td><td align="left" valign="top">0.711</td><td align="left" valign="top">0.734</td><td align="left" valign="top">0.806</td><td align="left" valign="top">0.661</td><td align="left" valign="top">0.755</td><td align="left" valign="top">0.695</td></tr><tr><td align="left" valign="top">Gingiva</td><td align="left" valign="top">0.723</td><td align="left" valign="top">0.780</td><td align="left" valign="top">0.782</td><td align="left" valign="top">0.671</td><td align="left" valign="top">0.751</td><td align="left" valign="top">0.721</td></tr></tbody></table></table-wrap><fig position="float" id="figure5"><label>Figure 5.</label><caption><p>Spider plot illustrating the per-class <italic>F</italic><sub>1</sub>-score performance of different deep learning models for oral tissue segmentation.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v13i1e76148_fig05.png"/></fig></sec><sec id="s3-3"><title>Interannotator Agreement</title><p>The analysis of segmentation annotations revealed an overall pixel accuracy of 80.84%, indicating a high level of interannotator agreement. Cohen Kappa, a more robust measure of interannotator agreement, was 77.56% when including the background class, indicating substantial agreement. When the background was excluded, Kappa increased to 82.91%, highlighting even stronger agreement for the segmentation of foreground structures. Dice coefficients per class ranged from near-zero (in classes with minimal or inconsistent labeling) to 93% in well-defined classes, where the overlap between annotations was nearly perfect. Since the Dice Coefficient measures spatial overlap between segmentation masks, high Dice scores (above 90%) indicate excellent agreement, whereas lower values suggest potential ambiguity or underrepresentation of certain structures.</p></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Principal Findings and Comparison With Previous Works</title><p>The objective of this study was to establish a digital map of the oral cavity using endoscopic HSI in conjunction with advanced machine learning techniques and therefore to develop a comprehensive endoscopic HSI dataset. This approach aimed to enable reliable and automated differentiation of various tissue types and objects based on the spectral data acquired from the endoscopic HSI system. Despite the continuous interest in faster, minimally invasive diagnostic techniques, methods, such as brush biopsies and in vivo fluorescence procedures have faced challenges in establishing themselves as reliable alternatives due to their comparatively lower sensitivity and specificity [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref5">5</xref>,<xref ref-type="bibr" rid="ref9">9</xref>]. Modern methods, such as endoscopic HSI, represent a state-of-the-art, innovative approach in the field of automated image and tissue classification. In our preliminary ex vivo studies, we have already demonstrated that HSI can differentiate between various tissue types and states based on specific wavelength patterns [<xref ref-type="bibr" rid="ref20">20</xref>,<xref ref-type="bibr" rid="ref47">47</xref>]. However, to effectively differentiate between pathological and healthy tissue conditions in vivo, a substantial dataset of hyperspectral signatures from healthy tissues is essential [<xref ref-type="bibr" rid="ref47">47</xref>].</p><p>The study introduces a comprehensive collection of 1,130,751 endoscopic HSI&#x2013;cubes of healthy oral mucosa, captured in vivo from various angles, creating a representative digital map that includes relevant tissues and objects of the oral cavity. This dataset forms a representative digital endoscopic HSI map, encompassing relevant tissues and structures of the oral cavity. By establishing a robust reference for healthy tissue, this dataset lays the groundwork for advancing HSI-based diagnostics, particularly in the identification of premalignant and malignant mucosal lesions and the precise definition of tumor resection margins.</p><p>The findings revealed elevated CV across most classes, indicating substantial spectral diversity in endoscopic HSI data. This high variability poses challenges for threshold-based classification methods, as intraclass variation can lead to class overlap. However, this diversity also holds valuable information that, when leveraged by advanced computational models, can enhance tissue classification and pathological anomaly detection with high accuracy.</p><p>To address this complexity, this study evaluated several state-of-the-art segmentation models, including DeepLabv3, FCN, PSPNet, and U-Net, with different backbones, such as ResNet-50, ResNet-101, VGG16, and EfficientNet-B0. These architectures were chosen for their effectiveness in semantic segmentation, balancing feature extraction capability, computational efficiency, and global context recognition. DeepLabv3 was selected for its ability to handle scale variability and capture fine details while FCN served as a strong baseline due to its foundational role in image segmentation. PSPNet was included for its strong capability in capturing global context, which is crucial for recognizing complex structures in medical imaging. U-Net, widely used in medical imaging, was chosen for its ability to achieve accurate segmentation even with limited data.</p><p>The choice of backbones was guided by their specific strengths. ResNet-50 and ResNet-101 were selected for their robust feature extraction, VGG16 for its simplicity and high-resolution detail, and EfficientNet-B0 for its optimized architecture that balances performance and computational efficiency. By incorporating models with varying depths and parameter complexities, this study ensures a comprehensive evaluation of segmentation performance while maintaining a focus on efficient training and inference times. The integration of the DeepLabv3 model with the ResNet-50 as well as with the ResNet-101 backbone offers a robust semantic segmentation approach for endoscopic HSI data interpretation. The primary modification to all the models, including DeepLabv3, FCN, PSPNet, and U-Net (with backbones, such as ResNet-50, ResNet-101, EfficientNet-B0, and VGG16), involved adapting the first convolutional layer to handle the variable number of channels in the dataset. This adjustment was essential to accommodate the multidimensional spectral data of the HSI dataset, as opposed to the standard 3-channel RGB imagery. This modification enables the models to use the unique spectral information in the HSI data cubes beyond the visible spectrum. Despite these adjustments, the deep residual learning architectures of ResNet-50 and ResNet-101 retain their capacity to extract high-level features&#x2014;an essential aspect of HSI data analysis [<xref ref-type="bibr" rid="ref43">43</xref>]. These backbones excel at identifying subtle spectral patterns crucial for accurate disease detection and classification. Leveraging atrous convolution and atrous spatial pyramid pooling, the DeepLabv3 model efficiently captures multiscale information [<xref ref-type="bibr" rid="ref48">48</xref>]. This capability facilitates precise image segmentation, an essential feature for diagnosing oral health conditions with subtle and overlapping visual signatures. Likewise, FCN and PSPNet, with their robust architectures, and U-Net, recognized for its effectiveness in medical imaging, also benefited from the capacity to process and interpret multidimensional spectral data.</p><p>The endoscopic HSI procedure implemented in this study achieves results that are comparable to those obtained in preliminary HSI-ex vivo trials conducted by our research group in 2021 [<xref ref-type="bibr" rid="ref47">47</xref>]. In this study, using a similar approach with a lightweight 6-layer deep neural network containing 10,445 parameters trained over 4000 epochs, tissue samples of fat, muscle, and oral mucosa could be differentiated with an overall class accuracy of over 80%. Similarly, in a study by Ma et al, tissue from various organs, including the kidney, liver, lung, muscle, salivary gland, and spleen, was identified using automated polarized HSI with an accuracy of up to 87% [<xref ref-type="bibr" rid="ref49">49</xref>]. Furthermore, a deep convolutional neural network established by Poonkuzhali et al [<xref ref-type="bibr" rid="ref50">50</xref>] could accurately identify brain tissue with an <italic>F</italic><sub>1</sub>-score precision of 97.3% using HSI in a recent study conducted in 2023. Unlike previous studies that focused on ex vivo tissue samples and organ-specific HSI data [<xref ref-type="bibr" rid="ref47">47</xref>], the present work introduces the first large-scale in vivo annotated endoscopic HSI dataset of the oral cavity acquired under realistic clinical conditions. While earlier studies demonstrated the technical feasibility of HSI-based tissue differentiation, these approaches often lacked anatomical complexity, semantic annotations, and clinical variability. This study addresses these deficiencies by providing a large annotated in vivo HSI dataset and evaluating multiple DL models especially adapted for HSI data analysis. This provides the basis for clinically applicable segmentation of oral tissues and paves the way for future studies involving pathological lesions.</p><p>These findings affirm that HSI, when integrated with appropriate modalities, can serve as a reliable tool for differentiating various tissues in both in vivo and ex vivo setups. In oncological surgery, fast, reliable, and minimally invasive diagnosis of pathological tissue conditions is of paramount importance. Numerous studies in this field have shown that HSI can dependably differentiate between tumor and healthy tissue by analyzing histopathological sections and their spatial-spectral features. Moreover, the combination of HSI and DL has shown superior results compared to the use of RGB images and conventional support-vector-machine approaches [<xref ref-type="bibr" rid="ref51">51</xref>-<xref ref-type="bibr" rid="ref56">56</xref>].</p><p>In this study, the established DL and neural network approach was able to accurately identify different types of oral mucosa with an overall precision of approximately 91%. DeepLabv3 (ResNet-101) and U-Net (EfficientNet-B0) emerged as the top performers, demonstrating robust segmentation across key anatomical classes. While all models could benefit from further refinement in gingiva and lip segmentation, their consistent accuracy in identifying retractors and teeth underscores the strength of the dataset. While there is no universally accepted threshold for clinical applicability, <italic>F</italic><sub>1</sub>-scores above 0.85 are generally considered promising in similar biomedical imaging tasks. Scores in this range suggest that the method may already approach a level of accuracy relevant for clinical decision-making, though further validation in real-world settings would be required. Notably, U-Net (EfficientNet-B0) exhibited exceptional segmentation of retractors, mucosa, and teeth, indicating its strong potential for medical image segmentation applications. The results demonstrate the dataset&#x2019;s suitability for training advanced neural networks, particularly in challenging medical segmentation tasks.</p><p>Limitations of our study specifically include the acquisition of a higher amount of data, which depended heavily on the individual patient volume of the clinic conducting the study. This represents a major challenge, particularly when aiming to document various oral mucosal lesions and precancerous mucosal conditions. To address the issue of data scarcity, the raw data generated in this study will be made publicly available. This initiative aims to provide other research groups with the opportunity to use and build upon the dataset for further investigations. Furthermore, while the study cohort was intentionally designed to include a broad cross-section of patients from a university-based outpatient clinic, selection bias cannot be fully excluded. As recruitment was confined to a single university clinic, the cohort may not fully reflect the demographic and clinical variability found in primary care or the general population. Other limitations include variable lighting conditions, motion artifacts, and spectral overlaps, as well as overexposure and underexposure. Overexposure can occur due to various factors, such as the variability in patient anatomy, the movement during image capture, or the fluctuation of lighting conditions in an operating room. While overexposed images are typically regarded as artifacts and excluded from datasets, our approach incorporates them, acknowledging that such occurrences are an inherent part of clinical practice. Understanding that these instances present both challenges and opportunities for advanced image processing techniques, we have carefully indexed these overexposed images for further analysis. This approach enhances the resilience and adaptability of the models trained on this dataset, ensuring they perform effectively in real clinical environments. By incorporating overexposed images, the models are better equipped to handle the full spectrum of data variability, including common environmental factors. This combination highlights the substantial potential of endoscopic HSI data to significantly enhance the precision and reliability in detecting and categorizing healthy and pathological oral conditions. This convergence not only paves the way for significant advancements in diagnostic methodologies but also holds the potential to substantially enhance patient care. External validation is currently planned as part of future work, particularly for mucosa segmentation. We intend to validate our model on an ex vivo dataset that includes mucosa samples captured under different acquisition conditions. At this stage, the focus of this study was the initial validation on the internally collected dataset. Corresponding follow-up studies are currently in preparation and will include HSI of pathological tissue conditions, such as oral mucosal lesions and neoplastic changes, to assess the transferability and diagnostic robustness of the proposed models in clinically relevant scenarios. By enabling earlier detection and more precise characterization of oral health conditions, this approach facilitates the development of more effective treatment strategies, ultimately leading to improved patient outcomes and a higher standard of care.</p></sec><sec id="s4-2"><title>Conclusion</title><p>This study presents the first large-scale in vivo annotated dataset of the oral cavity using endoscopic HSI under realistic clinical conditions. By combining hyperspectral datasets and DL-based segmentation, it was possible to demonstrate the feasibility of automated, noninvasive tissue classification across important anatomical intraoral structures. Comparative analysis reveals DeepLabv3 and U-Net as robust architectures for oral tissue classification. The comprehensive dataset consequently provides a sufficient foundation for future work on pathological tissue detection, intraoperative margin assessment in oncology, and early, individualized diagnostics in oral medicine.</p></sec></sec></body><back><ack><p>The results of the present study are part of the doctoral thesis of the co-author and student of the University of Mainz, Katharina Kloster, and as such will be made publicly available as part of the doctorate. It was funded by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) &#x2013; Project number 516210826.</p></ack><notes><sec><title>Data Availability</title><p>All raw data on which this study is based will be made available by the corresponding author upon request. The datasets used and/or analyzed during the current study are available from the corresponding author on reasonable request.</p></sec></notes><fn-group><fn fn-type="con"><p>The conceptualization of the study was carried out by DT, JJP, and PR. Methodology development was led by DT, JJP, BP, PR, KK and TK. Validation was performed by DT, JJP, BP, and PR. Formal analysis was conducted by DT, PWK, JJP, BP, PR, and TK. Investigation was carried out by DT, BP, KK and JJP. Resources were provided by DT, JJP, TK and BAN. Data curation was managed by DT, KK, PR and JJP. The original draft was prepared by DT, JJP, PR, SV. Manuscript review and editing were done by DT, JJP, BP, SV, PWK, TK and BAN. Visualization was prepared by DT, JJP and PR. Supervision was conducted by DT, JJP, PR and TK. Project administration was managed by DT, JJP, PR and TK.</p></fn><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">CV</term><def><p>coefficient of variation</p></def></def-item><def-item><term id="abb2">DL</term><def><p>deep learning</p></def></def-item><def-item><term id="abb3">FCN</term><def><p>fully convolutional network</p></def></def-item><def-item><term id="abb4">HSI</term><def><p>hyperspectral imaging</p></def></def-item><def-item><term id="abb5">PSPNet</term><def><p>pyramid scene parsing network</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chen</surname><given-names>S</given-names> </name><name name-style="western"><surname>Forman</surname><given-names>M</given-names> </name><name name-style="western"><surname>Sadow</surname><given-names>PM</given-names> </name><name name-style="western"><surname>August</surname><given-names>M</given-names> </name></person-group><article-title>The diagnostic accuracy of incisional biopsy in the oral cavity</article-title><source>J Oral Maxillofac Surg</source><year>2016</year><month>05</month><volume>74</volume><issue>5</issue><fpage>959</fpage><lpage>964</lpage><pub-id pub-id-type="doi">10.1016/j.joms.2015.11.006</pub-id><pub-id pub-id-type="medline">26682520</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Pentenero</surname><given-names>M</given-names> </name><name name-style="western"><surname>Carrozzo</surname><given-names>M</given-names> </name><name name-style="western"><surname>Pagano</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Oral mucosal dysplastic lesions and early squamous cell carcinomas: underdiagnosis from incisional biopsy</article-title><source>Oral Dis</source><year>2003</year><month>03</month><volume>9</volume><issue>2</issue><fpage>68</fpage><lpage>72</lpage><pub-id pub-id-type="doi">10.1034/j.1601-0825.2003.02875.x</pub-id><pub-id pub-id-type="medline">12657031</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Macey</surname><given-names>R</given-names> </name><name name-style="western"><surname>Walsh</surname><given-names>T</given-names> </name><name name-style="western"><surname>Brocklehurst</surname><given-names>P</given-names> </name><etal/></person-group><article-title>Diagnostic tests for oral cancer and potentially malignant disorders in patients presenting with clinically evident lesions</article-title><source>Cochrane Database Syst Rev</source><year>2015</year><month>05</month><day>29</day><volume>2015</volume><issue>5</issue><fpage>CD010276</fpage><pub-id pub-id-type="doi">10.1002/14651858.CD010276.pub2</pub-id><pub-id pub-id-type="medline">26021841</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Rashid</surname><given-names>A</given-names> </name><name name-style="western"><surname>Warnakulasuriya</surname><given-names>S</given-names> </name></person-group><article-title>The use of light-based (optical) detection systems as adjuncts in the detection of oral cancer and oral potentially malignant disorders: a systematic review</article-title><source>J Oral Pathol Med</source><year>2015</year><month>05</month><volume>44</volume><issue>5</issue><fpage>307</fpage><lpage>328</lpage><pub-id pub-id-type="doi">10.1111/jop.12218</pub-id><pub-id pub-id-type="medline">25183259</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>K&#x00E4;mmerer</surname><given-names>PW</given-names> </name><name name-style="western"><surname>Rahimi-Nedjat</surname><given-names>RK</given-names> </name><name name-style="western"><surname>Ziebart</surname><given-names>T</given-names> </name><etal/></person-group><article-title>A chemiluminescent light system in combination with toluidine blue to assess suspicious oral lesions-clinical evaluation and review of the literature</article-title><source>Clin Oral Investig</source><year>2015</year><month>03</month><volume>19</volume><issue>2</issue><fpage>459</fpage><lpage>466</lpage><pub-id pub-id-type="doi">10.1007/s00784-014-1252-z</pub-id><pub-id pub-id-type="medline">24888605</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bray</surname><given-names>F</given-names> </name><name name-style="western"><surname>Ferlay</surname><given-names>J</given-names> </name><name name-style="western"><surname>Soerjomataram</surname><given-names>I</given-names> </name><name name-style="western"><surname>Siegel</surname><given-names>RL</given-names> </name><name name-style="western"><surname>Torre</surname><given-names>LA</given-names> </name><name name-style="western"><surname>Jemal</surname><given-names>A</given-names> </name></person-group><article-title>Global cancer statistics 2018: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries</article-title><source>CA Cancer J Clin</source><year>2018</year><month>11</month><volume>68</volume><issue>6</issue><fpage>394</fpage><lpage>424</lpage><pub-id pub-id-type="doi">10.3322/caac.21492</pub-id><pub-id pub-id-type="medline">30207593</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Panarese</surname><given-names>I</given-names> </name><name name-style="western"><surname>Aquino</surname><given-names>G</given-names> </name><name name-style="western"><surname>Ronchi</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Oral and Oropharyngeal squamous cell carcinoma: prognostic and predictive parameters in the etiopathogenetic route</article-title><source>Expert Rev Anticancer Ther</source><year>2019</year><month>02</month><volume>19</volume><issue>2</issue><fpage>105</fpage><lpage>119</lpage><pub-id pub-id-type="doi">10.1080/14737140.2019.1561288</pub-id><pub-id pub-id-type="medline">30582397</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Barsouk</surname><given-names>A</given-names> </name><name name-style="western"><surname>Aluru</surname><given-names>JS</given-names> </name><name name-style="western"><surname>Rawla</surname><given-names>P</given-names> </name><name name-style="western"><surname>Saginala</surname><given-names>K</given-names> </name><name name-style="western"><surname>Barsouk</surname><given-names>A</given-names> </name></person-group><article-title>Epidemiology, risk factors, and prevention of head and neck squamous cell carcinoma</article-title><source>Med Sci (Basel)</source><year>2023</year><month>06</month><day>13</day><volume>11</volume><issue>2</issue><fpage>42</fpage><pub-id pub-id-type="doi">10.3390/medsci11020042</pub-id><pub-id pub-id-type="medline">37367741</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Brocklehurst</surname><given-names>P</given-names> </name><name name-style="western"><surname>Kujan</surname><given-names>O</given-names> </name><name name-style="western"><surname>O&#x2019;Malley</surname><given-names>LA</given-names> </name><name name-style="western"><surname>Ogden</surname><given-names>G</given-names> </name><name name-style="western"><surname>Shepherd</surname><given-names>S</given-names> </name><name name-style="western"><surname>Glenny</surname><given-names>AM</given-names> </name></person-group><article-title>Screening programmes for the early detection and prevention of oral cancer</article-title><source>Cochrane Database Syst Rev</source><year>2013</year><month>11</month><day>19</day><volume>2013</volume><issue>11</issue><fpage>CD004150</fpage><pub-id pub-id-type="doi">10.1002/14651858.CD004150.pub4</pub-id><pub-id pub-id-type="medline">24254989</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sankaranarayanan</surname><given-names>R</given-names> </name><name name-style="western"><surname>Ramadas</surname><given-names>K</given-names> </name><name name-style="western"><surname>Thomas</surname><given-names>G</given-names> </name><etal/></person-group><article-title>Effect of screening on oral cancer mortality in Kerala, India: a cluster-randomised controlled trial</article-title><source>Lancet</source><year>2005</year><volume>365</volume><issue>9475</issue><fpage>1927</fpage><lpage>1933</lpage><pub-id pub-id-type="doi">10.1016/S0140-6736(05)66658-5</pub-id><pub-id pub-id-type="medline">15936419</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sankaranarayanan</surname><given-names>R</given-names> </name><name name-style="western"><surname>Ramadas</surname><given-names>K</given-names> </name><name name-style="western"><surname>Thara</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Long term effect of visual screening on oral cancer incidence and mortality in a randomized trial in Kerala, India</article-title><source>Oral Oncol</source><year>2013</year><month>04</month><volume>49</volume><issue>4</issue><fpage>314</fpage><lpage>321</lpage><pub-id pub-id-type="doi">10.1016/j.oraloncology.2012.11.004</pub-id><pub-id pub-id-type="medline">23265945</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Halicek</surname><given-names>M</given-names> </name><name name-style="western"><surname>Fabelo</surname><given-names>H</given-names> </name><name name-style="western"><surname>Ortega</surname><given-names>S</given-names> </name><name name-style="western"><surname>Callico</surname><given-names>GM</given-names> </name><name name-style="western"><surname>Fei</surname><given-names>B</given-names> </name></person-group><article-title>In-vivo and ex-vivo tissue analysis through hyperspectral imaging techniques: revealing the invisible features of cancer</article-title><source>Cancers (Basel)</source><year>2019</year><month>05</month><day>30</day><volume>11</volume><issue>6</issue><fpage>31151223</fpage><pub-id pub-id-type="doi">10.3390/cancers11060756</pub-id><pub-id pub-id-type="medline">31151223</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ishida</surname><given-names>T</given-names> </name><name name-style="western"><surname>Kurihara</surname><given-names>J</given-names> </name><name name-style="western"><surname>Viray</surname><given-names>FA</given-names> </name><etal/></person-group><article-title>A novel approach for vegetation classification using UAV-based hyperspectral imaging</article-title><source>Comput Electron Agric</source><year>2018</year><month>01</month><volume>144</volume><fpage>80</fpage><lpage>85</lpage><pub-id pub-id-type="doi">10.1016/j.compag.2017.11.027</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Miljkovi&#x0107;</surname><given-names>V</given-names> </name><name name-style="western"><surname>Gajski</surname><given-names>D</given-names> </name></person-group><article-title>Adaptation of industrial hyperspectral line scanner for archaeological applications</article-title><source>Int Arch Photogramm Remote Sens Spatial Inf Sci</source><volume>XLI-B5</volume><fpage>343</fpage><lpage>345</lpage><pub-id pub-id-type="doi">10.5194/isprs-archives-XLI-B5-343-2016</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Pavurala</surname><given-names>N</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>X</given-names> </name><name name-style="western"><surname>Krishnaiah</surname><given-names>YSR</given-names> </name></person-group><article-title>Hyperspectral imaging using near infrared spectroscopy to monitor coat thickness uniformity in the manufacture of a transdermal drug delivery system</article-title><source>Int J Pharm</source><year>2017</year><month>05</month><day>15</day><volume>523</volume><issue>1</issue><fpage>281</fpage><lpage>290</lpage><pub-id pub-id-type="doi">10.1016/j.ijpharm.2017.03.022</pub-id><pub-id pub-id-type="medline">28330736</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fern&#x00E1;ndez de la Ossa</surname><given-names>M&#x00C1;</given-names> </name><name name-style="western"><surname>Amigo</surname><given-names>JM</given-names> </name><name name-style="western"><surname>Garc&#x00ED;a-Ruiz</surname><given-names>C</given-names> </name></person-group><article-title>Detection of residues from explosive manipulation by near infrared hyperspectral imaging: a promising forensic tool</article-title><source>Forensic Sci Int</source><year>2014</year><month>09</month><volume>242</volume><fpage>228</fpage><lpage>235</lpage><pub-id pub-id-type="doi">10.1016/j.forsciint.2014.06.023</pub-id><pub-id pub-id-type="medline">25086347</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Sakarya</surname><given-names>U</given-names> </name><name name-style="western"><surname>Teke</surname><given-names>M</given-names> </name><name name-style="western"><surname>Demirkesen</surname><given-names>C</given-names> </name><etal/></person-group><article-title>A short survey of hyperspectral remote sensing and hyperspectral remote sensing research at t&#x00FC;b&#x0131;tak uzay</article-title><conf-name>2015 7th International Conference on Recent Advances in Space Technologies (RAST)</conf-name><conf-date>Jun 16-19, 2015</conf-date><conf-loc>Istanbul, Turkey</conf-loc><fpage>187</fpage><lpage>192</lpage><pub-id pub-id-type="doi">10.1109/RAST.2015.7208339</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Qin</surname><given-names>J</given-names> </name><name name-style="western"><surname>Kim</surname><given-names>M</given-names> </name><name name-style="western"><surname>Chao</surname><given-names>K</given-names> </name><name name-style="western"><surname>Chan</surname><given-names>D</given-names> </name><name name-style="western"><surname>Delwiche</surname><given-names>S</given-names> </name><name name-style="western"><surname>Cho</surname><given-names>BK</given-names> </name></person-group><article-title>Line-scan hyperspectral imaging techniques for food safety and quality applications</article-title><source>Appl Sci (Basel)</source><year>2017</year><volume>7</volume><issue>2</issue><fpage>125</fpage><pub-id pub-id-type="doi">10.3390/app7020125</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lu</surname><given-names>G</given-names> </name><name name-style="western"><surname>Fei</surname><given-names>B</given-names> </name></person-group><article-title>Medical hyperspectral imaging: a review</article-title><source>J Biomed Opt</source><year>2014</year><month>01</month><volume>19</volume><issue>1</issue><fpage>10901</fpage><pub-id pub-id-type="doi">10.1117/1.JBO.19.1.010901</pub-id><pub-id pub-id-type="medline">24441941</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>R&#x00F6;mer</surname><given-names>P</given-names> </name><name name-style="western"><surname>Blatt</surname><given-names>S</given-names> </name><name name-style="western"><surname>Siegberg</surname><given-names>F</given-names> </name><etal/></person-group><article-title>Intraoral perfusion assessment using endoscopic hyperspectral imaging (EHSI)- first description of a novel approach</article-title><source>Clin Oral Investig</source><year>2025</year><month>02</month><day>5</day><volume>29</volume><issue>2</issue><fpage>115</fpage><pub-id pub-id-type="doi">10.1007/s00784-025-06197-5</pub-id><pub-id pub-id-type="medline">39907805</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lu</surname><given-names>G</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>D</given-names> </name><name name-style="western"><surname>Qin</surname><given-names>X</given-names> </name><etal/></person-group><article-title>Histopathology feature mining and association with hyperspectral imaging for the detection of squamous neoplasia</article-title><source>Sci Rep</source><year>2019</year><month>11</month><day>28</day><volume>9</volume><issue>1</issue><fpage>17863</fpage><pub-id pub-id-type="doi">10.1038/s41598-019-54139-5</pub-id><pub-id pub-id-type="medline">31780698</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ortega</surname><given-names>S</given-names> </name><name name-style="western"><surname>Halicek</surname><given-names>M</given-names> </name><name name-style="western"><surname>Fabelo</surname><given-names>H</given-names> </name><name name-style="western"><surname>Callico</surname><given-names>GM</given-names> </name><name name-style="western"><surname>Fei</surname><given-names>B</given-names> </name></person-group><article-title>Hyperspectral and multispectral imaging in digital and computational pathology: a systematic review [Invited]</article-title><source>Biomed Opt Express</source><year>2020</year><month>06</month><day>1</day><volume>11</volume><issue>6</issue><fpage>3195</fpage><lpage>3233</lpage><pub-id pub-id-type="doi">10.1364/BOE.386338</pub-id><pub-id pub-id-type="medline">32637250</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bhargava</surname><given-names>R</given-names> </name><name name-style="western"><surname>Falahkheirkhah</surname><given-names>K</given-names> </name></person-group><article-title>Enhancing hyperspectral imaging</article-title><source>Nat Mach Intell</source><year>2021</year><volume>3</volume><issue>4</issue><fpage>279</fpage><lpage>280</lpage><pub-id pub-id-type="doi">10.1038/s42256-021-00336-9</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Cui</surname><given-names>R</given-names> </name><name name-style="western"><surname>Yu</surname><given-names>H</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>T</given-names> </name><etal/></person-group><article-title>Deep learning in medical hyperspectral images: a review</article-title><source>Sensors (Basel)</source><year>2022</year><month>12</month><day>13</day><volume>22</volume><issue>24</issue><fpage>9790</fpage><pub-id pub-id-type="doi">10.3390/s22249790</pub-id><pub-id pub-id-type="medline">36560157</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Madooei</surname><given-names>A</given-names> </name><name name-style="western"><surname>Abdlaty</surname><given-names>RM</given-names> </name><name name-style="western"><surname>Doerwald-Munoz</surname><given-names>L</given-names> </name><etal/></person-group><article-title>Hyperspectral image processing for detection and grading of skin erythema</article-title><year>2017</year><month>02</month><day>24</day><conf-name>SPIE Medical Imaging</conf-name><conf-date>Feb 11-16, 2017</conf-date><conf-loc>Orlando, Florida, United States</conf-loc><fpage>1013322</fpage><pub-id pub-id-type="doi">10.1117/12.2254132</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Krizhevsky</surname><given-names>A</given-names> </name><name name-style="western"><surname>Sutskever</surname><given-names>I</given-names> </name><name name-style="western"><surname>Hinton</surname><given-names>GE</given-names> </name></person-group><article-title>ImageNet classification with deep convolutional neural networks</article-title><source>Commun ACM</source><year>2017</year><month>05</month><day>24</day><volume>60</volume><issue>6</issue><fpage>84</fpage><lpage>90</lpage><pub-id pub-id-type="doi">10.1145/3065386</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>W</given-names> </name></person-group><article-title>Cell classification using convolutional neural networks in medical hyperspectral imagery</article-title><conf-name>2017 2nd International Conference on Image, Vision and Computing (ICIVC)</conf-name><conf-date>Jun 2-4, 2017</conf-date><conf-loc>Chengdu, China</conf-loc><fpage>501</fpage><lpage>504</lpage><pub-id pub-id-type="doi">10.1109/ICIVC.2017.7984606</pub-id></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Jeyaraj</surname><given-names>PR</given-names> </name><name name-style="western"><surname>Samuel Nadar</surname><given-names>ER</given-names> </name></person-group><article-title>Computer-assisted medical image classification for early diagnosis of oral cancer employing deep learning algorithm</article-title><source>J Cancer Res Clin Oncol</source><year>2019</year><month>04</month><volume>145</volume><issue>4</issue><fpage>829</fpage><lpage>837</lpage><pub-id pub-id-type="doi">10.1007/s00432-018-02834-7</pub-id><pub-id pub-id-type="medline">30603908</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Gangadharan</surname><given-names>SMP</given-names> </name><name name-style="western"><surname>Dharani</surname><given-names>M</given-names> </name><name name-style="western"><surname>Thapliyal</surname><given-names>N</given-names> </name><name name-style="western"><surname>Yamsani</surname><given-names>N</given-names> </name><name name-style="western"><surname>Singh</surname><given-names>J</given-names> </name><name name-style="western"><surname>Singh</surname><given-names>P</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Singh</surname><given-names>P</given-names> </name></person-group><article-title>Comparative analysis of deep learning-based brain tumor prediction models using MRI scan</article-title><conf-name>2023 3rd International Conference on Innovative Sustainable Computational Technologies (CISCT)</conf-name><conf-date>Sep 8-9, 2023</conf-date><conf-loc>Dehradun, India</conf-loc><fpage>1</fpage><lpage>6</lpage><pub-id pub-id-type="doi">10.1109/CISCT57197.2023.10351227</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Diwakar</surname><given-names>M</given-names> </name><name name-style="western"><surname>Singh</surname><given-names>P</given-names> </name><name name-style="western"><surname>Garg</surname><given-names>D</given-names> </name></person-group><article-title>Edge-guided filtering based CT image denoising using fractional order total variation</article-title><source>Biomed Signal Process Control</source><year>2024</year><month>06</month><volume>92</volume><fpage>106072</fpage><pub-id pub-id-type="doi">10.1016/j.bspc.2024.106072</pub-id></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Diwakar</surname><given-names>M</given-names> </name><name name-style="western"><surname>Kumar</surname><given-names>P</given-names> </name><name name-style="western"><surname>Singh</surname><given-names>P</given-names> </name><name name-style="western"><surname>Tripathi</surname><given-names>A</given-names> </name><name name-style="western"><surname>Singh</surname><given-names>L</given-names> </name></person-group><article-title>An efficient reversible data hiding using SVD over a novel weighted iterative anisotropic total variation based denoised medical images</article-title><source>Biomed Signal Process Control</source><year>2023</year><month>04</month><volume>82</volume><fpage>104563</fpage><pub-id pub-id-type="doi">10.1016/j.bspc.2022.104563</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Diwakar</surname><given-names>M</given-names> </name><name name-style="western"><surname>Pandey</surname><given-names>NK</given-names> </name><name name-style="western"><surname>Singh</surname><given-names>R</given-names> </name><etal/></person-group><article-title>Low-dose COVID-19 CT Image Denoising Using CNN and its Method Noise Thresholding</article-title><source>Curr Med Imaging</source><year>2023</year><volume>19</volume><issue>2</issue><fpage>182</fpage><lpage>193</lpage><pub-id pub-id-type="doi">10.2174/1573405618666220404162241</pub-id><pub-id pub-id-type="medline">35379137</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Agrawal</surname><given-names>T</given-names> </name><name name-style="western"><surname>Choudhary</surname><given-names>P</given-names> </name><name name-style="western"><surname>Shankar</surname><given-names>A</given-names> </name><name name-style="western"><surname>Singh</surname><given-names>P</given-names> </name><name name-style="western"><surname>Diwakar</surname><given-names>M</given-names> </name></person-group><article-title>MultiFeNet: multi-scale feature scaling in deep neural network for the brain tumour classification in MRI images</article-title><source>Int J Imaging Syst Technol</source><year>2024</year><volume>34</volume><issue>1</issue><fpage>e22956</fpage><pub-id pub-id-type="doi">10.1002/ima.22956</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sicher</surname><given-names>C</given-names> </name><name name-style="western"><surname>Rutkowski</surname><given-names>R</given-names> </name><name name-style="western"><surname>Lutze</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Hyperspectral imaging as a possible tool for visualization of changes in hemoglobin oxygenation in patients with deficient hemodynamics - proof of concept</article-title><source>Biomed Tech (Berl)</source><year>2018</year><month>10</month><day>25</day><volume>63</volume><issue>5</issue><fpage>609</fpage><lpage>616</lpage><pub-id pub-id-type="doi">10.1515/bmt-2017-0084</pub-id><pub-id pub-id-type="medline">30210054</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yudovsky</surname><given-names>D</given-names> </name><name name-style="western"><surname>Nouvong</surname><given-names>A</given-names> </name><name name-style="western"><surname>Pilon</surname><given-names>L</given-names> </name></person-group><article-title>Hyperspectral imaging in diabetic foot wound care</article-title><source>J Diabetes Sci Technol</source><year>2010</year><month>09</month><day>1</day><volume>4</volume><issue>5</issue><fpage>1099</fpage><lpage>1113</lpage><pub-id pub-id-type="doi">10.1177/193229681000400508</pub-id><pub-id pub-id-type="medline">20920429</pub-id></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Ponciano</surname><given-names>C</given-names> </name><name name-style="western"><surname>Schaffert</surname><given-names>M</given-names> </name><name name-style="western"><surname>Ponciano</surname><given-names>JJ</given-names> </name></person-group><article-title>Deep learning datasets challenges for semantic segmentation-a survey</article-title><conf-name>P337 - INFORMATIK 2023 - Designing Futures: Shaping the Future</conf-name><conf-date>Sep 26-29, 2023</conf-date><conf-loc>Berlin, Germany</conf-loc><fpage>57</fpage><lpage>70</lpage><pub-id pub-id-type="doi">10.18420/inf2023_04</pub-id></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Gupta</surname><given-names>P</given-names> </name><name name-style="western"><surname>Bagchi</surname><given-names>A</given-names> </name></person-group><source>Essentials of Python for Artificial Intelligence and Machine Learning</source><year>2024</year><pub-id pub-id-type="other">3031437241</pub-id></nlm-citation></ref><ref id="ref38"><label>38</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Lu</surname><given-names>H</given-names> </name><name name-style="western"><surname>Cai</surname><given-names>J</given-names> </name></person-group><article-title>Artificial intelligence and robotics</article-title><source>8th International Symposium, ISAIR 2023</source><year>2024</year><volume>15</volume><publisher-name>Springer</publisher-name><fpage>535</fpage><comment><ext-link ext-link-type="uri" xlink:href="https://link.springer.com/10.1007/978-981-99-9109-9">https://link.springer.com/10.1007/978-981-99-9109-9</ext-link></comment><pub-id pub-id-type="doi">10.1007/978-981-99-9109-9</pub-id></nlm-citation></ref><ref id="ref39"><label>39</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Chen</surname><given-names>LC</given-names> </name></person-group><article-title>Rethinking atrous convolution for semantic image segmentation</article-title><source>arXiv</source><comment>Preprint posted online on  Sep 5, 2017</comment><pub-id pub-id-type="doi">10.48550/arXiv.1706.05587</pub-id></nlm-citation></ref><ref id="ref40"><label>40</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Chen</surname><given-names>LC</given-names> </name><name name-style="western"><surname>Zhu</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Papandreou</surname><given-names>G</given-names> </name><name name-style="western"><surname>Schroff</surname><given-names>F</given-names> </name><collab>editors</collab></person-group><article-title>Encoder-decoder with atrous separable convolution for semantic image segmentation</article-title><source>Proceedings of the European Conference on Computer Vision (ECCV</source><year>2018</year><volume>1121</volume><publisher-name>Springer</publisher-name><fpage>833</fpage><lpage>851</lpage><pub-id pub-id-type="doi">10.1007/978-3-030-01234-2_49</pub-id></nlm-citation></ref><ref id="ref41"><label>41</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Long</surname><given-names>J</given-names> </name><name name-style="western"><surname>Shelhamer</surname><given-names>E</given-names> </name><name name-style="western"><surname>Darrell</surname><given-names>T</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Long</surname><given-names>J</given-names> </name><name name-style="western"><surname>Shelhamer</surname><given-names>E</given-names> </name><name name-style="western"><surname>Darrell</surname><given-names>T</given-names> </name></person-group><article-title>Fully convolutional networks for semantic segmentation</article-title><conf-name>Proceedings of the IEEE conference on computer vision and pattern recognition</conf-name><conf-date>Jun 7-12, 2015</conf-date><conf-loc>Boston, MA, USA</conf-loc><fpage>3431</fpage><lpage>3440</lpage><pub-id pub-id-type="doi">10.1109/CVPR.2015.7298965</pub-id></nlm-citation></ref><ref id="ref42"><label>42</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Zhao</surname><given-names>H</given-names> </name><name name-style="western"><surname>Shi</surname><given-names>J</given-names> </name><name name-style="western"><surname>Qi</surname><given-names>X</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>X</given-names> </name><name name-style="western"><surname>Jia</surname><given-names>J</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Jia</surname><given-names>J</given-names> </name></person-group><article-title>Pyramid scene parsing network</article-title><conf-name>2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</conf-name><conf-date>Jul 21-26, 2017</conf-date><conf-loc>Honolulu, HI</conf-loc><fpage>6230</fpage><lpage>6239</lpage><pub-id pub-id-type="doi">10.1109/CVPR.2017.660</pub-id></nlm-citation></ref><ref id="ref43"><label>43</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>He</surname><given-names>K</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>X</given-names> </name><name name-style="western"><surname>Ren</surname><given-names>S</given-names> </name><name name-style="western"><surname>Sun</surname><given-names>J</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Sun</surname><given-names>J</given-names> </name></person-group><article-title>Deep residual learning for image recognition</article-title><conf-name>2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</conf-name><conf-date>Jun 27-30, 2016</conf-date><conf-loc>Las Vegas, NV, USA</conf-loc><fpage>770</fpage><lpage>778</lpage><pub-id pub-id-type="doi">10.1109/CVPR.2016.90</pub-id></nlm-citation></ref><ref id="ref44"><label>44</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Simonyan</surname><given-names>K</given-names> </name><name name-style="western"><surname>Zisserman</surname><given-names>A</given-names> </name></person-group><article-title>Very deep convolutional networks for large-scale image recognition</article-title><source>arXiv</source><comment>Preprint posted online on  Apr 10, 2015</comment><pub-id pub-id-type="doi">10.48550/arXiv.1409.1556</pub-id></nlm-citation></ref><ref id="ref45"><label>45</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Ronneberger</surname><given-names>O</given-names> </name><name name-style="western"><surname>Fischer</surname><given-names>P</given-names> </name><name name-style="western"><surname>Brox</surname><given-names>T</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Navab</surname><given-names>N</given-names> </name><name name-style="western"><surname>Hornegger</surname><given-names>J</given-names> </name><name name-style="western"><surname>Wells</surname><given-names>W</given-names> </name><name name-style="western"><surname>Frangi</surname><given-names>A</given-names> </name></person-group><article-title>U-net: convolutional networks for biomedical image segmentation. medical image computing and computer-assisted intervention&#x2013;MICCAI 2015</article-title><source>Medical Image Computing and Computer-Assisted Intervention &#x2013; MICCAI 2015</source><year>2015</year><volume>9351</volume><publisher-name>Springer</publisher-name><fpage>234</fpage><lpage>241</lpage><pub-id pub-id-type="doi">10.1007/978-3-319-24574-4_28</pub-id></nlm-citation></ref><ref id="ref46"><label>46</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Tan</surname><given-names>M</given-names> </name><name name-style="western"><surname>Le</surname><given-names>Q</given-names> </name></person-group><article-title>Efficientnet: rethinking model scaling for convolutional neural networks</article-title><access-date>2025-09-03</access-date><conf-name>Proceedings of the 36th International Conference on Machine Learning</conf-name><conf-date>Jun 9-15, 2019</conf-date><conf-loc>Long Beach, CA</conf-loc><fpage>6105</fpage><lpage>6114</lpage><comment><ext-link ext-link-type="uri" xlink:href="https://proceedings.mlr.press/v97/tan19a.html">https://proceedings.mlr.press/v97/tan19a.html</ext-link></comment></nlm-citation></ref><ref id="ref47"><label>47</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Thiem</surname><given-names>DGE</given-names> </name><name name-style="western"><surname>R&#x00F6;mer</surname><given-names>P</given-names> </name><name name-style="western"><surname>Gielisch</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Hyperspectral imaging and artificial intelligence to detect oral malignancy - part 1 - automated tissue classification of oral muscle, fat and mucosa using a light-weight 6-layer deep neural network</article-title><source>Head Face Med</source><year>2021</year><month>09</month><day>3</day><volume>17</volume><issue>1</issue><fpage>38</fpage><pub-id pub-id-type="doi">10.1186/s13005-021-00292-0</pub-id><pub-id pub-id-type="medline">34479595</pub-id></nlm-citation></ref><ref id="ref48"><label>48</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chen</surname><given-names>LC</given-names> </name><name name-style="western"><surname>Papandreou</surname><given-names>G</given-names> </name><name name-style="western"><surname>Kokkinos</surname><given-names>I</given-names> </name><name name-style="western"><surname>Murphy</surname><given-names>K</given-names> </name><name name-style="western"><surname>Yuille</surname><given-names>AL</given-names> </name></person-group><article-title>DeepLab: semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRF</article-title><source>IEEE Trans Pattern Anal Mach Intell</source><year>2018</year><month>04</month><volume>40</volume><issue>4</issue><fpage>834</fpage><lpage>848</lpage><pub-id pub-id-type="doi">10.1109/TPAMI.2017.2699184</pub-id><pub-id pub-id-type="medline">28463186</pub-id></nlm-citation></ref><ref id="ref49"><label>49</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Ma</surname><given-names>L</given-names> </name><name name-style="western"><surname>Srinivas</surname><given-names>A</given-names> </name><name name-style="western"><surname>Krishnamurthy</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Automated polarized hyperspectral imaging (PHSI) for ex-vivo and in-vivo tissue assessment</article-title><year>2023</year><month>01</month><fpage>123910F</fpage><pub-id pub-id-type="doi">10.1117/12.2651011</pub-id><pub-id pub-id-type="medline">38476292</pub-id></nlm-citation></ref><ref id="ref50"><label>50</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Poonkuzhali</surname><given-names>P</given-names> </name><name name-style="western"><surname>Helen Prabha</surname><given-names>K</given-names> </name></person-group><article-title>Deep convolutional neural network based hyperspectral brain tissue classification</article-title><source>J Xray Sci Technol</source><year>2023</year><volume>31</volume><issue>4</issue><fpage>777</fpage><lpage>796</lpage><pub-id pub-id-type="doi">10.3233/XST-230045</pub-id><pub-id pub-id-type="medline">37182861</pub-id></nlm-citation></ref><ref id="ref51"><label>51</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hu</surname><given-names>B</given-names> </name><name name-style="western"><surname>Du</surname><given-names>J</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>Q</given-names> </name></person-group><article-title>Tumor tissue classification based on micro-hyperspectral technology and deep learning</article-title><source>Biomed Opt Express</source><year>2019</year><month>12</month><day>1</day><volume>10</volume><issue>12</issue><fpage>6370</fpage><lpage>6389</lpage><pub-id pub-id-type="doi">10.1364/BOE.10.006370</pub-id><pub-id pub-id-type="medline">31853405</pub-id></nlm-citation></ref><ref id="ref52"><label>52</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ma</surname><given-names>L</given-names> </name><name name-style="western"><surname>Zhou</surname><given-names>X</given-names> </name><name name-style="western"><surname>Little</surname><given-names>JV</given-names> </name><etal/></person-group><article-title>Hyperspectral microscopic imaging for the detection of head and neck squamous cell carcinoma in histologic images</article-title><source>Proc SPIE Int Soc Opt Eng</source><year>2021</year><month>02</month><volume>11603</volume><fpage>35783088</fpage><pub-id pub-id-type="doi">10.1117/12.2581970</pub-id><pub-id pub-id-type="medline">35783088</pub-id></nlm-citation></ref><ref id="ref53"><label>53</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ma</surname><given-names>L</given-names> </name><name name-style="western"><surname>Lu</surname><given-names>G</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>D</given-names> </name><name name-style="western"><surname>Qin</surname><given-names>X</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>ZG</given-names> </name><name name-style="western"><surname>Fei</surname><given-names>B</given-names> </name></person-group><article-title>Adaptive deep learning for head and neck cancer detection using hyperspectral imaging</article-title><source>Vis Comput Ind Biomed Art</source><year>2019</year><volume>2</volume><issue>1</issue><fpage>18</fpage><pub-id pub-id-type="doi">10.1186/s42492-019-0023-8</pub-id><pub-id pub-id-type="medline">32190408</pub-id></nlm-citation></ref><ref id="ref54"><label>54</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhou</surname><given-names>X</given-names> </name><name name-style="western"><surname>Ma</surname><given-names>L</given-names> </name><name name-style="western"><surname>Mubarak</surname><given-names>HK</given-names> </name><etal/></person-group><article-title>Automatic detection of head and neck squamous cell carcinoma on pathologic slides using polarized hyperspectral imaging and deep learning</article-title><source>Proc SPIE Int Soc Opt Eng</source><year>2022</year><volume>12039</volume><fpage>Feb</fpage><lpage>Mar</lpage><pub-id pub-id-type="doi">10.1117/12.2614624</pub-id><pub-id pub-id-type="medline">36798940</pub-id></nlm-citation></ref><ref id="ref55"><label>55</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Eggert</surname><given-names>D</given-names> </name><name name-style="western"><surname>Bengs</surname><given-names>M</given-names> </name><name name-style="western"><surname>Westermann</surname><given-names>S</given-names> </name><etal/></person-group><article-title>In vivo detection of head and neck tumors by hyperspectral imaging combined with deep learning methods</article-title><source>J Biophotonics</source><year>2022</year><month>03</month><volume>15</volume><issue>3</issue><fpage>e202100167</fpage><pub-id pub-id-type="doi">10.1002/jbio.202100167</pub-id><pub-id pub-id-type="medline">34889065</pub-id></nlm-citation></ref><ref id="ref56"><label>56</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Trajanovski</surname><given-names>S</given-names> </name><name name-style="western"><surname>Shan</surname><given-names>C</given-names> </name><name name-style="western"><surname>Weijtmans</surname><given-names>PJC</given-names> </name><name name-style="western"><surname>de Koning</surname><given-names>SGB</given-names> </name><name name-style="western"><surname>Ruers</surname><given-names>TJM</given-names> </name></person-group><article-title>Tongue tumor detection in hyperspectral images using deep learning semantic segmentation</article-title><source>IEEE Trans Biomed Eng</source><year>2021</year><month>04</month><volume>68</volume><issue>4</issue><fpage>1330</fpage><lpage>1340</lpage><pub-id pub-id-type="doi">10.1109/TBME.2020.3026683</pub-id><pub-id pub-id-type="medline">32976092</pub-id></nlm-citation></ref></ref-list></back></article>