<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Med Inform</journal-id><journal-id journal-id-type="publisher-id">medinform</journal-id><journal-id journal-id-type="index">7</journal-id><journal-title>JMIR Medical Informatics</journal-title><abbrev-journal-title>JMIR Med Inform</abbrev-journal-title><issn pub-type="epub">2291-9694</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v14i1e83785</article-id><article-id pub-id-type="doi">10.2196/83785</article-id><article-categories><subj-group subj-group-type="heading"><subject>Original Paper</subject></subj-group></article-categories><title-group><article-title>An Entity-Based Visual Analytics System Enhancing Medical Expertise Acquisition: Development and Verification Study</article-title></title-group><contrib-group><contrib contrib-type="author"><name name-style="western"><surname>Pang</surname><given-names>Xiao</given-names></name><degrees>MS</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Liu</surname><given-names>Chang</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Huang</surname><given-names>Yan</given-names></name><degrees>MS</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Liu</surname><given-names>MingYou</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author" corresp="yes" equal-contrib="yes"><name name-style="western"><surname>Liu</surname><given-names>Jiyuan</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib></contrib-group><aff id="aff1"><institution>Department of Information Management, State Key Laboratory of Oral Diseases &#x0026; National Center for Stomatology &#x0026; National Clinical Research Center for Oral Diseases, West China Hospital of Stomatology, Sichuan University</institution><addr-line>No. 14, 3rd Section of Ren Min Nan Rd.</addr-line><addr-line>Chengdu</addr-line><country>China</country></aff><aff id="aff2"><institution>Department of Information and Network Security, Peking Union Medical College Hospital, Chinese Academy of Medical Sciences</institution><addr-line>Beijing</addr-line><country>China</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Benis</surname><given-names>Arriel</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Weng</surname><given-names>Di</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Megaritis</surname><given-names>Dimitrios</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Wen</surname><given-names>Xiaolin</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Jiyuan Liu, MD, Department of Information Management, State Key Laboratory of Oral Diseases &#x0026; National Center for Stomatology &#x0026; National Clinical Research Center for Oral Diseases, West China Hospital of Stomatology, Sichuan University, No. 14, 3rd Section of Ren Min Nan Rd., Chengdu, 610041, China, 86 13880737189; <email>wchsljy@scu.edu.cn</email></corresp><fn fn-type="equal" id="equal-contrib1"><label>*</label><p>these authors contributed equally</p></fn></author-notes><pub-date pub-type="collection"><year>2026</year></pub-date><pub-date pub-type="epub"><day>15</day><month>5</month><year>2026</year></pub-date><volume>14</volume><elocation-id>e83785</elocation-id><history><date date-type="received"><day>08</day><month>09</month><year>2025</year></date><date date-type="rev-recd"><day>06</day><month>03</month><year>2026</year></date><date date-type="accepted"><day>06</day><month>03</month><year>2026</year></date></history><copyright-statement>&#x00A9; Xiao Pang, Chang Liu, Yan Huang, MingYou Liu, Jiyuan Liu. Originally published in JMIR Medical Informatics (<ext-link ext-link-type="uri" xlink:href="https://medinform.jmir.org">https://medinform.jmir.org</ext-link>), 15.5.2026. </copyright-statement><copyright-year>2026</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://medinform.jmir.org/">https://medinform.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://medinform.jmir.org/2026/1/e83785"/><abstract><sec><title>Background</title><p>Acquiring medical expertise from the vast body of medical text is a critical component of medical education. However, the majority of medical knowledge resides in unstructured texts. Data heterogeneity across institutions and strict privacy regulations hinder the use of general-purpose analysis tools. This creates a substantial barrier to the efficient acquisition of expertise for learners.</p></sec><sec><title>Objective</title><p>This study aimed to design, develop, and evaluate MExplore, an interactive visual analytics system to facilitate the acquisition of medical expertise from unstructured clinical texts.</p></sec><sec sec-type="methods"><title>Methods</title><p>We propose a localized, cost-effective workflow for the automatic extraction of medical entities. Building on this workflow, MExplore provides a multilevel visual framework featuring coordinated views for progressive, entity-centered exploration. The system was evaluated through case studies conducted with domain experts, a user study, and semistructured interviews.</p></sec><sec sec-type="results"><title>Results</title><p>The evaluation demonstrated that MExplore significantly enhances the medical expertise acquisition process. Our findings confirm that MExplore provides an effective and interactive approach for structuring complex knowledge, facilitating the construction of illness scripts, and strengthening knowledge retention.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>MExplore provides an intuitive and powerful approach for acquiring medical expertise. The results suggest that it effectively supports medical learners in conducting in-depth data exploration and developing robust clinical reasoning skills.</p></sec></abstract><kwd-group><kwd>medical text</kwd><kwd>medical expertise</kwd><kwd>illness script</kwd><kwd>visualization</kwd><kwd>visual analytics</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>The rapid expansion of medical education has intensified the demand for efficient expertise acquisition [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref2">2</xref>]. Medical expertise is characterized by a highly organized and differentiated knowledge base, technical skills, and perceptual capabilities acquired through extensive domain-related practice [<xref ref-type="bibr" rid="ref3">3</xref>]. The acquisition of medical expertise is a complex cognitive process that involves constructing abstract knowledge networks and iteratively refining them into narrative structures known as &#x201C;illness scripts&#x201D; [<xref ref-type="bibr" rid="ref4">4</xref>]&#x2014;cognitive frameworks that link clinical features to diagnoses and enable clinicians to perform complex reasoning tasks efficiently [<xref ref-type="bibr" rid="ref5">5</xref>]. Traditional learning relies heavily on textbooks and often overlooks real-world medical documents (MDs) such as electronic medical records. Although these documents provide rich, detailed cases for refining illness scripts, their unstructured nature and volume make it difficult for novices to extract and synthesize information.</p><p>While large language models (LLMs) have emerged as tools for information synthesis, their utility is limited by practical barriers, such as privacy regulations and computational costs [<xref ref-type="bibr" rid="ref6">6</xref>]. Furthermore, their direct application in medical education is fraught with risks. Research indicates that LLMs are prone to &#x201C;deceptive expertise,&#x201D; generating plausible yet factually incorrect information while failing to recognize their own limitations [<xref ref-type="bibr" rid="ref7">7</xref>,<xref ref-type="bibr" rid="ref8">8</xref>]. Overreliance may lead to cognitive offloading, impairing learners&#x2019; ability to critically evaluate artificial intelligence&#x2013;generated outputs [<xref ref-type="bibr" rid="ref9">9</xref>-<xref ref-type="bibr" rid="ref12">12</xref>]&#x2014;a particularly concerning issue in medical education, where independent clinical reasoning is essential. Consequently, there is a pressing need for frameworks that support exploratory learning and active engagement with data, rather than systems that encourage the passive acceptance of output [<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref13">13</xref>].</p><p>Visual analytics, which integrates human expertise with machine intelligence in a &#x201C;human-in-the-loop&#x201D; paradigm, offers a viable approach for such exploratory frameworks. By visualizing complex data, these systems can facilitate the construction of mental models. However, the application of visual analytics to medical expertise acquisition remains underdeveloped. Much of the existing work in this domain has focused on analyzing individual patient electronic medical records to support clinical decision-making for specific cases [<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref15">15</xref>], which does not facilitate the higher-level organization of generalizable medical knowledge required for education. Other studies have visualized health care data but often neglect the underlying semantics of the content [<xref ref-type="bibr" rid="ref16">16</xref>,<xref ref-type="bibr" rid="ref17">17</xref>], limiting the user&#x2019;s ability to understand deep clinical relationships. Therefore, a gap exists for a system that can aggregate unstructured text and present it in a semantically meaningful, multilevel structure suitable for learning.</p><p>Named entity recognition (NER) serves as a key technique for transforming unstructured text into structured representations. It involves identifying and classifying domain-specific concepts, such as symptoms, diseases, and treatments, within a text [<xref ref-type="bibr" rid="ref18">18</xref>,<xref ref-type="bibr" rid="ref19">19</xref>]. While recent advancements in deep learning and transformer-based architectures such as BERT (Bidirectional Encoder Representations From Transformers) have significantly improved the accuracy of extracting these medical entities (MEs) [<xref ref-type="bibr" rid="ref20">20</xref>-<xref ref-type="bibr" rid="ref23">23</xref>], the potential of using extracted entities as &#x201C;anchors&#x201D; to construct visual knowledge networks remains largely unexplored. Integrating NER with visual analytics offers a novel pathway to transform raw clinical text into structured learning pathways, thereby accelerating the construction of illness scripts.</p><p>We present MExplore, a system that leverages NER to extract MEs from unstructured text and organizes them into a coherent visual structure based on a multilevel metaphor. This design allows learners to explore medical knowledge at varying granularities&#x2014;including MDs, medical paragraphs (MPs), medical entity sets (MESs), and individual MEs. This approach actively engages users in reasoning, reducing the passive learning common in chat-based interfaces [<xref ref-type="bibr" rid="ref24">24</xref>,<xref ref-type="bibr" rid="ref25">25</xref>], and reinforcing the cognitive links necessary for robust illness scripts. We validate the system through a multifaceted evaluation involving domain expert case studies, a controlled user study with medical students, and expert interviews to assess clinical validity and user acceptance [<xref ref-type="bibr" rid="ref26">26</xref>].</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Study Design</title><sec id="s2-1-1"><title>Design Methodology and Workflow</title><p>We used a user-centered design process [<xref ref-type="bibr" rid="ref27">27</xref>], as illustrated in <xref ref-type="fig" rid="figure1">Figure 1</xref>. Initially, we collaborated with domain experts to analyze requirements and define specific visual analytics tasks. Guided by these tasks, we established the hierarchy, structural organization, and computational metrics for the textual data. We subsequently designed and developed a multilevel visual analytics framework tailored to these data specifications. Throughout the development lifecycle, we engaged in close collaboration with experts and users, conducting multiple rounds of testing and validation. The feedback and recommendations from these sessions drove iterative redesigns, ensuring continuous system optimization.</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>The MExplore workflow encompasses visual design, data processing, storage, and query handling, culminating in multilevel visual analysis. The process incorporates iterative redesigns driven by user and expert feedback. MD: medical document; MP: medical paragraph.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e83785_fig01.png"/></fig></sec><sec id="s2-1-2"><title>Requirements Analysis</title><sec id="s2-1-2-1"><title>Requirement Elicitation</title><p>We collaborated closely with 10 domain experts (E1-E10). E1 and E2 are PhD candidates specializing in medical studies, with 4 and 5 years of focused research experience; E3 is a professor at a medical college with 8 years of teaching experience; E4 is a medical researcher with a decade of research expertise; E5 is a dentist with 10 years of professional practice; E6-E8 are experienced physicians with 5&#x2010;10 years of clinical practice; and E9 and E10 are visualization researchers, with 5&#x2010;10 years of experience in visual analytics.</p><p>Through iterative discussions, the experts (specifically E3-E8) emphasized that acquiring expertise involves refining knowledge into &#x201C;illness scripts&#x201D; [<xref ref-type="bibr" rid="ref4">4</xref>]. These scripts enable clinicians to integrate disease-relevant information and enhance both recall and application [<xref ref-type="bibr" rid="ref28">28</xref>,<xref ref-type="bibr" rid="ref29">29</xref>]. To support this, we identified 4 key requirements.</p></sec><sec id="s2-1-2-2"><title>R1: Extraction of Core Knowledge Units From Complex Texts</title><p>Unstructured medical texts create a high cognitive load for novice learners due to their density and complexity [<xref ref-type="bibr" rid="ref30">30</xref>]. Novices often face a &#x201C;reading bottleneck&#x201D; [<xref ref-type="bibr" rid="ref31">31</xref>] where resources are consumed by decoding information rather than deep comprehension. Therefore, it is essential to extract core knowledge units and filter out redundant information to mitigate extraneous cognitive load [<xref ref-type="bibr" rid="ref32">32</xref>,<xref ref-type="bibr" rid="ref33">33</xref>].</p></sec><sec id="s2-1-2-3"><title>R2: Organization of Text With Varying Knowledge Density to Support Gradual Exploration</title><p>Presenting excessive details simultaneously causes cognitive overload [<xref ref-type="bibr" rid="ref32">32</xref>,<xref ref-type="bibr" rid="ref34">34</xref>]. Experts suggest that information should be presented at appropriate levels of abstraction [<xref ref-type="bibr" rid="ref35">35</xref>]. Medical texts must be restructured with varying knowledge densities to support an incremental learning process, moving from high-level overviews to granular details [<xref ref-type="bibr" rid="ref36">36</xref>].</p></sec><sec id="s2-1-2-4"><title>R3: Revealing the Interconnections Between Knowledge Units</title><p>Rote memorization is insufficient for long-term retention [<xref ref-type="bibr" rid="ref37">37</xref>]. Learners must actively integrate new information with existing schemas [<xref ref-type="bibr" rid="ref38">38</xref>]. Consequently, there is a need to visually reveal the logical interconnections and inclusion relationships between knowledge units to support structural understanding.</p></sec><sec id="s2-1-2-5"><title>R4: Focused Analysis of the Key Knowledge Units</title><p>Effective knowledge acquisition relies on identifying and analyzing key knowledge units that serve as cognitive &#x201C;anchors&#x201D; [<xref ref-type="bibr" rid="ref39">39</xref>,<xref ref-type="bibr" rid="ref40">40</xref>]. Focused analysis of these units not only deepens understanding but also facilitates the establishment of broader connections within the medical field [<xref ref-type="bibr" rid="ref40">40</xref>], thereby reinforcing long-term retention.</p></sec></sec></sec><sec id="s2-2"><title>Visual Analytics Tasks</title><p>Based on the requirements (R1-R4), we derived 4 visual analytics tasks (T1-T4) to guide the system implementation.</p><sec id="s2-2-1"><title>T1: Extract MEs From Real-World Medical Texts</title><p>The system must automatically extract MEs from raw medical texts and categorize them. This transforms unstructured data into discrete, manageable knowledge units (R1), which serve as the foundational elements for visualization.</p></sec><sec id="s2-2-2"><title>T2: Support for Cascading Visual Analysis and Exploration of Texts With Varying Knowledge Densities</title><p>The system should organize extracted data into a hierarchical structure. It must provide cascading views that allow users to transition seamlessly from a high-level thematic overview to coarse-grained paragraph relations, and finally to fine-grained entity details. This capability facilitates incremental exploration and understanding of complex medical knowledge (R2).</p></sec><sec id="s2-2-3"><title>T3: Establishing a Clear Structure for Association Analysis</title><p>The system needs to construct and visualize a clear mapping of relationships, specifically the inclusion of MEs within paragraphs and the co-occurrence of elements. This visualization should clarify how knowledge units are structurally connected (R3).</p></sec><sec id="s2-2-4"><title>T4: Support for Entity-Centric Pattern Analysis</title><p>The system must provide interaction mechanisms to select a specific ME or MES as a focal point. Upon selection, the visualization should be dynamically reconfigured to display the distribution, composition, and associated context centered around the target knowledge unit (R4).</p></sec></sec><sec id="s2-3"><title>Dataset Construction</title><sec id="s2-3-1"><title>Data Source</title><p>In this paper, we analyze and process 2 realistic datasets. The first dataset is the Chinese Biomedical Language Understanding Evaluation (CBLUE) provided by the Key Laboratory of Computational Linguistics (Peking University, China) [<xref ref-type="bibr" rid="ref41">41</xref>]. CBLUE is a comprehensive benchmark compiled from authoritative medical textbooks and clinical practice records. The dataset contains approximately 96,000 MDs, covering a wide range of more than 500 common diseases. To ensure high quality, the MEs within the dataset were meticulously annotated by medical experts. These MEs are categorized into 9 classes on the basis of Chinese ME annotation standards [<xref ref-type="bibr" rid="ref41">41</xref>]: disease (dis), clinical symptoms (sym), drugs (dru), medical equipment (equ), medical procedures (pro), body (bod), medical examination items (ite), microorganisms (mic), and department (dep).</p><p>The second dataset, the Medical Text of West China School (Hospital) of Stomatology (MWCSS), was collected between 2023 and 2025 and provided by the West China School (Hospital) of Stomatology (Sichuan University, China). This dataset contains a large volume of unstructured text data from clinical practice. It includes approximately 100,000 MDs, covering various facets of medical processes, such as clinical diagnoses, treatment processes, specialist examinations, auxiliary examinations, treatment plans, interventions, and drug instructions.</p></sec><sec id="s2-3-2"><title>ME Extraction Comparative Analysis</title><p>The accuracy of ME extraction is critical to the effectiveness of data processing and visual analytics. Therefore, we systematically assessed the performance and computational resource demands of several models for ME extraction.</p><p>Due to strict data confidentiality protocols, which preclude the transmission of data to external servers, models must be deployable locally. In light of the practical limitations of resource-constrained environments, and based on the latest benchmark list from FlagEval [<xref ref-type="bibr" rid="ref42">42</xref>], we selected high-performing, open-source LLMs for evaluation: DeepSeek-70B and Qwen-32B. Additionally, we included MacBERT, an enhanced BERT model with a novel masked language modeling correction pretraining task [<xref ref-type="bibr" rid="ref43">43</xref>], which has been identified as the top-performing BERT-based model according to the CBLUE benchmark, for comparison.</p><p>Fine-tuning and inference were performed via two NVIDIA A10 GPUs or a single NVIDIA RTX 4060 Ti GPU for MacBERT. The fine-tuning parameters were as follows: learning rate of 3e-5, 2 epochs, batch size of 16, and warm-up ratio of 0.1. For DeepSeek-70B and Qwen-32B, both models required two A10 GPUs and were instruction-tuned via 4-bit quantized models with QLoRA (Quantized Low-Rank Adaptation) [<xref ref-type="bibr" rid="ref44">44</xref>] (LoRA [Low-Rank Adaptation] settings: <italic>r</italic>=16, <italic>&#x03B1;</italic>=64, dropout=0.05). All the fine-tuning parameters and LLM prompt words were selected based on benchmark tests [<xref ref-type="bibr" rid="ref41">41</xref>] and relevant comparative experiments [<xref ref-type="bibr" rid="ref45">45</xref>] and have been validated. The computational resource consumption for each model is summarized in <xref ref-type="table" rid="table1">Table 1</xref>.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Computational resource consumption comparison.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Model</td><td align="left" valign="bottom">GPU configuration (NVIDIA)</td><td align="left" valign="bottom">Fine-tuning time (hours)</td><td align="left" valign="bottom">Inference time (hours)</td></tr></thead><tbody><tr><td align="left" valign="top">DeepSeek-70B</td><td align="char" char="." valign="top">2 A10</td><td align="char" char="." valign="top">167.23</td><td align="char" char="." valign="top">134.27</td></tr><tr><td align="left" valign="top">Qwen-32B</td><td align="char" char="." valign="top">2 A10</td><td align="char" char="." valign="top">101.46</td><td align="char" char="." valign="top">83.58</td></tr><tr><td align="left" valign="top">MacBERT</td><td align="char" char="." valign="top">2 A10</td><td align="char" char="." valign="top">0.77</td><td align="char" char="." valign="top">0.18</td></tr><tr><td align="left" valign="top">MacBERT</td><td align="char" char="." valign="top">1 RTX 4060 Ti</td><td align="char" char="." valign="top">5.16</td><td align="char" char="." valign="top">0.35</td></tr></tbody></table></table-wrap><p>The results highlight a significant disparity in resource demand. While LLMs (DeepSeek-70B and Qwen-32B) require substantial fine-tuning and inference time on dual A10 GPUs, MacBERT demonstrates remarkable efficiency, completing the same tasks in a fraction of the time. Notably, MacBERT&#x2019;s inference speed on consumer-grade hardware (a single RTX 4060 Ti) suggests that it is highly suitable for local deployment in practical medical environments where high-end computational clusters may not be available.</p><p><xref ref-type="table" rid="table2">Table 2</xref> summarizes the performance metrics for each model. Fine-tuning consistently improved performance across all the models. Notably, the fine-tuned MacBERT model achieves the highest <italic>F</italic><sub>1</sub>-score (0.605), surpassing both LLMs, and its precision approaches human annotation levels [<xref ref-type="bibr" rid="ref41">41</xref>].</p><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>Performance comparison of models for medical entity extraction.</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Model</td><td align="left" valign="bottom">Type</td><td align="left" valign="bottom">Precision</td><td align="left" valign="bottom">Recall</td><td align="left" valign="bottom"><italic>F</italic><sub>1</sub>-score</td></tr></thead><tbody><tr><td align="left" valign="top">DeepSeek-70B</td><td align="left" valign="top">Base</td><td align="char" char="." valign="top">0.548</td><td align="char" char="." valign="top">0.487</td><td align="char" char="." valign="top">0.516</td></tr><tr><td align="left" valign="top">DeepSeek-70B</td><td align="left" valign="top">Fine-tuned</td><td align="char" char="." valign="top">0.593</td><td align="char" char="." valign="top">0.521</td><td align="char" char="." valign="top">0.535</td></tr><tr><td align="left" valign="top">Qwen-32B</td><td align="left" valign="top">Base</td><td align="char" char="." valign="top">0.554</td><td align="char" char="." valign="top">0.491</td><td align="char" char="." valign="top">0.525</td></tr><tr><td align="left" valign="top">Qwen-32B</td><td align="left" valign="top">Fine-tuned</td><td align="char" char="." valign="top">0.608</td><td align="char" char="." valign="top">0.542</td><td align="char" char="." valign="top">0.561</td></tr><tr><td align="left" valign="top">MacBERT</td><td align="left" valign="top">Base</td><td align="char" char="." valign="top">0.595</td><td align="char" char="." valign="top">0.549</td><td align="char" char="." valign="top">0.571</td></tr><tr><td align="left" valign="top">MacBERT</td><td align="left" valign="top">Fine-tuned</td><td align="char" char="." valign="top">0.636</td><td align="char" char="." valign="top">0.579</td><td align="char" char="." valign="top">0.605</td></tr></tbody></table></table-wrap><p>This analysis demonstrates that MacBERT not only delivers the best performance after fine-tuning but also requires significantly fewer computational resources than the LLMs. MacBERT, therefore, emerges as the optimal choice for ME extraction tasks. This conclusion aligns with the literature [<xref ref-type="bibr" rid="ref46">46</xref>], which emphasizes that while LLMs excel in generative tasks, BERT-based models retain distinct advantages in NER and other specialized domains [<xref ref-type="bibr" rid="ref47">47</xref>] and sentiment analysis [<xref ref-type="bibr" rid="ref48">48</xref>].</p><p>Beyond common medical conditions, the system&#x2019;s robustness concerning rare diseases and less common terminology warrants further discussion. Although our evaluation used a dataset of common diseases&#x2014;where MacBERT demonstrated strong performance&#x2014;recent research [<xref ref-type="bibr" rid="ref49">49</xref>] suggests that the performance landscape may shift when encountering &#x201C;long-tail&#x201D; medical data. Specifically, while fine-tuned BERT-based models generally maintain overall superiority, LLMs can exhibit superior capabilities in identifying rare disease entities in few-shot settings. Therefore, for specialized tasks focusing on rare diseases where relevant samples and sufficient computational resources are available, LLMs may be used to achieve higher precision. Nevertheless, within our current visual analytics framework centered on general-purpose unstructured medical text, MacBERT provides the most reliable performance-to-cost ratio.</p><p>For NER requiring deep semantic insight, language-specific pretrained models often outperform general multilingual models [<xref ref-type="bibr" rid="ref50">50</xref>]. In this study, we use MacBERT&#x2014;the top-performing model on the CBLUE dataset&#x2014;to maximize extraction accuracy for Chinese corpora. Notably, the underlying pipeline remains highly adaptable; the base encoder can be seamlessly replaced (eg, substituting BioBERT [<xref ref-type="bibr" rid="ref51">51</xref>] for MacBERT) to process English or other language datasets.</p></sec><sec id="s2-3-3"><title>Data Processing</title><p><xref ref-type="fig" rid="figure2">Figure 2</xref> shows the pipeline of data processing. The process can be described as follows.</p><fig position="float" id="figure2"><label>Figure 2.</label><caption><p>Pipeline of data processing. First, MacBERT was fine-tuned using labeled data from CBLUE. Individual MPs within the MDs are processed via the fine-tuned MacBERT model to extract MEs, which are subsequently organized into MESs. These texts are input into ESimCSE-BERT to generate embeddings, which are then stored in a vector database. The MD embeddings serve as the basis for topic clustering in BERTopic. The resulting topics of MDs, along with all medical texts and their inclusion and co-occurrence relationships, are then structured into a graph and stored in a graph database. CBLUE: Chinese Biomedical Language Understanding Evaluation; <italic>D</italic>: medical document; <italic>E</italic>: medical entity; MD: medical document; ME: medical entity; MES: medical entity set; MP: medical paragraph; MWCSS: Medical Text of West China School (Hospital) of Stomatology; <italic>P</italic>: medical paragraph; <italic>SE</italic>: medical entity set.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e83785_fig02.png"/></fig><sec id="s2-3-3-1"><title>Step 1: ME Extraction and Structuring</title><p>According to the results of the ME Extraction comparative analysis, we used the MacBERT model fine-tuned with the CBLUE dataset to extract MEs from MWCSS, using MP of MD <italic>D<sub>i</sub></italic> as the extraction unit, thereby constructing MESs.</p></sec><sec id="s2-3-3-2"><title>Step 2: Vectorization</title><p>All the text from CBLUE and MWCSS, including MDs, MPs, MESs, and MEs, is processed via ESimCSE-BERT [<xref ref-type="bibr" rid="ref52">52</xref>]&#x2014;an efficient and improved unsupervised sentence embedding method that generates high-quality sentence vectors.</p></sec><sec id="s2-3-3-3"><title>Step 3: Topic Clustering</title><p>The high-quality MD vectors from step 2 were input into the BERTopic model. By leveraging these superior embeddings, the clustering process produces topic assignments for each MD. The results yielded a topic diversity score [<xref ref-type="bibr" rid="ref53">53</xref>] of 0.975, indicating a high degree of distinctness among the generated categories. To further validate the clinical utility, a 5-point Likert scale assessment was conducted by medical experts (E3, E4, E6, E7, and E8) to evaluate the effectiveness of the generated topics and keywords (<xref ref-type="table" rid="table3">Table 3</xref>). Following the evaluation framework proposed in [<xref ref-type="bibr" rid="ref54">54</xref>], the results suggest that the interpretability, distinctiveness, and relevance of the topic clusters are at a high level. These findings further corroborate the superiority of transformer-based topic models as discussed in [<xref ref-type="bibr" rid="ref54">54</xref>].</p><table-wrap id="t3" position="float"><label>Table 3.</label><caption><p>Evaluation of topic clustering (1&#x2010;5 Likert scale).</p></caption><table id="table3" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Evaluation</td><td align="left" valign="bottom">Explanation</td><td align="left" valign="bottom">Score, mean (SD)</td></tr></thead><tbody><tr><td align="left" valign="top">Interpretability</td><td align="left" valign="top">How easily a human can assign a coherent meaning to the topic.</td><td align="char" char="parenthesis" valign="top">4.3 (0.57)</td></tr><tr><td align="left" valign="top">Distinctiveness</td><td align="left" valign="top">How well the topic is differentiated from other topics.</td><td align="char" char="parenthesis" valign="top">4.4 (0.41)</td></tr><tr><td align="left" valign="top">Relevance</td><td align="left" valign="top">How well the topic captures meaningful domain content.</td><td align="char" char="parenthesis" valign="top">4.6 (0.22)</td></tr></tbody></table></table-wrap></sec><sec id="s2-3-3-4"><title>Step 4: Relationship Mapping and Storage</title><p>To support complex analytical queries, we implemented a dual-storage strategy addressing the inherent relational complexity and interconnectedness of medical data.</p><sec id="s2-3-3-4-1"><title>Semantic Storage in the Vector Database</title><p>All textual elements (MDs, MPs, MESs, and MEs) are embedded into high-dimensional vectors and stored in a vector database. This infrastructure leverages various optimization algorithms to enable the efficient execution of similarity-based queries and fast calculations across datasets in the tens of millions [<xref ref-type="bibr" rid="ref55">55</xref>].</p></sec><sec id="s2-3-3-4-2"><title>Structural Storage in the Graph Database</title><p>Each text unit is modeled as a vertex, with MD vertices carrying the topic attributes derived from clustering. The inclusion relationships between texts and the co-occurrence of MEs within MESs are represented as edges. This graph structure is stored in a graph database to facilitate high-performance queries across complex entity relationships [<xref ref-type="bibr" rid="ref56">56</xref>]. The scalability and query efficiency of this graph-based approach have become increasingly evident with larger datasets [<xref ref-type="bibr" rid="ref57">57</xref>].</p></sec></sec></sec></sec><sec id="s2-4"><title>Visual Analytics Systems Development</title><sec id="s2-4-1"><title>System Overview and Workflow</title><p>Building upon the constructed dataset, we developed MExplore, a multilevel visual analytics system designed to facilitate progressive exploration and cascading visual analysis (T2). The system adopts a 3-tiered metaphor&#x2014;cosmic space, star map, and planet cross-section&#x2014;which inspires the design of the MD space view (<xref ref-type="fig" rid="figure3">Figure 3A</xref>), the MP star map (<xref ref-type="fig" rid="figure3">Figure 3B</xref>), and the focused sectional view (<xref ref-type="fig" rid="figure3">Figure 3C</xref>). We use a hierarchical layout that systematically guides users from macrolevel document retrieval to microlevel detail exploration.</p><p>The analytical workflow follows a top-down trajectory (<xref ref-type="fig" rid="figure3">Figure 3</xref>), starting from the MD space view, where users filter and define a subset of MDs as the basis for all subsequent analysis based on cluster and spatial distributions. Once the analytical scope is established, the system transitions to the MP star map, enabling users to identify structural patterns and semantic clusters within MP subgraphs. From there, users can incorporate specific subgraphs into the association analysis view to reveal intricate relationships between MESs and MEs. For the most granular inquiry, users can select individual nodes within this view to generate a corresponding focused sectional view, facilitating a deep dive into the target knowledge unit&#x2019;s context.</p><p>By structuring the analysis as an interactive, stepwise process, MExplore allows users to progressively increase analytical granularity without losing global contextual coherence [<xref ref-type="bibr" rid="ref58">58</xref>]. Furthermore, the design leverages grounded metaphors to map complex information onto pre-existing cognitive schemas [<xref ref-type="bibr" rid="ref59">59</xref>], effectively mitigating the visual complexity encountered by the user and enhancing memory encoding for knowledge retention [<xref ref-type="bibr" rid="ref60">60</xref>]. Ultimately, this multitiered approach facilitates a continuous analytical loop.</p><fig position="float" id="figure3"><label>Figure 3.</label><caption><p>The MExplore framework: (A) the MD space view, inspired by cosmic space, where users select MDs and construct and partition graphs of the contained MPs; (B) the MP Star map, inspired by the star map, where users choose corresponding MP subgraphs and perform relational analysis; (C) the focused sectional view, inspired by the planet cross-section, which enables detailed, focused analysis of the selected data. <italic>F</italic><sub><italic>collision</italic></sub>: the collision force; <italic>F</italic><sub><italic>gravity</italic></sub>: the gravity force; <italic>F</italic><sub><italic>spring</italic></sub>: the spring force; MD: medical document; ME: medical entity; MES: medical entity set; MP: medical paragraph.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e83785_fig03.png"/></fig></sec><sec id="s2-4-2"><title>MD Space View</title><p>As shown in <xref ref-type="fig" rid="figure4">Figure 4A</xref>, upon entering keywords, the system executes a full-text vector search to retrieve relevant MDs. The retrieved MDs are mapped into a 3D force-directed graph where each MD is represented as a planet&#x2014;its size encodes text length, and its position is determined by semantic similarity. Interdocument links visualize this similarity. Users can adjust a threshold slider to filter connections by similarity, aggregating MDs into topic-based clusters represented as nebulae labeled with topic titles. This layout emphasizes core structural relationships and allows users to follow a progressive approach starting from these nebulae, clicking a specific nebula to enter Focus Mode to explore individual MDs in detail. This interaction mechanism leverages spatial proximity to facilitate the efficient identification of areas of interest [<xref ref-type="bibr" rid="ref61">61</xref>].</p><p>The colors of the nebulae and their constituent planets represent their respective topics. This color encoding is discarded upon transitioning to detailed analysis views, preventing semantic interference between the MD topic and ME attributes.</p><p>To facilitate comprehensive discovery, a guide for unexplored nebulae is displayed in the lower-left corner, allowing for thorough exploration and rapid focusing by clicking. Simultaneously, the panel below ranks MDs by vector similarity scores to support instant focused viewing and prevent users from overlooking important MDs. Throughout this process, users iteratively select MDs to build an analytical collection, which serves as the foundation for all subsequent analysis views.</p><fig position="float" id="figure4"><label>Figure 4.</label><caption><p>The MExplore system: (A) MD space view; (B) MP star map; (C) association analysis view; (D) focused sectional view; (E) provenance view. MD: medical document; MP: medical paragraph.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e83785_fig04.png"/></fig></sec><sec id="s2-4-3"><title>MP Star Map</title><p>Relying solely on full-text embedding similarity can introduce bias, as document representations may disproportionately reflect longer sections while overshadowing concise but critical information. For instance, detailed medical histories in outpatient records can overshadow concise diagnosis sections during similarity computation.</p><p>To mitigate this, MDs are decomposed into MPs, each modeled as a graph vertex. Intradocument edges <italic>E<sub>d</sub></italic> connect the MPs within the same MD. Cross-document MP pairs (<xref ref-type="fig" rid="figure5">Figure 5A</xref>) exceeding the similarity threshold set in <xref ref-type="fig" rid="figure4">Figure 4A</xref> are linked by similarity edges <italic>E<sub>s</sub></italic>. The KaFFPa algorithm [<xref ref-type="bibr" rid="ref62">62</xref>] then partitions the graph, segregating semantically distinct MPs into subgraphs (<xref ref-type="fig" rid="figure5">Figure 5B</xref>).</p><fig position="float" id="figure5"><label>Figure 5.</label><caption><p>For detailed analysis, MDs are decomposed into MPs. (A) MPs are represented as vertices, with <italic>E<sub>d</sub></italic> connecting MPs within the same MD, and the similarity of MPs across connected MDs is computed. (B) MPs whose similarity is greater <italic>&#x03B8;</italic> than are connected by <italic>E<sub>s</sub></italic>, and the graph is partitioned by KaFFPa. (C) The partitioned subgraphs are visualized via the constellation metaphor, where <italic>E<sub>d</sub></italic> are shown as lines and the MPs are shown as stars. Each star is subject to <italic>F</italic><sub><italic>intra</italic></sub> from <italic>E<sub>d</sub></italic> and <italic>F</italic><sub><italic>similarity</italic></sub> from <italic>E<sub>s</sub></italic>. MD: medical document; MP: medical paragraph.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e83785_fig05.png"/></fig><p>Each resulting subgraph is visualized as a constellation in the MP star map (<xref ref-type="fig" rid="figure4">Figure 4B</xref>), the MPs are depicted as stars, and <italic>E<sub>d</sub></italic> is rendered as a constellation line, whereas <italic>E</italic><sub><italic>s</italic></sub> is omitted to reduce clutter but is retained as latent factors influencing the layout, the similarity-based gravitational force <italic>F</italic><sub><italic>similarity</italic></sub>, and the structural gravitational force <italic>F</italic><sub><italic>intra</italic></sub> (<xref ref-type="fig" rid="figure4">Figure 4C</xref>). The combined gravitational force of the star <italic>i</italic> is calculated as follows:</p><disp-formula id="E1"><label>(1)</label><mml:math id="eqn1"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mstyle displaystyle="true" scriptlevel="0"><mml:msubsup><mml:mi>F</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>v</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:msubsup><mml:mi>F</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msubsup><mml:mo>+</mml:mo><mml:msubsup><mml:mi>F</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>m</mml:mi><mml:mi>i</mml:mi><mml:mi>l</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msubsup></mml:mstyle></mml:mrow></mml:mstyle></mml:math></disp-formula><disp-formula id="E2"><label>(2)</label><mml:math id="eqn2"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:msubsup><mml:mi>F</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>m</mml:mi><mml:mi>i</mml:mi><mml:mi>l</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>m</mml:mi><mml:mi>i</mml:mi><mml:mi>l</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo>&#x003E;</mml:mo><mml:mi>&#x03B8;</mml:mi></mml:mrow></mml:munderover><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo><mml:mi>F</mml:mi><mml:mo symmetric="true">&#x2016;</mml:mo></mml:mrow><mml:mo>&#x22C5;</mml:mo><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>m</mml:mi><mml:mi>i</mml:mi><mml:mi>l</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo stretchy="false">)</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:mo>&#x22C5;</mml:mo><mml:mrow><mml:mover><mml:msubsup><mml:mi>d</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo stretchy="false">^</mml:mo></mml:mover></mml:mrow></mml:mstyle></mml:mstyle></mml:mrow></mml:mstyle></mml:math></disp-formula><p>where <italic>F</italic><sub><italic>intra</italic></sub> is the link force connecting the MPs within the same MD, with a magnitude of unit force <inline-formula><mml:math id="ieqn1"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo><mml:mi>F</mml:mi><mml:mo symmetric="true">&#x2016;</mml:mo></mml:mrow></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula>, <italic>&#x03B8;</italic> is the threshold set by the user, and <inline-formula><mml:math id="ieqn2"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>m</mml:mi><mml:mi>i</mml:mi><mml:mi>l</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula> is the similarity between MP <italic>i</italic> and MP <italic>j</italic>. <inline-formula><mml:math id="ieqn3"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:msubsup><mml:mrow><mml:mover><mml:mi>d</mml:mi><mml:mo stretchy="false">^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula> is the unit vector of the force direction from MP <italic>i</italic> to MP <italic>j</italic>.</p><p>In addition, each star <italic>i</italic> is subjected to spring forces <inline-formula><mml:math id="ieqn4"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:msubsup><mml:mi>F</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>p</mml:mi><mml:mi>r</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msubsup></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula> from 9 ME-type poles uniformly distributed on the circular boundary of the star map:</p><disp-formula id="E3"><label>(3)</label><mml:math id="eqn3"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:msubsup><mml:mi>F</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>p</mml:mi><mml:mi>r</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>9</mml:mn></mml:mrow></mml:munderover><mml:mfrac><mml:mrow><mml:mi>n</mml:mi><mml:mi>u</mml:mi><mml:msubsup><mml:mi>m</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>n</mml:mi><mml:mi>u</mml:mi><mml:msub><mml:mi>m</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfrac><mml:mo>&#x22C5;</mml:mo><mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo><mml:mi>F</mml:mi><mml:mo symmetric="true">&#x2016;</mml:mo></mml:mrow><mml:mo>&#x22C5;</mml:mo><mml:msubsup><mml:mrow><mml:mover><mml:mi>d</mml:mi><mml:mo stretchy="false">^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup></mml:mstyle></mml:mstyle></mml:mrow></mml:mstyle></mml:math></disp-formula><p>where <inline-formula><mml:math id="ieqn5"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:mi>n</mml:mi><mml:mi>u</mml:mi><mml:msubsup><mml:mi>m</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula> is the number of MEs of type <italic>j</italic> in MP <italic>i</italic>, and <italic>num<sub>i</sub></italic> is the total number of MEs in MP <italic>i</italic>. <inline-formula><mml:math id="ieqn6"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:msubsup><mml:mrow><mml:mover><mml:mi>d</mml:mi><mml:mo stretchy="false">^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula> is the unit vector of the force direction from MP <italic>i</italic> to <italic>V<sub>j</sub></italic>.</p><p>To prevent occlusion, each star <italic>i</italic> is subject to the collision force <inline-formula><mml:math id="ieqn7"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:msubsup><mml:mi>F</mml:mi><mml:mrow><mml:mi>c</mml:mi><mml:mi>o</mml:mi><mml:mi>l</mml:mi><mml:mi>l</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msubsup></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula> of stars that may overlap it. The final combined force <inline-formula><mml:math id="ieqn8"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:msubsup><mml:mi>F</mml:mi><mml:mrow><mml:mi>c</mml:mi><mml:mi>o</mml:mi><mml:mi>l</mml:mi><mml:mi>l</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msubsup></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula> is calculated as follows:</p><disp-formula id="E4"><label>(4)</label><mml:math id="eqn4"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mstyle displaystyle="true" scriptlevel="0"><mml:msubsup><mml:mi>F</mml:mi><mml:mrow><mml:mi>c</mml:mi><mml:mi>o</mml:mi><mml:mi>m</mml:mi><mml:mi>b</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>e</mml:mi><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:msubsup><mml:mi>F</mml:mi><mml:mrow><mml:mi>g</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>v</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msubsup><mml:mo>+</mml:mo><mml:msubsup><mml:mi>F</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>p</mml:mi><mml:mi>r</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msubsup><mml:mo>+</mml:mo><mml:msubsup><mml:mi>F</mml:mi><mml:mrow><mml:mi>c</mml:mi><mml:mi>o</mml:mi><mml:mi>l</mml:mi><mml:mi>l</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msubsup></mml:mstyle></mml:mrow></mml:mstyle></mml:math></disp-formula><p>This multiforce design achieved semantically coherent clustering (<italic>F<sub>spring</sub></italic>), representation of cross-document similarity and structural relationships (<italic>F<sub>gravity</sub></italic>), and clear visualization (<italic>F<sub>collision</sub></italic>) (<xref ref-type="fig" rid="figure4">Figure 4B</xref>). Stars within the same constellation are assigned a uniform color, allowing users to quickly identify them even on a densely populated map. This color is modulated by a sequential luminance gradient [<xref ref-type="bibr" rid="ref63">63</xref>] to encode the star count: brighter constellations denote a greater number of constituent stars, enabling users to efficiently pinpoint pattern-dense regions. The luminance of star borders encodes the ME count, providing information density cues (T2). To preserve macrocontext awareness during the analysis of MPs and subsequent tasks, a topic navigation list at the bottom of the sidebar (<xref ref-type="fig" rid="figure4">Figure 4</xref>) visualizes the topic information of selected MDs. In the MP star map and subsequent views, users can click on MP/MES nodes or document cards to highlight the topic of the corresponding document. This mechanism enables on-demand backtracking, ensuring users remain oriented within the macrolevel topic context. This design facilitates progressive schema construction [<xref ref-type="bibr" rid="ref33">33</xref>,<xref ref-type="bibr" rid="ref34">34</xref>] and supports iterative exploratory learning.</p></sec><sec id="s2-4-4"><title>Association Analysis View</title><p>Leveraging the concepts within MPs to create visual representations of hierarchical structures can significantly increase the efficiency of knowledge navigation [<xref ref-type="bibr" rid="ref64">64</xref>,<xref ref-type="bibr" rid="ref65">65</xref>]. Therefore, we extract MEs from each MP (T1), compose the MES, and visualize the resulting data via a radial dendrogram. Compared with alternative text visualization methods, such as word clouds or Sankey diagrams, this approach effectively captures both inclusion and relational connections among elements while accurately displaying hierarchical structures. To further enhance visualization, we propose an algorithm that groups MESs containing shared MEs into the same tree branch, thereby emphasizing co-occurrence relationships between the MESs (T3). The algorithm is outlined in <xref ref-type="other" rid="box1">Textbox 1</xref>.</p><boxed-text id="box1"><title> Algorithm for tree construction.</title><p><bold>Input:</bold></p><list list-type="order"><list-item><p>Root node of the tree: <inline-formula><mml:math id="ieqn9"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula></p></list-item><list-item><p>Set of nodes to be added: <inline-formula><mml:math id="ieqn10"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:msub><mml:mi>N</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula></p></list-item></list><p><bold>Output:</bold> Root node of the tree: <inline-formula><mml:math id="ieqn11"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula></p><list list-type="order"><list-item><p><bold>for</bold> each node <inline-formula><mml:math id="ieqn12"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:msub><mml:mi>n</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula> in <inline-formula><mml:math id="ieqn13"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:msub><mml:mi>N</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula> <bold>do</bold></p></list-item><list-item><p><inline-formula><mml:math id="ieqn14"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:mi>C</mml:mi><mml:mo>=</mml:mo><mml:mi>G</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi><mml:mi>M</mml:mi><mml:mi>E</mml:mi><mml:mi>s</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>n</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula></p></list-item><list-item><p><inline-formula><mml:math id="ieqn15"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:mi>N</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mi>M</mml:mi><mml:mrow><mml:mi>c</mml:mi><mml:mi>h</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>k</mml:mi><mml:mi>e</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>T</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>v</mml:mi><mml:mi>e</mml:mi><mml:mi>r</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi><mml:mi>T</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>e</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>C</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula></p></list-item><list-item><p>end for</p></list-item><list-item><p><inline-formula><mml:math id="ieqn16"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:msub><mml:mi>N</mml:mi><mml:mrow><mml:mi>c</mml:mi><mml:mi>h</mml:mi><mml:mi>i</mml:mi><mml:mi>l</mml:mi><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>r</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">&#x27F5;</mml:mo><mml:mi>N</mml:mi><mml:mo>&#x222A;</mml:mo><mml:msub><mml:mi>M</mml:mi><mml:mrow><mml:mi>c</mml:mi><mml:mi>h</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>k</mml:mi><mml:mi>e</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula></p></list-item><list-item><p><bold>for</bold> each node <inline-formula><mml:math id="ieqn17"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:msub><mml:mi>n</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula> in <inline-formula><mml:math id="ieqn18"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:mi>N</mml:mi></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula> <bold>do</bold></p></list-item><list-item><p><inline-formula><mml:math id="ieqn19"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>m</mml:mi><mml:mi>o</mml:mi><mml:mi>v</mml:mi><mml:mi>e</mml:mi><mml:mi>F</mml:mi><mml:mi>r</mml:mi><mml:mi>o</mml:mi><mml:mi>m</mml:mi><mml:mi>T</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>e</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>e</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mi>n</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula></p></list-item><list-item><p><bold>end for</bold></p></list-item><list-item><p><bold>if</bold> <inline-formula><mml:math id="ieqn20"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:msub><mml:mo>&#x003E;</mml:mo><mml:mn>1</mml:mn></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula> <bold>then</bold></p></list-item><list-item><p><inline-formula><mml:math id="ieqn21"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:msub><mml:mi>n</mml:mi><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">&#x27F5;</mml:mo><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>m</mml:mi><mml:mi>m</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mi>F</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mi>h</mml:mi><mml:mi>e</mml:mi><mml:mi>r</mml:mi><mml:mi>N</mml:mi><mml:mi>o</mml:mi><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>N</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula></p></list-item><list-item><p><bold>if</bold> <inline-formula><mml:math id="ieqn22"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:msub><mml:mi>n</mml:mi><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula> is not <inline-formula><mml:math id="ieqn23"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:mo stretchy="false">(</mml:mo><mml:mi>n</mml:mi><mml:mi>u</mml:mi><mml:mi>l</mml:mi><mml:mi>l</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula> <bold>then</bold></p></list-item><list-item><p><inline-formula><mml:math id="ieqn24"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:msub><mml:mi>n</mml:mi><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub><mml:mo>&#x22C5;</mml:mo><mml:mi>C</mml:mi><mml:mi>h</mml:mi><mml:mi>i</mml:mi><mml:mi>l</mml:mi><mml:mi>d</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi><mml:mo stretchy="false">&#x27F5;</mml:mo><mml:msub><mml:mi>N</mml:mi><mml:mrow><mml:mi>c</mml:mi><mml:mi>h</mml:mi><mml:mi>i</mml:mi><mml:mi>l</mml:mi><mml:mi>d</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula></p></list-item><list-item><p><bold>else</bold></p></list-item><list-item><p><inline-formula><mml:math id="ieqn25"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mo>&#x22C5;</mml:mo><mml:mi>c</mml:mi><mml:mi>h</mml:mi><mml:mi>i</mml:mi><mml:mi>l</mml:mi><mml:mi>d</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi><mml:mo stretchy="false">&#x27F5;</mml:mo><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mo>&#x22C5;</mml:mo><mml:mi>C</mml:mi><mml:mi>h</mml:mi><mml:mi>i</mml:mi><mml:mi>l</mml:mi><mml:mi>d</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:mo>&#x222A;</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:msub><mml:mi>N</mml:mi><mml:mrow><mml:mi>c</mml:mi><mml:mi>h</mml:mi><mml:mi>i</mml:mi><mml:mi>l</mml:mi><mml:mi>d</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula></p></list-item><list-item><p><bold>end if</bold></p></list-item><list-item><p><bold>end if</bold></p></list-item><list-item><p><bold>return</bold> <inline-formula><mml:math id="ieqn26"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula></p></list-item></list></boxed-text><p>For each MP, the extracted MEs form an MES, which acts as a node. The newly added node set <italic>N<sub>add</sub></italic>, along with the root node of the tree <italic>r<sub>tree</sub></italic>, is provided as input. For each node <italic>n<sub>i</sub></italic> in <italic>N<sub>add</sub></italic>, the algorithm retrieves its associated MEs via <italic>GetMEs</italic>(<italic>n<sub>i</sub></italic>) function. It then traverses the tree through <italic>TraverseTree</italic> to identify the set <italic>M<sub>checked</sub></italic> of co-occurring MEs, along with the node set <italic>N</italic> containing these MEs. They are all grouped into the set <italic>N<sub>children</sub></italic>, which are subsequently added as children to the appropriate parent nodes.</p><p>If multiple nodes intersect, their common ancestor is identified by <italic>CommonFatherNode</italic>, and the new nodes are added as children of this ancestor. The nodes are added directly to the root if no common ancestor is found. The process concludes by returning the updated tree structure.</p><p>Based on the algorithm described above, the association analysis view is constructed (<xref ref-type="fig" rid="figure4">Figure 4C</xref>). To manage the cognitive load and ensure visual consistency, the system uses a synchronized color encoding strategy and a progressive disclosure mechanism. Categorical colors are assigned to ME nodes based on their entity types, while for MES nodes, the color is derived from a weighted mixture of its constituent ME types. To maintain cross-view coherence, this encoding scheme is applied uniformly across both the association analysis and focused sectional views, supported by a persistent legend in the sidebar that ensures color-to-type mappings remain readily accessible throughout the analytical process.</p><p>The risk of information overload is further mitigated by initially displaying only the aggregated MES nodes, allowing users to perceive the high-level relational structure without being overwhelmed by individual entities. Users can then interactively expand specific MES nodes to reveal their constituent MEs as independent child nodes. This progressive disclosure strategy enables users to manage information density dynamically according to their analytical needs.</p><p>These design choices facilitate the perception of complex hierarchical and relational patterns, thereby supporting schema construction and clinical reasoning by externalizing the learner&#x2019;s cognitive processes [<xref ref-type="bibr" rid="ref66">66</xref>,<xref ref-type="bibr" rid="ref67">67</xref>]. Furthermore, the system architecture supports the dynamic addition of new MPs, enabling real-time reconstruction of the association tree. This capability allows seamless integration of new clinical information into the existing cognitive framework of the user, promoting iterative, exploratory learning and long-term memory retrieval [<xref ref-type="bibr" rid="ref68">68</xref>].</p></sec><sec id="s2-4-5"><title>Focused Sectional View</title><p>To enable granular analysis of a specific ME or MES, the focused sectional view (<xref ref-type="fig" rid="figure4">Figure 4D</xref>) uses a &#x201C;planetary layer&#x201D; metaphor to organize complex co-occurrence data. The central ring functions as a donut chart that encodes the proportional distribution of ME classifications. If the analytical focus is a single ME, this layer identifies its specific classification. Surrounding this core, the mantle layer uses a polar-axis-aligned area chart to visualize co-occurring MESs. In this arrangement, each radial axis represents a distinct MES, with the axis height being <inline-formula><mml:math id="ieqn27"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:msubsup><mml:mi>N</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:mo>&#x2217;</mml:mo><mml:msub><mml:mi>h</mml:mi><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula>, which represents the number of MEs belonging to class <italic>k</italic> in the <italic>i</italic>th MES. To emphasize semantic continuity across the dataset, areas representing the same ME classification are connected across adjacent axes. Compared with traditional area or stream charts, this radial layout effectively portrays the compositional richness and semantic associations of the target entity within a compact structural signature.</p><p>To bridge the gap between visualization and raw text, the system allows users to click a radial axis to query the underlying database and retrieve the corresponding MD, which is then displayed as a document card in the panel below. This design enables users to revert to the original evidence for detailed, word-by-word analysis. To further prevent highlighting overload in the text view, the system implements a selective filtering strategy. While the card highlights various extracted MEs, users can click specific semantic regions in the polar area chart to isolate and highlight only the ME category of interest. This interaction triggers the document card to automatically scroll to the relevant text segment, enabling an efficient and focused review of the original evidence without cognitive distraction.</p><p>The resulting structural signatures illustrate the contextual association patterns unique to each focused ME/MES (T4). Furthermore, the mantle&#x2019;s radial patterns facilitate the interactive comparison of MESs with similar ME configurations. This comparative analysis fosters deeper cognitive engagement and more enduring conceptual impressions, supporting the learner&#x2019;s ability to internalize complex clinical relationships [<xref ref-type="bibr" rid="ref67">67</xref>].</p></sec><sec id="s2-4-6"><title>Provenance View</title><p>To facilitate knowledge retention and structure the analytical process, we developed the provenance view (<xref ref-type="fig" rid="figure4">Figure 4E</xref>). This view enables users to capture high-fidelity snapshots from any analytical view on demand. These snapshots function as interactive nodes on a scalable, zoomable canvas, allowing users to freely organize layouts and establish logical links between visual findings. By externalizing exploration history into an explicit visual learning path, users can revisit and inspect detailed analytical states at any time, effectively transforming ephemeral interactions into organized knowledge assets.</p></sec></sec><sec id="s2-5"><title>Evaluation Design</title><sec id="s2-5-1"><title>Evaluation Overview</title><p>To comprehensively evaluate the usability, effectiveness, and expertise acquisition capabilities of MExplore, we used a mixed methods evaluation strategy comprising three components: (1) case studies with domain experts to assess the practicality, (2) a comparative user study with medical students to quantify learning outcomes, and (3) semistructured interviews with experts to evaluate usability and future potential of the system.</p></sec><sec id="s2-5-2"><title>Implementation and Apparatus</title><p>As a web-based visual analysis system, MExplore was developed using the d3.js and Django frameworks. A Microsoft Windows platform with a 2.71 GHz Intel Core i5-7300 CPU and 8 GB of memory was used as the front-end page server. The evaluation experiments were performed via a Google Chrome web browser.</p></sec><sec id="s2-5-3"><title>Case Study</title><p>The case study can demonstrate feasibility and usability in performing real-world tasks [<xref ref-type="bibr" rid="ref69">69</xref>]. Therefore, we conducted case studies with domain experts (E3 and E5) to evaluate MExplore within authentic analytical scenarios. Through extensive discussion, the experts identified three critical stages of medical expertise acquisition as core tasks: (1) the discovery of areas of interest, (2) the association analysis, and (3) the construction of illness scripts. The session commenced with a 20-minute tutorial on MExplore&#x2019;s features, followed by 1 hour of autonomous exploration focused on the predefined tasks. To capture nuanced qualitative feedback, a think-aloud protocol was used, which encouraged participants to externalize their thought process, ask questions, or provide verbal comments throughout the session.</p></sec><sec id="s2-5-4"><title>User Study</title><sec id="s2-5-4-1"><title>Study Overview</title><p>The primary objective of the user study was to quantitatively assess the effectiveness of MExplore in facilitating the acquisition of medical expertise. We adopted illness script construction as the evaluative metric to operationalize the measurement of expertise [<xref ref-type="bibr" rid="ref4">4</xref>]. The experimental design and comparative study protocols were informed by established methodologies in visual analytics and medical education research [<xref ref-type="bibr" rid="ref15">15</xref>,<xref ref-type="bibr" rid="ref70">70</xref>-<xref ref-type="bibr" rid="ref72">72</xref>].</p></sec><sec id="s2-5-4-2"><title>Participants</title><p>We recruited 20 second-year undergraduate medical students (10 male and 10 female). None of the participants had systematic prior knowledge of the specific diseases selected for the task, thereby minimizing potential confounding effects from background knowledge.</p></sec><sec id="s2-5-4-3"><title>Tasks and Materials</title><p>Three diseases&#x2014;oral candidiasis (D1), meningitis (D2), and herpes zoster (D3)&#x2014;were selected from a list of diseases for which illness scripts could be activated [<xref ref-type="bibr" rid="ref73">73</xref>]. The participants had not studied these diseases systematically beforehand, thereby eliminating the potential confounding effect of prior knowledge. Following prior studies [<xref ref-type="bibr" rid="ref70">70</xref>], the task began by providing the participants with a brief description of typical cases for each disease, including relevant medical history and examination results, excluding the disease name. Based on this information, participants were asked to identify the disease and complete an illness script template, the information of which was divided into 3 categories: enabling conditions (EC), fault (FT), and consequences (CQ) [<xref ref-type="bibr" rid="ref28">28</xref>,<xref ref-type="bibr" rid="ref73">73</xref>].</p></sec><sec id="s2-5-4-4"><title>Experimental Procedure</title><p>Participants were randomly assigned to two balanced groups (n=10 each, 5 males and 5 females). The experimental group (MEX) used MExplore to complete the tasks. The control group (OTH) was permitted to use external resources (eg, textbooks, case retrieval systems, and LLMs) but was restricted from using MExplore. Before the main task, the MEX group received a detailed introduction and a 20-minute training session to familiarize themselves with the visual views.</p></sec><sec id="s2-5-4-5"><title>Metrics</title><p>Performance was evaluated based on the accuracy and time taken to form illness scripts. Accuracy was scored by comparing responses to a standardized answer key established by the expert panel. Each correct information unit was awarded one point. The total score was normalized to a percentage [<xref ref-type="bibr" rid="ref74">74</xref>]. Furthermore, to assess the long-term impact of MExplore on expertise retention, a follow-up test was conducted 2 weeks after the initial experiment. The participants were asked to fill out the answer sheets again based solely on recall, without access to the system or external aids [<xref ref-type="bibr" rid="ref75">75</xref>].</p></sec><sec id="s2-5-4-6"><title>Statistical Analysis</title><p>All analyses were conducted via Python (version 3.10). The performance was evaluated via 3 metrics: accuracy, retention accuracy, and completion time. Both accuracy and retention accuracy are expressed as percentages. Independent sample <italic>t</italic> tests were used to compare the mean differences between the MEX and OTH groups. Two-sided <italic>P</italic>&#x003C;.05 was considered statistically significant.</p></sec></sec></sec><sec id="s2-6"><title>Expert Interview</title><p>We conducted semistructured interviews with domain experts (E1-E10); such qualitative feedback is essential to verify whether users benefit from the system support in their specific domain problems [<xref ref-type="bibr" rid="ref27">27</xref>]. Each expert subsequently completed a 10-item standardized questionnaire (<xref ref-type="table" rid="table4">Table 4</xref>) via a 5-point Likert scale to assess their attitudes toward MExplore.</p><table-wrap id="t4" position="float"><label>Table 4.</label><caption><p>Expert interview questionnaire<sup><xref ref-type="table-fn" rid="table4fn1">a</xref></sup>.</p></caption><table id="table4" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Question</td><td align="left" valign="bottom">Question content</td></tr></thead><tbody><tr><td align="left" valign="top">Q1</td><td align="left" valign="top">MExplore is very easy (difficult) to learn.</td></tr><tr><td align="left" valign="top">Q2</td><td align="left" valign="top">MExplore is very easy (difficult) to use.</td></tr><tr><td align="left" valign="top">Q3</td><td align="left" valign="top">The visual design of MExplore is easy (difficult) to understand.</td></tr><tr><td align="left" valign="top">Q4</td><td align="left" valign="top">The visual interactions of MExplore are easy (difficult) to use.</td></tr><tr><td align="left" valign="top">Q5</td><td align="left" valign="top">I am very willing (unwilling) to use MExplore in exploring and acquiring medical expertise.</td></tr><tr><td align="left" valign="top">Q6</td><td align="left" valign="top">Using MExplore, I can (cannot) efficiently identify core knowledge units within complex medical texts. (R1)</td></tr><tr><td align="left" valign="top">Q7</td><td align="left" valign="top">Using MExplore, I can (cannot) explore medical texts in a gradual, structured manner. (R2)</td></tr><tr><td align="left" valign="top">Q8</td><td align="left" valign="top">Using MExplore, I can (cannot) identify and analyze the interconnections between knowledge units. (R3)</td></tr><tr><td align="left" valign="top">Q9</td><td align="left" valign="top">Using MExplore, I can (cannot) focus on and analyze key knowledge units in detail. (R4)</td></tr><tr><td align="left" valign="top">Q10</td><td align="left" valign="top">MExplore can (cannot) help me construct a comprehensive and reliable illness script.</td></tr></tbody></table><table-wrap-foot><fn id="table4fn1"><p><sup>a</sup>Q1-Q5 focus on assessing the system performance of MExplore, Q6-Q9 evaluate whether the key analysis requirements (R1-R4) are satisfied, and Q10 is the overall objective.</p></fn></table-wrap-foot></table-wrap></sec><sec id="s2-7"><title>Ethical Considerations</title><p>The study received ethics approval from the Institutional Review Board of West China Hospital of Stomatology, Sichuan University (ethical approval number: WCHSIRB-D-2024&#x2010;335; approval date: August 9, 2024). Due to its retrospective nature, this study required no informed consent. All collected data were fully anonymized, and no personally identifiable information was included. No financial or material incentives were provided to the participants.</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Case Study</title><sec id="s3-1-1"><title>Case 1: Identification and Exploration of Areas of Interest</title><p>The case study begins with a real scenario with exposed bone in the facial area. A CT scan confirmed the presence of osteonecrosis. The experts entered the keywords <italic>bone</italic>, <italic>expose</italic>, <italic>CT</italic>, and <italic>osteonecrosis</italic> in MExplore. This query generated an MD space where, upon setting the similarity threshold to 0.5, a distinct structure comprising 3 nebulae (A1, A2, and A3) emerged (<xref ref-type="fig" rid="figure6">Figure 6A</xref>).</p><p>The color of nebula A1 corresponded to topic 2 (<italic>Gastrointestinal and Organ Disorders</italic>), whereas A2 and A3 aligned with topic 1 (<italic>Oral and Maxillofacial Surgery</italic>). Consistent with the patient&#x2019;s clinical presentation in the maxillofacial region, the experts disregarded A1 to concentrate their analysis on A2 and A3. A detailed exploration of A3 revealed that it contained MDs describing symptoms highly congruent with the patient&#x2019;s condition. Notably, these MDs demonstrated strong connectivity to nodes in topic 3 (<italic>Pharmacology and Immunology</italic>), which also appeared in the recommendation list, confirming their high relevance to the search query. Consequently, E3 selected these MDs, along with those in A3, and transitioned them to the MP star map. The maximum size of the subgraph was set to 10, which was determined during the experts&#x2019; free exploration and proved effective for segmenting common MDs.</p><p>In the MP star map (<xref ref-type="fig" rid="figure6">Figure 6B</xref>), E5 first examined subgraphs close to <italic>dis</italic> and <italic>sym</italic> to identify the specific conditions. Three subgraphs were selected to review the text. After these MDs were compared with the original MDs (<xref ref-type="fig" rid="figure6">Figure 6C</xref>), the experts concluded that the target disease was medication-related osteonecrosis of the jaw (MRONJ). E5 commented, &#x201C;Compared with standard search engines, MExplore facilitates the discovery of trustworthy MDs. The structured visual representations of knowledge units enable users to identify the target disease through interactive reasoning rather than randomly searching through uncertain sources.&#x201D;</p><fig position="float" id="figure6"><label>Figure 6.</label><caption><p>(A) Topics and distribution of keyword-related MDs, with a connection threshold set to 0.5. (B) Distribution of MP subgraphs within the MDs in A2. (C) The original MD texts of the selected MP subgraphs, with border color annotations indicating their corresponding constellation in the star map. MD: medical document; MP: medical paragraph.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e83785_fig06.png"/></fig></sec><sec id="s3-1-2"><title>Case 2: Association Analysis and Identification of Key Factors</title><p>Upon identifying the target disease, the experts conducted an association analysis to determine the key factors related to the illness script. As defined by Feltovich and Barrows [<xref ref-type="bibr" rid="ref29">29</xref>], illness scripts consist of 3 components: EC, FT, and CQ. In this case, the disease&#x2019;s name suggested a connection to the medication, leading the experts to select subgraphs that distribute bias toward <italic>dru</italic> (<xref ref-type="fig" rid="figure6">Figure 6B1</xref>). Additionally, the subgraphs adjacent to the target disease with a high total number of MEs were selected (<xref ref-type="fig" rid="figure6">Figure 6B2</xref>).</p><p>The selected subgraphs were organized into the association analysis view (<xref ref-type="fig" rid="figure7">Figure 7</xref>). The experts first analyzed the MES (<xref ref-type="fig" rid="figure7">Figure 7A</xref>), which is distinguished by a high proportion of <italic>dru</italic> MEs based on its color encoding. Upon clicking to expand this MES, it was found to contain the <italic>targeted therapy drugs</italic>, <italic>Anlotinib</italic>, which is closely associated with the <italic>liver cancer</italic> ME, indicating the patient had a history of malignancy and related drug treatment. The text in <xref ref-type="fig" rid="figure7">Figure 7D</xref> further corroborated this by containing <italic>lung cancer</italic> and <italic>targeted therapy</italic>, noting specifically that the patient was using <italic>zoledronic acid</italic>. Therefore, E3 concluded that cancer status and related medications were factors associated with EC and FT. Regarding CQ, E3 analyzed the MES characterized by high <italic>bod</italic> and <italic>sym</italic> proportions (<xref ref-type="fig" rid="figure7">Figure 7B and C</xref>), identifying bone exposure and exudate as primary CQ factors. E3 concluded, &#x201C;Performing association analyses on real-case factors can be integrated into clinical reasoning training, enhancing learners&#x2019; clinical analysis skills.&#x201D;</p><fig position="float" id="figure7"><label>Figure 7.</label><caption><p>The association analysis view of the MES and MEs within the selected subgraph facilitates the exploratory analysis of relationships. ME: medical entity; MES: medical entity set.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e83785_fig07.png"/></fig></sec><sec id="s3-1-3"><title>Case 3: In-Depth Analysis and Construction of Illness Scripts</title><p>To construct a comprehensive illness script, the experts began by analyzing the factors presented in case 2. E5 was initially interested in the role of <italic>zoledronic acid</italic> as an EC for the development of MRONJ, so he focused on it and generated a focused sectional view (<xref ref-type="fig" rid="figure8">Figure 8A1</xref>). The relevant MES includes a <italic>mic</italic> ME potentially related to the FT. A review of the corresponding MD (<xref ref-type="fig" rid="figure8">Figure 8A2</xref>) revealed that this ME pertains to <italic>osteoclasts</italic>, detailing the drug&#x2019;s pharmacological action of inhibiting osteoclast growth, which may impair bone repair. The MD further indicates that when bisphosphonates are used in cancer patients, concurrent dental procedures such as tooth extraction may trigger MRONJ (<xref ref-type="fig" rid="figure8">Figure 8A3</xref>).</p><p>Subsequently, E3 examined a medical history MES involving <italic>tooth extraction</italic>, such MESs typically feature a relatively high density of <italic>pro</italic> and <italic>dis</italic> MEs (<xref ref-type="fig" rid="figure8">Figure 8B1</xref>). By examining the MDs of associated chief complaints and other medical history MESs (<xref ref-type="fig" rid="figure8">Figure 8B2 and B3</xref>), multiple co-occurring <italic>pro</italic> MEs were found to mention <italic>osteomyelitis</italic>, which was also referenced in <xref ref-type="fig" rid="figure8">Figure 8A3</xref>, suggesting that osteomyelitis may have developed as a CQ of MRONJ, likely secondary to an underlying infection.</p><p>To obtain a more definitive CQ, E5 performed a focus analysis of MRONJ (<xref ref-type="fig" rid="figure8">Figure 8C1</xref>), reviewing the MDs corresponding to the MESs that have <italic>sym</italic> and <italic>equ</italic> MEs (<xref ref-type="fig" rid="figure8">Figure 8C2 and C3</xref>). The analysis revealed that bone exposure and pus discharge were prominent clinical symptoms, whereas necrotic bone and radiolucent areas observed on panoramic radiographs emerged as distinctive imaging features. These findings can be used as reliable criteria for clinical diagnosis.</p><fig position="float" id="figure8"><label>Figure 8.</label><caption><p>(A) Focused analysis of zoledronic acid. (B) Focused analysis of the medical history of MES, including tooth extraction. (C) Focused analysis of MRONJ. MES: medical entity set; MRONJ: medication-related osteonecrosis of the jaw.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e83785_fig08.png"/></fig><p>This analysis enabled the experts to construct a comprehensive illness script for MRONJ as follows:</p><list list-type="bullet"><list-item><p><bold>EC:</bold> Received treatment with bisphosphonates or related medications; a history of cancer or related systemic diseases; recent dental procedures.</p></list-item><list-item><p><bold>FT:</bold> Medications such as bisphosphonates inhibit osteoclast function, while dental procedures, particularly tooth extractions, can cause trauma, both of which contribute to the onset of MRONJ.</p></list-item><list-item><p><bold>CQ:</bold> Exposed bone, pus drainage, and potential for subsequent infections, such as osteomyelitis.</p></list-item></list><p>As E3 noted, &#x201C;MExplore enables learners to construct illness scripts in a short period, facilitating rapid clinical reasoning and enhancing long-term retention of script knowledge.&#x201D;</p></sec></sec><sec id="s3-2"><title>User Study</title><p><xref ref-type="table" rid="table5">Tables 5</xref> and <xref ref-type="table" rid="table6">6</xref> present the mean accuracy and retention scores, whereas <xref ref-type="table" rid="table7">Table 7</xref> provides the detailed completion time for each disease across various information classification methods. A more granular comparison of accuracy and retention performance is illustrated in <xref ref-type="fig" rid="figure9">Figure 9</xref>.</p><table-wrap id="t5" position="float"><label>Table 5.</label><caption><p>User study accuracy score results.</p></caption><table id="table5" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom"/><td align="left" valign="bottom">Max score</td><td align="left" valign="bottom" colspan="3">Accuracy score</td></tr><tr><td align="left" valign="top"/><td align="char" char="." valign="top"/><td align="char" char="(" valign="top">MEX<sup><xref ref-type="table-fn" rid="table5fn1">a</xref></sup>, mean (SD)</td><td align="char" char="(" valign="top">OTH<sup><xref ref-type="table-fn" rid="table5fn2">b</xref></sup>, mean (SD)</td><td align="char" char="." valign="top"><italic>P</italic> value</td></tr></thead><tbody><tr><td align="left" valign="top">Oral candidiasis</td><td align="char" char="." valign="top"/><td align="char" char="(" valign="top"/><td align="char" char="(" valign="top"/><td align="char" char="." valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Enabling conditions</td><td align="char" char="." valign="top">7</td><td align="char" char="parenthesis" valign="top">6.3 (0.48)</td><td align="char" char="parenthesis" valign="top">5.9 (0.57)</td><td align="char" char="." valign="top">.11</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Fault</td><td align="char" char="." valign="top">7</td><td align="char" char="parenthesis" valign="top">6.1 (0.32)</td><td align="char" char="parenthesis" valign="top">5.7 (0.48)</td><td align="char" char="." valign="top">.04</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Consequences</td><td align="char" char="." valign="top">8</td><td align="char" char="parenthesis" valign="top">7.3 (0.48)</td><td align="char" char="parenthesis" valign="top">7 (0.47)</td><td align="char" char="." valign="top">.18</td></tr><tr><td align="left" valign="top">Meningitis</td><td align="char" char="." valign="top"/><td align="char" char="(" valign="top"/><td align="char" char="(" valign="top"/><td align="char" char="." valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Enabling conditions</td><td align="char" char="." valign="top">7</td><td align="char" char="parenthesis" valign="top">6.1 (0.56)</td><td align="char" char="parenthesis" valign="top">5.8 (0.92)</td><td align="char" char="." valign="top">.39</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Fault</td><td align="char" char="." valign="top">10</td><td align="char" char="parenthesis" valign="top">9.1 (0.32)</td><td align="char" char="parenthesis" valign="top">8.4 (0.52)</td><td align="char" char="." valign="top">.002</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Consequences</td><td align="char" char="." valign="top">10</td><td align="char" char="parenthesis" valign="top">9.3 (0.67)</td><td align="char" char="parenthesis" valign="top">8 (0.47)</td><td align="char" char="." valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Herpes zoster</td><td align="char" char="." valign="top"/><td align="char" char="(" valign="top"/><td align="char" char="(" valign="top"/><td align="char" char="." valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Enabling conditions</td><td align="char" char="." valign="top">9</td><td align="char" char="parenthesis" valign="top">7.9 (0.88)</td><td align="char" char="parenthesis" valign="top">7.1 (0.57)</td><td align="char" char="." valign="top">.03</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Fault</td><td align="char" char="." valign="top">10</td><td align="char" char="parenthesis" valign="top">8.9 (0.57)</td><td align="char" char="parenthesis" valign="top">8.2 (0.42)</td><td align="char" char="." valign="top">.006</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Consequences</td><td align="char" char="." valign="top">8</td><td align="char" char="parenthesis" valign="top">7.4 (0.52)</td><td align="char" char="parenthesis" valign="top">6.9 (0.57)</td><td align="char" char="." valign="top">.05</td></tr></tbody></table><table-wrap-foot><fn id="table5fn1"><p><sup>a</sup>Acquiring expertise via MExplore.</p></fn><fn id="table5fn2"><p><sup>b</sup>Acquiring expertise without MExplore but using other resources.</p></fn></table-wrap-foot></table-wrap><table-wrap id="t6" position="float"><label>Table 6.</label><caption><p>User study retention accuracy score results.</p></caption><table id="table6" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom"/><td align="left" valign="bottom">Max score</td><td align="left" valign="bottom" colspan="3">Retention accuracy score</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top">MEX<sup><xref ref-type="table-fn" rid="table6fn1">a</xref></sup>, mean (SD)</td><td align="left" valign="top">OTH<sup><xref ref-type="table-fn" rid="table6fn2">b</xref></sup>, mean (SD)</td><td align="left" valign="top"><italic>P</italic> value</td></tr></thead><tbody><tr><td align="left" valign="top">Oral candidiasis</td><td align="char" char="." valign="top"/><td align="char" char="(" valign="top"/><td align="char" char="(" valign="top"/><td align="char" char="." valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Enabling conditions</td><td align="char" char="." valign="top">7</td><td align="char" char="parenthesis" valign="top">5.7 (0.67)</td><td align="char" char="parenthesis" valign="top">5 (1.05)</td><td align="char" char="." valign="top">.09</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Fault</td><td align="char" char="." valign="top">7</td><td align="char" char="parenthesis" valign="top">5.6 (0.70)</td><td align="char" char="parenthesis" valign="top">4.9 (0.74)</td><td align="char" char="." valign="top">.04</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Consequences</td><td align="char" char="." valign="top">8</td><td align="char" char="parenthesis" valign="top">6.7 (0.67)</td><td align="char" char="parenthesis" valign="top">5.7 (1.34)</td><td align="char" char="." valign="top">.05</td></tr><tr><td align="left" valign="top">Meningitis</td><td align="char" char="." valign="top"/><td align="char" char="(" valign="top"/><td align="char" char="(" valign="top"/><td align="char" char="." valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Enabling conditions</td><td align="char" char="." valign="top">7</td><td align="char" char="parenthesis" valign="top">5.5 (0.53)</td><td align="char" char="parenthesis" valign="top">5.1 (0.99)</td><td align="char" char="." valign="top">.28</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Fault</td><td align="char" char="." valign="top">10</td><td align="char" char="parenthesis" valign="top">8.4 (0.70)</td><td align="char" char="parenthesis" valign="top">7.1 (0.88)</td><td align="char" char="." valign="top">.002</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Consequences</td><td align="char" char="." valign="top">10</td><td align="char" char="parenthesis" valign="top">8.6 (0.84)</td><td align="char" char="parenthesis" valign="top">7.4 (0.97)</td><td align="char" char="." valign="top">.008</td></tr><tr><td align="left" valign="top">Herpes zoster</td><td align="char" char="." valign="top"/><td align="char" char="(" valign="top"/><td align="char" char="(" valign="top"/><td align="char" char="." valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Enabling conditions</td><td align="char" char="." valign="top">9</td><td align="char" char="parenthesis" valign="top">7.2 (1.03)</td><td align="char" char="parenthesis" valign="top">6.7 (1.06)</td><td align="char" char="." valign="top">.30</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Fault</td><td align="char" char="." valign="top">10</td><td align="char" char="parenthesis" valign="top">8.1 (0.74)</td><td align="char" char="parenthesis" valign="top">6.9 (0.99)</td><td align="char" char="." valign="top">.007</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Consequences</td><td align="char" char="." valign="top">8</td><td align="char" char="parenthesis" valign="top">6.8 (1.03)</td><td align="char" char="parenthesis" valign="top">6 (1.15)</td><td align="char" char="." valign="top">.11</td></tr></tbody></table><table-wrap-foot><fn id="table6fn1"><p><sup>a</sup>Acquiring expertise via MExplore.</p></fn><fn id="table6fn2"><p><sup>b</sup>Acquiring expertise without MExplore but using other resources.</p></fn></table-wrap-foot></table-wrap><table-wrap id="t7" position="float"><label>Table 7.</label><caption><p>User study completion time results.</p></caption><table id="table7" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom"/><td align="left" valign="bottom" colspan="2">Completion time (minutes)</td><td align="left" valign="bottom"><italic>P</italic> value</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">MEX<sup><xref ref-type="table-fn" rid="table7fn1">a</xref></sup>, mean (SD)</td><td align="left" valign="top">OTH<sup><xref ref-type="table-fn" rid="table7fn2">b</xref></sup>, mean (SD)</td><td align="left" valign="top"/></tr></thead><tbody><tr><td align="left" valign="top">Oral candidiasis</td><td align="char" char="parenthesis" valign="top">10.45 (1.41)</td><td align="char" char="parenthesis" valign="top">12.32 (1.88)</td><td align="char" char="." valign="top">.02</td></tr><tr><td align="left" valign="top">Meningitis</td><td align="char" char="parenthesis" valign="top">13.40 (1.73)</td><td align="char" char="parenthesis" valign="top">13.85 (2.41)</td><td align="char" char="." valign="top">.64</td></tr><tr><td align="left" valign="top">Herpes zoster</td><td align="left" valign="top">11.82 (1.76)</td><td align="left" valign="top">15.05 (2.54)</td><td align="left" valign="top">.004</td></tr></tbody></table><table-wrap-foot><fn id="table7fn1"><p><sup>a</sup>Acquiring expertise via MExplore.</p></fn><fn id="table7fn2"><p><sup>b</sup>Acquiring expertise without MExplore but using other resources.</p></fn></table-wrap-foot></table-wrap><p>The results showed that the MEX group demonstrated superior performance. Independent <italic>t</italic> tests confirmed significant improvements in overall mean accuracy (68.3 out of 76 (90%) vs 63.1 out of 76 (83%), <italic>P</italic>&#x003C;.001). This advantage persisted in long-term retention: the MEX group achieved an overall mean retention accuracy of 62.5 out of 76 (82%), significantly surpassing the 54.7 out of 76 (72%) observed in the OTH group (<italic>P</italic>=.003). These findings suggest that the multilevel metaphors and interactive exploration in MEX may facilitate knowledge consolidation into long-term memory. Regarding efficiency, the MEX group had significantly shorter completion times for D1 (<italic>P</italic>=.02) and D3 (<italic>P</italic>=.004). Beyond speed, the MEX group exhibited consistently lower SDs across all tasks (<xref ref-type="table" rid="table7">Table 7</xref>), suggesting that the structured exploration path may reduce user disorientation and cognitive load when navigating complex medical knowledge.</p><fig position="float" id="figure9"><label>Figure 9.</label><caption><p>User study results: comparison of accuracy and retention accuracy. CQ: consequences; D1: oral candidiasis; D2: meningitis; D3: herpes zoster; EC: enabling conditions; FT: fault; MEX: experimental group; OTH: control group.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e83785_fig09.png"/></fig></sec><sec id="s3-3"><title>Expert Interview</title><p>The results of the questionnaire are presented in <xref ref-type="fig" rid="figure10">Figure 10</xref>.</p><fig position="float" id="figure10"><label>Figure 10.</label><caption><p>Results of the questionnaire. The number at the top represents the number of expert responses. The color of the bar represents the degree of negative (1 and 2), positive (4 and 5), or neutral (3) response.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e83785_fig10.png"/></fig><sec id="s3-3-1"><title>System Performance</title><p>The experts all recommended MExplore for its user-friendly interface and interaction capabilities. The system&#x2019;s hierarchical visualization, organized by data granularity and incremental exploration ability, significantly enhances user understanding and analysis. Although some of the experts (E5-E8) had no prior experience with visual analytics systems, they were able to easily comprehend and effectively use the system after receiving minimal training. E4 concurred that the ME-based exploration learning process in MExplore allows learners to focus on the core concepts of the target domain while creating an independent exploration path. E2 suggested incorporating features such as exploration playback and snapshot functionality to further increase the system&#x2019;s usability. In summary, the experts highly praised the system&#x2019;s visual design and interaction approach, asserting that it holds significant potential to improve the efficiency of knowledge acquisition for medical professionals.</p></sec><sec id="s3-3-2"><title>Analysis Requirements</title><p>As shown in <xref ref-type="fig" rid="figure10">Figure 10</xref>, the experts believe that most of the key analytical requirements have been well met. E9 noted that the design of the association analysis view and the focused sectional view enabled users to correlate and concentrate their analysis on key knowledge points, thereby identifying relevant and valuable information&#x2014;an essential aspect of the knowledge acquisition process. E8 noted that compared with search-based learning methods or chatbots, which have become popular in recent years, this approach can exercise the learner&#x2019;s thinking ability and improve knowledge retention and accumulation. E5 further highlighted, &#x201C;In addition to assisting novices in acquiring expertise, MExplore can also support experts and researchers in disease studies by aiding in research, summarization, and the discovery of new features.&#x201D;</p></sec></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Lessons Learned</title><p>The key insight from this study is the importance of comprehensibility. Learners must be able to focus on acquiring knowledge throughout the learning process, and the visualizations and interactions should be designed in a way that does not overwhelm or distract users. Through an iterative design process involving both experts and users, we find a clear preference for simple and intuitive visual forms. Based on this feedback, we developed a multilevel metaphor visual analytics framework to reduce cognitive strain and facilitate a shift in paradigm from traditional learning methods to more interactive forms of knowledge acquisition.</p></sec><sec id="s4-2"><title>Generalizability</title><sec id="s4-2-1"><title>Language Adaptability</title><p>Although MExplore uses a model pretrained on Chinese corpora, the framework is not strictly bound by linguistic characteristics. Transformer-based architectures have demonstrated robust concept understanding across diverse languages when fine-tuned on domain-specific data [<xref ref-type="bibr" rid="ref76">76</xref>]. Consequently, the pipeline can support cross-lingual applications by incorporating domain-specific clinical corpora.</p></sec><sec id="s4-2-2"><title>User Adaptability</title><p>As experts indicated in the Analysis Requirements section, MExplore&#x2019;s extraction and knowledge association capabilities benefit a broad spectrum of users, not limited to novices. It can further assist senior clinicians and researchers in identifying novel disease features and advancing clinical studies.</p></sec><sec id="s4-2-3"><title>Data Adaptability</title><p>To ensure broad generalizability, our retrieval and filtering mechanisms primarily leverage the semantic embedding similarity of unstructured text. However, we recognize that practical medical education often requires precise constraints. Our framework allows users to incorporate a deterministic metadata filtering layer (eg, filtering by visit date, department, and document type) into the search module tailored to their specific datasets. This enables the framework to meet curriculum-specific requirements and adapt to diverse educational goals beyond unstructured content analysis.</p></sec><sec id="s4-2-4"><title>Domain Adaptability</title><p>The methodology can be extended beyond medicine to other domains requiring structured knowledge exploration, such as jurisprudence. By fine-tuning predictive models on annotated domain-specific texts, the MExplore framework can be repurposed to visualize entity relationships and facilitate knowledge acquisition in nonclinical fields.</p></sec></sec><sec id="s4-3"><title>Limitations</title><p>A notable limitation is the linguistic specificity of the current dataset. Clinical documents in different languages exhibit unique syntactic complexities and sublanguages. While our pipeline is architecture-agnostic, performance in non-Chinese contexts depends on the availability of high-quality domain-specific corpora. Future iterations could incorporate neural machine translation and annotation projection to automatically generate training resources for low-resource languages, thereby enhancing multilingual applicability. Additionally, we plan to expand the dataset through multi-institutional collaboration to cover a broader range of diseases. While the system currently supports large-scale queries, further optimizations&#x2014;such as approximate nearest neighbor search and WebGL rendering&#x2014;will be explored to accommodate expanding data volumes. Lastly, the user study was conducted with a cohort of medical students (n=20). While suitable for testing learnability, this demographic may not reflect expert clinical decision-making. Future work will expand evaluations to practicing clinicians to better validate the system&#x2019;s professional utility and generalizability.</p></sec><sec id="s4-4"><title>Conclusions</title><p>In this paper, we introduced MExplore, a visual analysis tool designed to support medical learners in exploring and acquiring medical knowledge. MExplore extracts MEs from large-scale medical texts and constructs a knowledge structure, offering a multilevel metaphor visual analytics framework. This framework includes 4 coordinated views: the MD space view, the MP star map, the association analysis view, and the focused sectional view. Together, these views enable users to tailor their learning paths, fostering deeper comprehension and improved retention. By facilitating the construction of complex, interconnected schemas typical of medicine, MExplore significantly supports learners in acquiring medical expertise.</p><p>We conducted 3 case studies, a user study with real-world datasets, and expert interviews. The results highlight MExplore&#x2019;s effectiveness in facilitating the exploration and analysis of medical texts, significantly enhancing learning efficiency and supporting the acquisition of medical expertise.</p></sec></sec></body><back><notes><sec><title>Funding</title><p>This work was supported in part by the Industry-University-Research Innovation Fund for Chinese Universities (2024XL004), the Sichuan Science and Technology Program (2025ZNSFSC0459), and the Comprehensive Reform Project on Innovation-Oriented Practical Education Empowered by Artificial Intelligence at Sichuan University.</p></sec><sec><title>Data Availability</title><p>The Chinese Biomedical Language Understanding Evaluation (CBLUE) dataset analyzed in this study is publicly available through the Aliyun platform [<xref ref-type="bibr" rid="ref77">77</xref>]. The Medical Text of West China School (Hospital) of Stomatology (MWCSS) dataset, which contains sensitive clinical data, is subject to strict confidentiality agreements and institutional privacy policies and is not publicly available. However, deidentified data may be shared by the authors upon reasonable request, subject to institutional approval and a formal data-use agreement. Both the source code and the fine-tuned MacBERT model weights are publicly accessible in our GitHub repository [<xref ref-type="bibr" rid="ref78">78</xref>].</p></sec></notes><fn-group><fn fn-type="con"><p>XP and CL conceptualized the study. JL and YH contributed to data curation. XP, CL, and YH developed and validated the software. XP drafted the original manuscript. JL and ML reviewed and edited the manuscript. All the authors have read and approved the final manuscript. JL and ML contributed equally as the corresponding authors of this manuscript.</p></fn><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">BERT</term><def><p>Bidirectional Encoder Representations From Transformers</p></def></def-item><def-item><term id="abb2">CBLUE</term><def><p>Chinese Biomedical Language Understanding Evaluation</p></def></def-item><def-item><term id="abb3">CQ</term><def><p>consequences</p></def></def-item><def-item><term id="abb4">EC</term><def><p>enabling conditions</p></def></def-item><def-item><term id="abb5">FT</term><def><p>fault</p></def></def-item><def-item><term id="abb6">LLM</term><def><p>large language model</p></def></def-item><def-item><term id="abb7">LoRA</term><def><p>Low-Rank Adaptation</p></def></def-item><def-item><term id="abb8">MD</term><def><p>medical document</p></def></def-item><def-item><term id="abb9">ME</term><def><p>medical entity</p></def></def-item><def-item><term id="abb10">MES</term><def><p>medical entity set</p></def></def-item><def-item><term id="abb11">MEX</term><def><p>experimental group</p></def></def-item><def-item><term id="abb12">MP</term><def><p>medical paragraph</p></def></def-item><def-item><term id="abb13">MRONJ</term><def><p>medication-related osteonecrosis of the jaw</p></def></def-item><def-item><term id="abb14">MWCSS</term><def><p>Medical Text of West China School (Hospital) of Stomatology</p></def></def-item><def-item><term id="abb15">NER</term><def><p>named entity recognition</p></def></def-item><def-item><term id="abb16">OTH</term><def><p>control group</p></def></def-item><def-item><term id="abb17">QLoRA</term><def><p>Quantized Low-Rank Adaptation</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="web"><article-title>Medical school enrollment reaches a new high</article-title><source>Association of American Medical Colleges</source><year>2025</year><month>01</month><day>9</day><access-date>2026-4-10</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.aamc.org/news/medical-school-enrollment-reaches-new-high">https://www.aamc.org/news/medical-school-enrollment-reaches-new-high</ext-link></comment></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="web"><article-title>China Health Statistics Yearbook 2022</article-title><source>National Health Commission of the People's Republic of China</source><year>2022</year><month>01</month><day>24</day><access-date>2026-4-10</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.nhc.gov.cn/mohwsbwstjxxzx/tjtjnj/202501/8193a8edda0f49df80eb5a8ef5e2547c.shtml">https://www.nhc.gov.cn/mohwsbwstjxxzx/tjtjnj/202501/8193a8edda0f49df80eb5a8ef5e2547c.shtml</ext-link></comment></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Patel</surname><given-names>VL</given-names> </name><name name-style="western"><surname>Kaufman</surname><given-names>DR</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Smelser</surname><given-names>NJ</given-names> </name><name name-style="western"><surname>Baltes</surname><given-names>PB</given-names> </name></person-group><article-title>Medical expertise, cognitive psychology of</article-title><source>International Encyclopedia of the Social &#x0026; Behavioral Sciences</source><year>2001</year><publisher-name>Pergamon</publisher-name><fpage>9515</fpage><lpage>9518</lpage><pub-id pub-id-type="other">978-0-08-043076-8</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Schmidt</surname><given-names>HG</given-names> </name><name name-style="western"><surname>Boshuizen</surname><given-names>HPA</given-names> </name></person-group><article-title>On acquiring expertise in medicine</article-title><source>Educ Psychol Rev</source><year>1993</year><month>09</month><volume>5</volume><issue>3</issue><fpage>205</fpage><lpage>221</lpage><pub-id pub-id-type="doi">10.1007/BF01323044</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Charlin</surname><given-names>B</given-names> </name><name name-style="western"><surname>Tardif</surname><given-names>J</given-names> </name><name name-style="western"><surname>Boshuizen</surname><given-names>HP</given-names> </name></person-group><article-title>Scripts and medical diagnostic knowledge: theory and applications for clinical reasoning instruction and research</article-title><source>Acad Med</source><year>2000</year><month>02</month><volume>75</volume><issue>2</issue><fpage>182</fpage><lpage>190</lpage><pub-id pub-id-type="doi">10.1097/00001888-200002000-00020</pub-id><pub-id pub-id-type="medline">10693854</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Xiao</surname><given-names>H</given-names> </name><name name-style="western"><surname>Zhou</surname><given-names>F</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>X</given-names> </name><etal/></person-group><article-title>A comprehensive survey of large language models and multimodal large language models in medicine</article-title><source>Inf Fusion</source><year>2025</year><month>05</month><volume>117</volume><fpage>102888</fpage><pub-id pub-id-type="doi">10.1016/j.inffus.2024.102888</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Griot</surname><given-names>M</given-names> </name><name name-style="western"><surname>Hemptinne</surname><given-names>C</given-names> </name><name name-style="western"><surname>Vanderdonckt</surname><given-names>J</given-names> </name><name name-style="western"><surname>Yuksel</surname><given-names>D</given-names> </name></person-group><article-title>Large language models lack essential metacognition for reliable medical reasoning</article-title><source>Nat Commun</source><year>2025</year><month>01</month><day>14</day><volume>16</volume><issue>1</issue><fpage>642</fpage><pub-id pub-id-type="doi">10.1038/s41467-024-55628-6</pub-id><pub-id pub-id-type="medline">39809759</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Kim</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Jeong</surname><given-names>H</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Medical hallucination in foundation models and their impact on healthcare</article-title><source>arXiv</source><comment>Preprint posted online on  Feb 26, 2025</comment><pub-id pub-id-type="doi">10.48550/arXiv.2503.05777</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Safranek</surname><given-names>CW</given-names> </name><name name-style="western"><surname>Sidamon-Eristoff</surname><given-names>AE</given-names> </name><name name-style="western"><surname>Gilson</surname><given-names>A</given-names> </name><name name-style="western"><surname>Chartash</surname><given-names>D</given-names> </name></person-group><article-title>The role of large language models in medical education: applications and implications</article-title><source>JMIR Med Educ</source><year>2023</year><volume>9</volume><fpage>e50945</fpage><pub-id pub-id-type="doi">10.2196/50945</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Dergaa</surname><given-names>I</given-names> </name><name name-style="western"><surname>Ben Saad</surname><given-names>H</given-names> </name><name name-style="western"><surname>Glenn</surname><given-names>JM</given-names> </name><etal/></person-group><article-title>From tools to threats: a reflection on the impact of artificial-intelligence chatbots on cognitive health</article-title><source>Front Psychol</source><year>2024</year><volume>15</volume><fpage>1259845</fpage><pub-id pub-id-type="doi">10.3389/fpsyg.2024.1259845</pub-id><pub-id pub-id-type="medline">38629037</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Stadler</surname><given-names>M</given-names> </name><name name-style="western"><surname>Bannert</surname><given-names>M</given-names> </name><name name-style="western"><surname>Sailer</surname><given-names>M</given-names> </name></person-group><article-title>Cognitive ease at a cost: LLMs reduce mental effort but compromise depth in student scientific inquiry</article-title><source>Comput Human Behav</source><year>2024</year><month>11</month><volume>160</volume><fpage>108386</fpage><pub-id pub-id-type="doi">10.1016/j.chb.2024.108386</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Lee</surname><given-names>HP</given-names> </name><name name-style="western"><surname>Sarkar</surname><given-names>A</given-names> </name><name name-style="western"><surname>Tankelevitch</surname><given-names>L</given-names> </name><etal/></person-group><article-title>The impact of generative AI on critical thinking: self-reported reductions in cognitive effort and confidence effects from a survey of knowledge workers</article-title><source>CHI &#x2019;25: Proceedings of the 2025 CHI Conference on Human Factors in Computing Systems</source><year>2025</year><publisher-name>Association for Computing Machinery</publisher-name><pub-id pub-id-type="doi">10.1145/3706598.3713778</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Roustan</surname><given-names>D</given-names> </name><name name-style="western"><surname>Bastardot</surname><given-names>F</given-names> </name></person-group><article-title>The clinicians&#x2019; guide to large language models: a general perspective with a focus on hallucinations</article-title><source>Interact J Med Res</source><year>2025</year><month>01</month><day>28</day><volume>14</volume><issue>1</issue><fpage>e59823</fpage><pub-id pub-id-type="doi">10.2196/59823</pub-id><pub-id pub-id-type="medline">39874574</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Jin</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Cui</surname><given-names>S</given-names> </name><name name-style="western"><surname>Guo</surname><given-names>S</given-names> </name><name name-style="western"><surname>Gotz</surname><given-names>D</given-names> </name><name name-style="western"><surname>Sun</surname><given-names>J</given-names> </name><name name-style="western"><surname>Cao</surname><given-names>N</given-names> </name></person-group><article-title>CarePre: an intelligent clinical decision assistance system</article-title><source>ACM Trans Comput Healthc</source><year>2020</year><volume>1</volume><issue>1</issue><fpage>1</fpage><lpage>20</lpage><pub-id pub-id-type="doi">10.1145/3344258</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Xu</surname><given-names>T</given-names> </name><name name-style="western"><surname>Ma</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Pan</surname><given-names>T</given-names> </name><etal/></person-group><article-title>Visual analytics of multidimensional oral health surveys: data mining study</article-title><source>JMIR Med Inform</source><year>2023</year><month>08</month><day>1</day><volume>11</volume><issue>1</issue><fpage>e46275</fpage><pub-id pub-id-type="doi">10.2196/46275</pub-id><pub-id pub-id-type="medline">37526971</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Permana</surname><given-names>B</given-names> </name><name name-style="western"><surname>Harris</surname><given-names>PNA</given-names> </name><name name-style="western"><surname>Roberts</surname><given-names>LW</given-names> </name><etal/></person-group><article-title>HAIviz: an interactive dashboard for visualising and integrating healthcare-associated genomic epidemiological data</article-title><source>Microb Genom</source><year>2024</year><month>02</month><volume>10</volume><issue>2</issue><fpage>001200</fpage><pub-id pub-id-type="doi">10.1099/mgen.0.001200</pub-id><pub-id pub-id-type="medline">38358326</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Siirtola</surname><given-names>H</given-names> </name><name name-style="western"><surname>Gracia-Tabuenca</surname><given-names>J</given-names> </name><name name-style="western"><surname>Raisamo</surname><given-names>R</given-names> </name><name name-style="western"><surname>Niemi</surname><given-names>M</given-names> </name><name name-style="western"><surname>Reeve</surname><given-names>MP</given-names> </name><name name-style="western"><surname>Laitinen</surname><given-names>T</given-names> </name></person-group><article-title>Glyph-based visualization of health trajectories</article-title><source>2022 26th International Conference on Information Visualisation</source><year>2022</year><publisher-name>IEEE</publisher-name><pub-id pub-id-type="doi">10.1109/IV56949.2022.00075</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Grishman</surname><given-names>R</given-names> </name><name name-style="western"><surname>Sundheim</surname><given-names>B</given-names> </name></person-group><article-title>Message understanding conference-6: a brief history</article-title><source>Proceedings of the 16th Conference on Computational Linguistics</source><year>1996</year><publisher-name>Association for Computational Linguistics</publisher-name><pub-id pub-id-type="doi">10.3115/992628.992709</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mehmood</surname><given-names>T</given-names> </name><name name-style="western"><surname>Serina</surname><given-names>I</given-names> </name><name name-style="western"><surname>Lavelli</surname><given-names>A</given-names> </name><name name-style="western"><surname>Putelli</surname><given-names>L</given-names> </name><name name-style="western"><surname>Gerevini</surname><given-names>A</given-names> </name></person-group><article-title>On the use of knowledge transfer techniques for biomedical named entity recognition</article-title><source>Future Internet</source><year>2023</year><volume>15</volume><issue>2</issue><fpage>79</fpage><pub-id pub-id-type="doi">10.3390/fi15020079</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Huang</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>W</given-names> </name><name name-style="western"><surname>Yu</surname><given-names>K</given-names> </name></person-group><article-title>Bidirectional LSTM-CRF models for sequence tagging</article-title><source>arXiv</source><comment>Preprint posted online on  Aug 9, 2015</comment><pub-id pub-id-type="doi">10.48550/arXiv.1508.01991</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Lample</surname><given-names>G</given-names> </name><name name-style="western"><surname>Ballesteros</surname><given-names>M</given-names> </name><name name-style="western"><surname>Subramanian</surname><given-names>S</given-names> </name><name name-style="western"><surname>Kawakami</surname><given-names>K</given-names> </name><name name-style="western"><surname>Dyer</surname><given-names>C</given-names> </name></person-group><article-title>Neural architectures for named entity recognition</article-title><source>arXiv</source><comment>Preprint posted online on  Mar 4, 2016</comment><pub-id pub-id-type="doi">10.48550/arXiv.1603.01360</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Devlin</surname><given-names>J</given-names> </name><name name-style="western"><surname>Chang</surname><given-names>MW</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>K</given-names> </name><name name-style="western"><surname>Toutanova</surname><given-names>K</given-names> </name></person-group><article-title>BERT: pre-training of deep bidirectional transformers for language understanding</article-title><source>Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source><year>2019</year><publisher-name>Association for Computational Linguistics</publisher-name><pub-id pub-id-type="doi">10.18653/v1/N19-1423</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Vaswani</surname><given-names>A</given-names> </name><name name-style="western"><surname>Shazeer</surname><given-names>N</given-names> </name><name name-style="western"><surname>Parmar</surname><given-names>N</given-names> </name><etal/></person-group><article-title>Attention is all you need</article-title><source>Proceedings of the 31st International Conference on Neural Information Processing Systems</source><year>2017</year><publisher-name>Curran Associates Inc</publisher-name><pub-id pub-id-type="doi">10.5555/3295222.3295349</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fall</surname><given-names>LH</given-names> </name><name name-style="western"><surname>English</surname><given-names>R</given-names> </name><name name-style="western"><surname>Fulton</surname><given-names>TB</given-names> </name><etal/></person-group><article-title>Thinking slow more quickly: development of integrated illness scripts to support cognitively integrated learning and improve clinical decision-making</article-title><source>Med Sci Educ</source><year>2021</year><month>06</month><volume>31</volume><issue>3</issue><fpage>1005</fpage><lpage>1007</lpage><pub-id pub-id-type="doi">10.1007/s40670-021-01293-z</pub-id><pub-id pub-id-type="medline">34457943</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Cloude</surname><given-names>EB</given-names> </name><name name-style="western"><surname>Ballelos</surname><given-names>NAM</given-names> </name><name name-style="western"><surname>Azevedo</surname><given-names>R</given-names> </name><etal/></person-group><article-title>Designing intelligent systems to support medical diagnostic reasoning using process data</article-title><source>Artificial Intelligence in Education</source><year>2021</year><publisher-name>Springer</publisher-name><pub-id pub-id-type="doi">10.1007/978-3-030-78270-2_19</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Likert</surname><given-names>R</given-names> </name></person-group><article-title>A technique for the measurement of attitudes</article-title><source>Arch Psychol</source><year>1932</year><access-date>2026-04-10</access-date><volume>22</volume><issue>140</issue><fpage>1</fpage><lpage>55</lpage><comment><ext-link ext-link-type="uri" xlink:href="https://psycnet.apa.org/record/1933-01885-001">https://psycnet.apa.org/record/1933-01885-001</ext-link></comment></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Munzner</surname><given-names>T</given-names> </name></person-group><article-title>A nested model for visualization design and validation</article-title><source>IEEE Trans Vis Comput Graph</source><year>2009</year><volume>15</volume><issue>6</issue><fpage>921</fpage><lpage>928</lpage><pub-id pub-id-type="doi">10.1109/TVCG.2009.111</pub-id><pub-id pub-id-type="medline">19834155</pub-id></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Custers</surname><given-names>EJFM</given-names> </name></person-group><article-title>Thirty years of illness scripts: theoretical origins and practical applications</article-title><source>Med Teach</source><year>2015</year><month>05</month><volume>37</volume><issue>5</issue><fpage>457</fpage><lpage>462</lpage><pub-id pub-id-type="doi">10.3109/0142159X.2014.956052</pub-id><pub-id pub-id-type="medline">25180878</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Feltovich</surname><given-names>PJ</given-names> </name><name name-style="western"><surname>Barrows</surname><given-names>HS</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Schmidt</surname><given-names>HG</given-names> </name><name name-style="western"><surname>Volder</surname><given-names>ML</given-names> </name></person-group><article-title>Issues of generality in medical problem solving</article-title><source>Tutorials in Problem-Based Learning: New Directions in Training for the Health Professions</source><year>1984</year><publisher-name>Van Gorcum</publisher-name><fpage>128</fpage><lpage>142</lpage><pub-id pub-id-type="other">9023220641</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Bagheri</surname><given-names>A</given-names> </name><name name-style="western"><surname>Giachanou</surname><given-names>A</given-names> </name><name name-style="western"><surname>Mosteiro</surname><given-names>P</given-names> </name><name name-style="western"><surname>Verberne</surname><given-names>S</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Denaxas</surname><given-names>S</given-names> </name><name name-style="western"><surname>Oberski</surname><given-names>D</given-names> </name><name name-style="western"><surname>Moore</surname><given-names>J</given-names> </name></person-group><article-title>Natural language processing and text mining (turning unstructured data into structured)</article-title><source>Clinical Applications of Artificial Intelligence in Real-World Data</source><year>2023</year><publisher-name>Springer</publisher-name><fpage>69</fpage><lpage>93</lpage><pub-id pub-id-type="other">978-3-031-36678-9_5</pub-id></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Perfetti</surname><given-names>CA</given-names> </name></person-group><article-title>Cognitive research can inform reading education</article-title><source>J Res Read</source><year>1995</year><month>09</month><volume>18</volume><issue>2</issue><fpage>106</fpage><lpage>115</lpage><pub-id pub-id-type="doi">10.1111/j.1467-9817.1995.tb00076.x</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sweller</surname><given-names>J</given-names> </name></person-group><article-title>Cognitive load during problem solving: effects on learning</article-title><source>Cogn Sci</source><year>1988</year><month>04</month><volume>12</volume><issue>2</issue><fpage>257</fpage><lpage>285</lpage><pub-id pub-id-type="doi">10.1207/s15516709cog1202_4</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="web"><article-title>Cognitive load theory: a guide to applying cognitive load theory to your teaching</article-title><source>Office of Educational Improvement, Medical College of Wisconsin</source><year>2022</year><month>05</month><access-date>2026-4-10</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.mcw.edu/-/media/MCW/Education/Academic-Affairs/OEI/Faculty-Quick-Guides/Cognitive-Load-Theory.pdf">https://www.mcw.edu/-/media/MCW/Education/Academic-Affairs/OEI/Faculty-Quick-Guides/Cognitive-Load-Theory.pdf</ext-link></comment></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Natesan</surname><given-names>S</given-names> </name><name name-style="western"><surname>Bailitz</surname><given-names>J</given-names> </name><name name-style="western"><surname>King</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Clinical teaching: an evidence-based guide to best practices from the council of emergency medicine residency directors</article-title><source>West J Emerg Med</source><year>2020</year><month>07</month><day>3</day><volume>21</volume><issue>4</issue><fpage>985</fpage><lpage>998</lpage><pub-id pub-id-type="doi">10.5811/westjem.2020.4.46060</pub-id><pub-id pub-id-type="medline">32726274</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ganascia</surname><given-names>JG</given-names> </name></person-group><article-title>Abstraction of levels of abstraction</article-title><source>J Exp Theor Artif Intell</source><year>2015</year><month>01</month><day>2</day><volume>27</volume><issue>1</issue><fpage>23</fpage><lpage>35</lpage><pub-id pub-id-type="doi">10.1080/0952813X.2014.940685</pub-id></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Qiao</surname><given-names>YQ</given-names> </name><name name-style="western"><surname>Shen</surname><given-names>J</given-names> </name><name name-style="western"><surname>Liang</surname><given-names>X</given-names> </name><etal/></person-group><article-title>Using cognitive theory to facilitate medical education</article-title><source>BMC Med Educ</source><year>2014</year><month>04</month><day>14</day><volume>14</volume><issue>1</issue><fpage>79</fpage><pub-id pub-id-type="doi">10.1186/1472-6920-14-79</pub-id><pub-id pub-id-type="medline">24731433</pub-id></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lujan</surname><given-names>HL</given-names> </name><name name-style="western"><surname>DiCarlo</surname><given-names>SE</given-names> </name></person-group><article-title>The paradox of knowledge: why medical students know more but understand less</article-title><source>Med Sci Educ</source><year>2025</year><month>06</month><volume>35</volume><issue>3</issue><fpage>1761</fpage><lpage>1766</lpage><pub-id pub-id-type="doi">10.1007/s40670-025-02379-8</pub-id><pub-id pub-id-type="medline">40626000</pub-id></nlm-citation></ref><ref id="ref38"><label>38</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mayer</surname><given-names>A</given-names> </name><name name-style="western"><surname>Hege</surname><given-names>I</given-names> </name><name name-style="western"><surname>Kononowicz</surname><given-names>AA</given-names> </name><name name-style="western"><surname>M&#x00FC;ller</surname><given-names>A</given-names> </name><name name-style="western"><surname>Sudacka</surname><given-names>M</given-names> </name></person-group><article-title>Collaborative development of feedback concept maps for virtual patient-based clinical reasoning education: mixed methods study</article-title><source>JMIR Med Educ</source><year>2025</year><month>01</month><day>30</day><volume>11</volume><fpage>e57331</fpage><pub-id pub-id-type="doi">10.2196/57331</pub-id><pub-id pub-id-type="medline">39883032</pub-id></nlm-citation></ref><ref id="ref39"><label>39</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Etukakpan</surname><given-names>AU</given-names> </name><name name-style="western"><surname>Waldhuber</surname><given-names>MG</given-names> </name><name name-style="western"><surname>Janke</surname><given-names>KK</given-names> </name><name name-style="western"><surname>Netere</surname><given-names>AK</given-names> </name><name name-style="western"><surname>Angelo</surname><given-names>T</given-names> </name><name name-style="western"><surname>White</surname><given-names>PJ</given-names> </name></person-group><article-title>Core concept identification in STEM and related domain education: a scoping review of rationales, methods, and outputs</article-title><source>Front Educ</source><year>2025</year><volume>10</volume><fpage>1547994</fpage><pub-id pub-id-type="doi">10.3389/feduc.2025.1547994</pub-id></nlm-citation></ref><ref id="ref40"><label>40</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Medwell</surname><given-names>J</given-names> </name><name name-style="western"><surname>Wray</surname><given-names>D</given-names> </name></person-group><article-title>CONCEPT-based teaching and learning: a review of the research literature</article-title><source>13th Annual International Conference of Education, Research and Innovation</source><year>2020</year><publisher-name>IATED</publisher-name><pub-id pub-id-type="doi">10.21125/iceri.2020.0144</pub-id></nlm-citation></ref><ref id="ref41"><label>41</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>N</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>M</given-names> </name><name name-style="western"><surname>Bi</surname><given-names>Z</given-names> </name><etal/></person-group><article-title>CBLUE: a Chinese biomedical language understanding evaluation benchmark</article-title><source>Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics</source><year>2022</year><publisher-name>Association for Computational Linguistics</publisher-name><pub-id pub-id-type="doi">10.18653/v1/2022.acl-long.544</pub-id></nlm-citation></ref><ref id="ref42"><label>42</label><nlm-citation citation-type="web"><article-title>FlagEval</article-title><source>Beijing Academy of Artificial Intelligence</source><access-date>2026-4-10</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://flageval.baai.ac.cn/">https://flageval.baai.ac.cn/</ext-link></comment></nlm-citation></ref><ref id="ref43"><label>43</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Cui</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Che</surname><given-names>W</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>T</given-names> </name><name name-style="western"><surname>Qin</surname><given-names>B</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>S</given-names> </name><name name-style="western"><surname>Hu</surname><given-names>G</given-names> </name></person-group><article-title>Revisiting pre-trained models for Chinese natural language processing</article-title><source>arXiv</source><comment>Preprint posted online on  Apr 29, 2020</comment><pub-id pub-id-type="doi">10.48550/arXiv.2004.13922</pub-id></nlm-citation></ref><ref id="ref44"><label>44</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Dettmers</surname><given-names>T</given-names> </name><name name-style="western"><surname>Holtzman</surname><given-names>A</given-names> </name><name name-style="western"><surname>Pagnoni</surname><given-names>A</given-names> </name><name name-style="western"><surname>Zettlemoyer</surname><given-names>L</given-names> </name></person-group><article-title>QLORA: efficient finetuning of quantized LLMs</article-title><source>Proceedings of the 37th International Conference on Neural Information Processing Systems</source><year>2023</year><publisher-name>Curran Associates Inc</publisher-name><pub-id pub-id-type="doi">10.5555/3666122.3666563</pub-id></nlm-citation></ref><ref id="ref45"><label>45</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Hu</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Zuo</surname><given-names>X</given-names> </name><name name-style="western"><surname>Zhou</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>Information extraction from clinical notes: are we ready to switch to large language models?</article-title><source>arXiv</source><comment>Preprint posted online on  Nov 15, 2024</comment><pub-id pub-id-type="doi">10.48550/arXiv.2411.10020</pub-id></nlm-citation></ref><ref id="ref46"><label>46</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Keraghel</surname><given-names>I</given-names> </name><name name-style="western"><surname>Morbieu</surname><given-names>S</given-names> </name><name name-style="western"><surname>Nadif</surname><given-names>M</given-names> </name></person-group><article-title>A survey on recent advances in named entity recognition</article-title><source>arXiv</source><comment>Preprint posted online on  Dec 20, 2024</comment><pub-id pub-id-type="doi">10.13140/RG.2.2.35314.00965</pub-id></nlm-citation></ref><ref id="ref47"><label>47</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Mohan</surname><given-names>GB</given-names> </name><name name-style="western"><surname>Kumar</surname><given-names>RP</given-names> </name><name name-style="western"><surname>Elakkiya</surname><given-names>R</given-names> </name><name name-style="western"><surname>Pendem</surname><given-names>M</given-names> </name></person-group><article-title>Fine-tuned BERT based multilingual model for named entity recognition in native Indian languages</article-title><source>2023 International Conference on Evolutionary Algorithms and Soft Computing Techniques (EASCT)</source><year>2023</year><publisher-name>IEEE</publisher-name><pub-id pub-id-type="doi">10.1109/EASCT59475.2023.10392937</pub-id></nlm-citation></ref><ref id="ref48"><label>48</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Anas</surname><given-names>M</given-names> </name><name name-style="western"><surname>Saiyeda</surname><given-names>A</given-names> </name><name name-style="western"><surname>Sohail</surname><given-names>SS</given-names> </name><name name-style="western"><surname>Cambria</surname><given-names>E</given-names> </name><name name-style="western"><surname>Hussain</surname><given-names>A</given-names> </name></person-group><article-title>Can generative AI models extract deeper sentiments as compared to traditional deep learning algorithms?</article-title><source>IEEE Intell Syst</source><year>2024</year><volume>39</volume><issue>2</issue><fpage>5</fpage><lpage>10</lpage><pub-id pub-id-type="doi">10.1109/MIS.2024.3374582</pub-id></nlm-citation></ref><ref id="ref49"><label>49</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Shyr</surname><given-names>C</given-names> </name><name name-style="western"><surname>Hu</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Bastarache</surname><given-names>L</given-names> </name><etal/></person-group><article-title>Identifying and extracting rare diseases and their phenotypes with large language models</article-title><source>J Healthc Inform Res</source><year>2024</year><month>06</month><volume>8</volume><issue>2</issue><fpage>438</fpage><lpage>461</lpage><pub-id pub-id-type="doi">10.1007/s41666-023-00155-0</pub-id><pub-id pub-id-type="medline">38681753</pub-id></nlm-citation></ref><ref id="ref50"><label>50</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kim</surname><given-names>K</given-names> </name><name name-style="western"><surname>Park</surname><given-names>S</given-names> </name><name name-style="western"><surname>Min</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Multifaceted natural language processing task-based evaluation of bidirectional encoder representations from transformers models for bilingual (Korean and English) clinical notes: algorithm development and validation</article-title><source>JMIR Med Inform</source><year>2024</year><month>10</month><day>30</day><volume>12</volume><fpage>e52897</fpage><pub-id pub-id-type="doi">10.2196/52897</pub-id><pub-id pub-id-type="medline">39475725</pub-id></nlm-citation></ref><ref id="ref51"><label>51</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lee</surname><given-names>J</given-names> </name><name name-style="western"><surname>Yoon</surname><given-names>W</given-names> </name><name name-style="western"><surname>Kim</surname><given-names>S</given-names> </name><etal/></person-group><article-title>BioBERT: a pre-trained biomedical language representation model for biomedical text mining</article-title><source>Bioinformatics</source><year>2020</year><month>02</month><day>15</day><volume>36</volume><issue>4</issue><fpage>1234</fpage><lpage>1240</lpage><pub-id pub-id-type="doi">10.1093/bioinformatics/btz682</pub-id><pub-id pub-id-type="medline">31501885</pub-id></nlm-citation></ref><ref id="ref52"><label>52</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Wu</surname><given-names>X</given-names> </name><name name-style="western"><surname>Gao</surname><given-names>C</given-names> </name><name name-style="western"><surname>Zang</surname><given-names>L</given-names> </name><name name-style="western"><surname>Han</surname><given-names>J</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Hu</surname><given-names>S</given-names> </name></person-group><article-title>ESimCSE: enhanced sample building method for contrastive learning of unsupervised sentence embedding</article-title><source>arXiv</source><comment>Preprint posted online on  Sep 9, 2021</comment><pub-id pub-id-type="doi">10.48550/arXiv.2109.04380</pub-id></nlm-citation></ref><ref id="ref53"><label>53</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Dieng</surname><given-names>AB</given-names> </name><name name-style="western"><surname>Ruiz</surname><given-names>FJR</given-names> </name><name name-style="western"><surname>Blei</surname><given-names>DM</given-names> </name></person-group><article-title>Topic modeling in embedding spaces</article-title><source>Trans Assoc Comput Linguist</source><year>2020</year><month>12</month><volume>8</volume><fpage>439</fpage><lpage>453</lpage><pub-id pub-id-type="doi">10.1162/tacl_a_00325</pub-id></nlm-citation></ref><ref id="ref54"><label>54</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ajinaja</surname><given-names>MO</given-names> </name><name name-style="western"><surname>Fakoya</surname><given-names>JT</given-names> </name><name name-style="western"><surname>Ogunwale</surname><given-names>YE</given-names> </name><etal/></person-group><article-title>A comparative evaluation of probabilistic and transformer-based topic models across diverse and multilingual text corpora</article-title><source>Neural Process Lett</source><year>2026</year><volume>58</volume><issue>1</issue><fpage>9</fpage><pub-id pub-id-type="doi">10.1007/s11063-025-11820-3</pub-id></nlm-citation></ref><ref id="ref55"><label>55</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Yi</surname><given-names>X</given-names> </name><name name-style="western"><surname>Guo</surname><given-names>R</given-names> </name><etal/></person-group><article-title>Milvus: a purpose-built vector data management system</article-title><source>Proceedings of the 2021 International Conference on Management of Data</source><year>2021</year><publisher-name>Association for Computing Machinery</publisher-name><pub-id pub-id-type="doi">10.1145/3448016.3457550</pub-id></nlm-citation></ref><ref id="ref56"><label>56</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Hodler</surname><given-names>AE</given-names> </name><name name-style="western"><surname>Needham</surname><given-names>M</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Bader</surname><given-names>DA</given-names> </name></person-group><article-title>Graph data science using Neo4j</article-title><source>Massive Graph Analytics</source><year>2022</year><publisher-name>Chapman and Hall/CRC</publisher-name><fpage>433</fpage><lpage>457</lpage><pub-id pub-id-type="other">9781003033707-20</pub-id></nlm-citation></ref><ref id="ref57"><label>57</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Robinson</surname><given-names>I</given-names> </name><name name-style="western"><surname>Webber</surname><given-names>J</given-names> </name><name name-style="western"><surname>Eifrem</surname><given-names>E</given-names> </name></person-group><source>Graph Databases: New Opportunities for Connected Data</source><year>2015</year><publisher-name>O&#x2019;Reilly Media</publisher-name><pub-id pub-id-type="other">1491930896</pub-id></nlm-citation></ref><ref id="ref58"><label>58</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Roberts</surname><given-names>JC</given-names> </name></person-group><article-title>State of the art: coordinated &#x0026; multiple views in exploratory visualization</article-title><source>Proceedings of the Fifth International Conference on Coordinated and Multiple Views in Exploratory Visualization</source><year>2007</year><publisher-name>IEEE Computer Society</publisher-name><pub-id pub-id-type="doi">10.1109/CMV.2007.20</pub-id></nlm-citation></ref><ref id="ref59"><label>59</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Carroll</surname><given-names>JM</given-names> </name><name name-style="western"><surname>Mack</surname><given-names>RL</given-names> </name><name name-style="western"><surname>Kellogg</surname><given-names>WA</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Helander</surname><given-names>M</given-names> </name></person-group><article-title>Chapter 3: Interface metaphors and user interface design</article-title><source>Handbook of Human-Computer Interaction</source><year>1988</year><publisher-name>North-Holland</publisher-name><fpage>67</fpage><lpage>85</lpage><pub-id pub-id-type="other">978-0-444-70536-5</pub-id></nlm-citation></ref><ref id="ref60"><label>60</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Shneiderman</surname><given-names>B</given-names> </name><name name-style="western"><surname>Plaisant</surname><given-names>C</given-names> </name><name name-style="western"><surname>Cohen</surname><given-names>M</given-names> </name><name name-style="western"><surname>Jacobs</surname><given-names>S</given-names> </name><name name-style="western"><surname>Elmqvist</surname><given-names>N</given-names> </name><name name-style="western"><surname>Diakopoulos</surname><given-names>N</given-names> </name></person-group><source>Designing the User Interface: Strategies for Effective Human-Computer Interaction</source><year>2016</year><publisher-name>Pearson</publisher-name><pub-id pub-id-type="other">013438038X</pub-id></nlm-citation></ref><ref id="ref61"><label>61</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Ware</surname><given-names>C</given-names> </name></person-group><source>Information Visualization: Perception for Design</source><year>2020</year><publisher-name>Morgan Kaufmann</publisher-name><pub-id pub-id-type="other">9780128128756</pub-id></nlm-citation></ref><ref id="ref62"><label>62</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Sanders</surname><given-names>P</given-names> </name><name name-style="western"><surname>Schulz</surname><given-names>C</given-names> </name></person-group><article-title>Engineering multilevel graph partitioning algorithms</article-title><source>Proceedings of the 19th European Conference on Algorithms</source><year>2011</year><publisher-name>Springer</publisher-name><pub-id pub-id-type="doi">10.5555/2040572.2040624</pub-id></nlm-citation></ref><ref id="ref63"><label>63</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>L</given-names> </name><name name-style="western"><surname>Giesen</surname><given-names>J</given-names> </name><name name-style="western"><surname>McDonnell</surname><given-names>KT</given-names> </name><name name-style="western"><surname>Zolliker</surname><given-names>P</given-names> </name><name name-style="western"><surname>Mueller</surname><given-names>K</given-names> </name></person-group><article-title>Color design for illustrative visualization</article-title><source>IEEE Trans Vis Comput Graph</source><year>2008</year><volume>14</volume><issue>6</issue><fpage>1739</fpage><lpage>1746</lpage><pub-id pub-id-type="doi">10.1109/TVCG.2008.118</pub-id><pub-id pub-id-type="medline">18989033</pub-id></nlm-citation></ref><ref id="ref64"><label>64</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Huang</surname><given-names>L</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>M</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>F</given-names> </name></person-group><article-title>Syllabus design for teacher education MOOCs (massive open online courses): a mixed methods approach</article-title><source>Technology in Education: Pedagogical Innovations</source><year>2019</year><publisher-name>Springer</publisher-name><pub-id pub-id-type="doi">10.1007/978-981-13-9895-7_14</pub-id></nlm-citation></ref><ref id="ref65"><label>65</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhou</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Cai</surname><given-names>L</given-names> </name><name name-style="western"><surname>Guo</surname><given-names>J</given-names> </name><etal/></person-group><article-title>ExeVis: concept-based visualization of exercises in online learning</article-title><source>J Vis</source><year>2024</year><month>04</month><volume>27</volume><issue>2</issue><fpage>235</fpage><lpage>254</lpage><pub-id pub-id-type="doi">10.1007/s12650-024-00956-4</pub-id></nlm-citation></ref><ref id="ref66"><label>66</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sweller</surname><given-names>J</given-names> </name></person-group><article-title>Cognitive load theory, learning difficulty, and instructional design</article-title><source>Learn Instr</source><year>1994</year><month>01</month><volume>4</volume><issue>4</issue><fpage>295</fpage><lpage>312</lpage><pub-id pub-id-type="doi">10.1016/0959-4752(94)90003-5</pub-id></nlm-citation></ref><ref id="ref67"><label>67</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fonseca</surname><given-names>M</given-names> </name><name name-style="western"><surname>Broeiro-Gon&#x00E7;alves</surname><given-names>P</given-names> </name><name name-style="western"><surname>Barosa</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Concept mapping to promote clinical reasoning in multimorbidity: a mixed methods study in undergraduate family medicine</article-title><source>BMC Med Educ</source><year>2024</year><month>12</month><day>18</day><volume>24</volume><issue>1</issue><fpage>1478</fpage><pub-id pub-id-type="doi">10.1186/s12909-024-06484-x</pub-id><pub-id pub-id-type="medline">39695556</pub-id></nlm-citation></ref><ref id="ref68"><label>68</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Dong</surname><given-names>H</given-names> </name><name name-style="western"><surname>Lio</surname><given-names>J</given-names> </name><name name-style="western"><surname>Sherer</surname><given-names>R</given-names> </name><name name-style="western"><surname>Jiang</surname><given-names>I</given-names> </name></person-group><article-title>Some learning theories for medical educators</article-title><source>Med Sci Educ</source><year>2021</year><month>06</month><volume>31</volume><issue>3</issue><fpage>1157</fpage><lpage>1172</lpage><pub-id pub-id-type="doi">10.1007/s40670-021-01270-6</pub-id><pub-id pub-id-type="medline">34457959</pub-id></nlm-citation></ref><ref id="ref69"><label>69</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Plaisant</surname><given-names>C</given-names> </name></person-group><article-title>The challenge of information visualization evaluation</article-title><source>Proceedings of the Working Conference on Advanced Visual Interfaces</source><year>2004</year><publisher-name>Association for Computing Machinery</publisher-name><pub-id pub-id-type="doi">10.1145/989863.989880</pub-id></nlm-citation></ref><ref id="ref70"><label>70</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Keemink</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Custers</surname><given-names>E</given-names> </name><name name-style="western"><surname>van Dijk</surname><given-names>S</given-names> </name><name name-style="western"><surname>Ten Cate</surname><given-names>O</given-names> </name></person-group><article-title>Illness script development in pre-clinical education through case-based clinical reasoning training</article-title><source>Int J Med Educ</source><year>2018</year><month>02</month><day>9</day><volume>9</volume><fpage>35</fpage><lpage>41</lpage><pub-id pub-id-type="doi">10.5116/ijme.5a5b.24a9</pub-id><pub-id pub-id-type="medline">29428911</pub-id></nlm-citation></ref><ref id="ref71"><label>71</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Moghadami</surname><given-names>M</given-names> </name><name name-style="western"><surname>Amini</surname><given-names>M</given-names> </name><name name-style="western"><surname>Moghadami</surname><given-names>M</given-names> </name><name name-style="western"><surname>Dalal</surname><given-names>B</given-names> </name><name name-style="western"><surname>Charlin</surname><given-names>B</given-names> </name></person-group><article-title>Teaching clinical reasoning to undergraduate medical students by illness script method: a randomized controlled trial</article-title><source>BMC Med Educ</source><year>2021</year><month>02</month><day>2</day><volume>21</volume><issue>1</issue><fpage>87</fpage><pub-id pub-id-type="doi">10.1186/s12909-021-02522-0</pub-id><pub-id pub-id-type="medline">33531017</pub-id></nlm-citation></ref><ref id="ref72"><label>72</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Young</surname><given-names>JQ</given-names> </name><name name-style="western"><surname>van Dijk</surname><given-names>SM</given-names> </name><name name-style="western"><surname>O&#x2019;Sullivan</surname><given-names>PS</given-names> </name><name name-style="western"><surname>Custers</surname><given-names>EJ</given-names> </name><name name-style="western"><surname>Irby</surname><given-names>DM</given-names> </name><name name-style="western"><surname>ten Cate</surname><given-names>O</given-names> </name></person-group><article-title>Influence of learner knowledge and case complexity on handover accuracy and cognitive load: results from a simulation study</article-title><source>Med Educ</source><year>2016</year><month>09</month><volume>50</volume><issue>9</issue><fpage>969</fpage><lpage>978</lpage><pub-id pub-id-type="doi">10.1111/medu.13107</pub-id></nlm-citation></ref><ref id="ref73"><label>73</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Custers</surname><given-names>E</given-names> </name><name name-style="western"><surname>Boshuizen</surname><given-names>HPA</given-names> </name><name name-style="western"><surname>Schmidt</surname><given-names>HG</given-names> </name></person-group><article-title>The role of illness scripts in the development of medical diagnostic expertise: results from an interview study</article-title><source>Cogn Instr</source><year>1998</year><month>12</month><volume>16</volume><issue>4</issue><fpage>367</fpage><lpage>398</lpage><pub-id pub-id-type="doi">10.1207/s1532690xci1604_1</pub-id></nlm-citation></ref><ref id="ref74"><label>74</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bland</surname><given-names>AC</given-names> </name><name name-style="western"><surname>Kreiter</surname><given-names>CD</given-names> </name><name name-style="western"><surname>Gordon</surname><given-names>JA</given-names> </name></person-group><article-title>The psychometric properties of five scoring methods applied to the script concordance test</article-title><source>Acad Med</source><year>2005</year><month>04</month><volume>80</volume><issue>4</issue><fpage>395</fpage><lpage>399</lpage><pub-id pub-id-type="doi">10.1097/00001888-200504000-00019</pub-id><pub-id pub-id-type="medline">15793026</pub-id></nlm-citation></ref><ref id="ref75"><label>75</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Cepeda</surname><given-names>NJ</given-names> </name><name name-style="western"><surname>Pashler</surname><given-names>H</given-names> </name><name name-style="western"><surname>Vul</surname><given-names>E</given-names> </name><name name-style="western"><surname>Wixted</surname><given-names>JT</given-names> </name><name name-style="western"><surname>Rohrer</surname><given-names>D</given-names> </name></person-group><article-title>Distributed practice in verbal recall tasks: a review and quantitative synthesis</article-title><source>Psychol Bull</source><year>2006</year><month>05</month><volume>132</volume><issue>3</issue><fpage>354</fpage><lpage>380</lpage><pub-id pub-id-type="doi">10.1037/0033-2909.132.3.354</pub-id><pub-id pub-id-type="medline">16719566</pub-id></nlm-citation></ref><ref id="ref76"><label>76</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Akcali</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Cubuk</surname><given-names>HS</given-names> </name><name name-style="western"><surname>Oguz</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Automated extraction of key entities from non-English mammography reports using named entity recognition with prompt engineering</article-title><source>Bioengineering (Basel)</source><year>2025</year><month>02</month><day>10</day><volume>12</volume><issue>2</issue><fpage>168</fpage><pub-id pub-id-type="doi">10.3390/bioengineering12020168</pub-id><pub-id pub-id-type="medline">40001686</pub-id></nlm-citation></ref><ref id="ref77"><label>77</label><nlm-citation citation-type="web"><article-title>Chinese medical information processing challenge list_CBLUE</article-title><source>Alibaba Cloud</source><year>2025</year><month>02</month><day>21</day><access-date>2026-04-11</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://tianchi.aliyun.com/cblue">https://tianchi.aliyun.com/cblue</ext-link></comment></nlm-citation></ref><ref id="ref78"><label>78</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Pang</surname><given-names>X</given-names> </name></person-group><article-title>MExplore</article-title><source>GitHub</source><access-date>2026-4-11</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://github.com/Px956678784/MExplore">https://github.com/Px956678784/MExplore</ext-link></comment></nlm-citation></ref></ref-list></back></article>