<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMI</journal-id>
      <journal-id journal-id-type="nlm-ta">JMIR Med Inform</journal-id>
      <journal-title>JMIR Medical Informatics</journal-title>
      <issn pub-type="epub">2291-9694</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v10i9e38414</article-id>
      <article-id pub-id-type="pmid">36053574</article-id>
      <article-id pub-id-type="doi">10.2196/38414</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>Leveraging Representation Learning for the Construction and Application of a Knowledge Graph for Traditional Chinese Medicine: Framework Development Study</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Hao</surname>
            <given-names>Tianyong</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Zhang</surname>
            <given-names>Chaoyang</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Li</surname>
            <given-names>Jianqiang</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author">
          <name name-style="western">
            <surname>Weng</surname>
            <given-names>Heng</given-names>
          </name>
          <degrees>MD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-5203-8663</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author">
          <name name-style="western">
            <surname>Chen</surname>
            <given-names>Jielong</given-names>
          </name>
          <degrees>MD</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-9833-265X</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author">
          <name name-style="western">
            <surname>Ou</surname>
            <given-names>Aihua</given-names>
          </name>
          <degrees>BA</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-6123-7931</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author" corresp="yes">
          <name name-style="western">
            <surname>Lao</surname>
            <given-names>Yingrong</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <address>
            <institution>State Key Laboratory of Dampness Syndrome of Chinese Medicine</institution>
            <institution>Second Affiliated Hospital of Guangzhou University of Chinese Medicine</institution>
            <addr-line>Dade road No. 111</addr-line>
            <addr-line>Guangzhou, 510120</addr-line>
            <country>China</country>
            <phone>86 81887233 ext 35933</phone>
            <email>laoyingrong@qq.com</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-0549-3655</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>State Key Laboratory of Dampness Syndrome of Chinese Medicine</institution>
        <institution>Second Affiliated Hospital of Guangzhou University of Chinese Medicine</institution>
        <addr-line>Guangzhou</addr-line>
        <country>China</country>
      </aff>
      <aff id="aff2">
        <label>2</label>
        <institution>School of Information Science</institution>
        <institution>Guangdong University of Finance &#38; Economics</institution>
        <addr-line>Guangzhou</addr-line>
        <country>China</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Yingrong Lao <email>laoyingrong@qq.com</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <month>9</month>
        <year>2022</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>2</day>
        <month>9</month>
        <year>2022</year>
      </pub-date>
      <volume>10</volume>
      <issue>9</issue>
      <elocation-id>e38414</elocation-id>
      <history>
        <date date-type="received">
          <day>31</day>
          <month>3</month>
          <year>2022</year>
        </date>
        <date date-type="rev-request">
          <day>9</day>
          <month>5</month>
          <year>2022</year>
        </date>
        <date date-type="rev-recd">
          <day>4</day>
          <month>7</month>
          <year>2022</year>
        </date>
        <date date-type="accepted">
          <day>27</day>
          <month>7</month>
          <year>2022</year>
        </date>
      </history>
      <copyright-statement>©Heng Weng, Jielong Chen, Aihua Ou, Yingrong Lao. Originally published in JMIR Medical Informatics (https://medinform.jmir.org), 02.09.2022.</copyright-statement>
      <copyright-year>2022</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on https://medinform.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://medinform.jmir.org/2022/9/e38414" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>Knowledge discovery from treatment data records from Chinese physicians is a dramatic challenge in the application of artificial intelligence (AI) models to the research of traditional Chinese medicine (TCM).</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>This paper aims to construct a TCM knowledge graph (KG) from Chinese physicians and apply it to the decision-making related to diagnosis and treatment in TCM.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>A new framework leveraging a representation learning method for TCM KG construction and application was designed. A transformer-based Contextualized Knowledge Graph Embedding (CoKE) model was applied to KG representation learning and knowledge distillation. Automatic identification and expansion of multihop relations were integrated with the CoKE model as a pipeline. Based on the framework, a TCM KG containing 59,882 entities (eg, diseases, symptoms, examinations, drugs), 17 relations, and 604,700 triples was constructed. The framework was validated through a link predication task.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>Experiments showed that the framework outperforms a set of baseline models in the link prediction task using the standard metrics mean reciprocal rank (MRR) and Hits@N. The knowledge graph embedding (KGE) multitagged TCM discriminative diagnosis metrics also indicated the improvement of our framework compared with the baseline models.</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>Experiments showed that the clinical KG representation learning and application framework is effective for knowledge discovery and decision-making assistance in diagnosis and treatment. Our framework shows superiority of application prospects in tasks such as KG-fused multimodal information diagnosis, KGE-based text classification, and knowledge inference–based medical question answering.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>knowledge graph</kwd>
        <kwd>knowledge embedding</kwd>
        <kwd>traditional Chinese medicine</kwd>
        <kwd>knowledge discovery</kwd>
        <kwd>medicine</kwd>
        <kwd>clinical</kwd>
        <kwd>framework</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <sec>
        <title>Background</title>
        <p>Having a long history of 5000 years, traditional Chinese medicine (TCM) is featured as the scientific thinking of holistic view and syndrome differentiation, as well as the long-time practice of technical methods of personalized treatment. TCM has the advantages of precise clinical efficacy, relatively safe medication, flexible treatment, and relatively low cost [<xref ref-type="bibr" rid="ref1">1</xref>]. However, a large amount of empirical knowledge exists with Chinese physicians, which is difficult to be applied directly in assisting clinical decision-making systems. At the same time, the dismantling of medical guidelines alone cannot cope with all situations, and existing clinical assisted decision-making systems cannot explain the ins and outs of diagnostic decisions as senior experts do.</p>
        <p>The combination of knowledge graphs (KGs) and artificial intelligence (AI) models has the bilateral advantages of “black box” and “logic.” Using knowledge graph embedding (KGE) techniques, KGE models may partially simulate the cognitive process of the human brain by representing massive entities, relations, and attributes. By combining with the causal events extracted from the text of event descriptions by causality extraction techniques, event information can be presented in structured form. KGs and machine learning models are expected to be integrated to assist machine understanding and concept interpretation, allowing the decision-making process of machines to be interpretable. However, how to construct a TCM KG and apply it with KGE models is still a challengeable problem.</p>
        <p>To that end, this paper proposes a new framework leveraging a representation learning method for TCM KG construction and application. TCM knowledge is extracted from Chinese physicians based on 1 of our previous works [<xref ref-type="bibr" rid="ref2">2</xref>] by using an automatic procedure of information extraction concept normalization, <italic>entity alignment</italic>. The framework collects multimodal information about Chinese medicines to support the automatic construction of personalized KGs according to clinical disease treatments by Chinese physicians. Our framework has application potential in text classification, KG-based question answering, and recommendations of practitioners and specialties.</p>
        <p>The main contributions of this paper are threefold: (1) A new framework for the construction and application of TCM KG by leveraging representation learning is proposed, (2) a transformer-based Contextualized Knowledge Graph Embedding (CoKE) model is applied to KG representation learning and knowledge distillation by integrating multihop relations, and (3) a TCM KG containing 59,882 entities, 17 relations, and 604,700 triples is constructed.</p>
      </sec>
      <sec>
        <title>Related Work</title>
        <sec>
          <title>Medical Knowledge Graph</title>
          <p>The concept of KG was proposed by Google in 2012. Research applications evolved by previously improving the capabilities of search engines and enhancing the search quality and experience of users related to finance, healthcare, geography, e-commerce, and medical care. There exist many KGs, including on Google Knowledge Graph [<xref ref-type="bibr" rid="ref3">3</xref>], DBpedia [<xref ref-type="bibr" rid="ref4">4</xref>], Yet Another Great Ontology (YAGO; Max Planck Institute for Computer Science) [<xref ref-type="bibr" rid="ref5">5</xref>], and FreeBase (Metaweb Technologies, Inc.) [<xref ref-type="bibr" rid="ref6">6</xref>]. In China, there are Zhi Cube (Sogou), Zhi Xin (Baidu), zhishi.me (Shanghai Jiao Tong University) [<xref ref-type="bibr" rid="ref7">7</xref>], and the GDM Lab Chinese KG project (Fudan University) [<xref ref-type="bibr" rid="ref8">8</xref>]. In the medical field, the KG of medicine NKIMed [<xref ref-type="bibr" rid="ref9">9</xref>] was developed by the Institute of Computer Technology of the Chinese Academy of Sciences, and the KG of Chinese medicine [<xref ref-type="bibr" rid="ref10">10</xref>] was constructed by the Institute of Chinese Medicine Information of the Chinese Academy of Traditional Chinese Medicine. The Traditional Chinese Medicine Language System (TCMLS) is a relatively large semantic network for the KG of Chinese medicine [<xref ref-type="bibr" rid="ref11">11</xref>], containing more than 100,000 concepts and 1 million semantic relations, which basically covers the conceptual system of TCM disciplines. The TCMLS was in the leading position of the TCM community in terms of its scale and completeness. Rotmensch et al [<xref ref-type="bibr" rid="ref12">12</xref>] extracted positive mentions of diseases and symptoms (concepts) from structured and unstructured data in electronic medical records (EMRs) and used them to construct a health KG automatically.</p>
        </sec>
        <sec>
          <title>Knowledge Graph Representation Learning</title>
          <p>Graph neural networks (GNNs) are deep learning architectures for graph-structured data, which combine end-to-end learning with inductive reasoning. GNNs are promising research topics of AI, and they are expected to solve the problems of causal inference and interpretability that cannot be handled by traditional deep learning models. KG representation learning is a critical branch of the research on GNNs and plays a nontrivial role in knowledge acquisition and downstream application. KG representation learning consists of elements such as representation spaces (pointwise space, complex vector space, gaussian distribution, manifold, and group), scoring functions (distance-based and semantic-matching scoring functions), and encoding models (linear/bilinear, factorization models and neural networks).</p>
          <p>Translational models leverage translational distances (eg, L1 or L2 norm) to model relations between head and tail entities. TransE is one of the representative translational models [<xref ref-type="bibr" rid="ref13">13</xref>]. Dealing with 1-to-N, N-to-1, and N-to-N relations, TransE suffered from inefficiency problems in representing head or tail entities. To alleviate such problems, KGE models, including TransH [<xref ref-type="bibr" rid="ref14">14</xref>], TransR [<xref ref-type="bibr" rid="ref15">15</xref>], and TransD [<xref ref-type="bibr" rid="ref16">16</xref>], were designed to impose translational distance constraints through different entity projection strategies. RotatE considers the embedding vectors of relations as rotations from source entities to target entities in a complex space [<xref ref-type="bibr" rid="ref17">17</xref>].</p>
          <p>The basic idea of factorization models is to decompose the matrix of each slice in a 3-way tensor into a product of entity vectors and relation matrices in the lower-dimensional space. The RESCAL model leveraged a relation-associated matrix to capture interactions between head and tail entities, which required a large number of parameters to model relations [<xref ref-type="bibr" rid="ref18">18</xref>]. Therefore, vector forms of relations were introduced in DistMult [<xref ref-type="bibr" rid="ref19">19</xref>] to decrease model parameters by restricting the interaction matrices to diagonal matrices. To increase the interactions between head and tail entities, a circular correlation operation was leveraged as the score function in the expressive HolE model [<xref ref-type="bibr" rid="ref20">20</xref>]. Inspired by DistMult, the ComplEx model extended the representations of entities and relations by utilizing embedding vectors in a complex space [<xref ref-type="bibr" rid="ref21">21</xref>]. An expressive KGE model named SimplE used 2 vectors for each entity to learn independent parameters through simplifying ComplEx by removing redundant computation [<xref ref-type="bibr" rid="ref22">22</xref>].</p>
          <p>In recent years, inspired by convolution operations, convolution-based KGE models, such as ConvE [<xref ref-type="bibr" rid="ref23">23</xref>], ConvKB [<xref ref-type="bibr" rid="ref24">24</xref>], and CapsE [<xref ref-type="bibr" rid="ref25">25</xref>], were designed as different strategies to capture features between entities and relations for KG representation learning. A KGE model named knowledge base attention (KBAT) extended the graph attention (GAT) network by exploring the multihop representation of a given entity for representation aggregation via multihead attention and graph attention mechanisms [<xref ref-type="bibr" rid="ref26">26</xref>]. The natural language pretraining model BERT [<xref ref-type="bibr" rid="ref27">27</xref>] learned to integrate contextual information in the KG based on the representation of the transformer [<xref ref-type="bibr" rid="ref28">28</xref>]. CoKE [<xref ref-type="bibr" rid="ref29">29</xref>] used a transformer to encode edge and path sequences. These promising methods have attracted much attention due to the high efficiency of convolution in representation learning. CoKE aimed to learn the dynamic adaptive representations of entities and relations based on a rich graph structure context. Compared with static representations, the performance of contextual models is state of the art, since the representations combined with contextual semantic information are richer and more flexible. Despite the use of a transformer, CoKE was still parameter-efficient to obtain competitive performance with fewer parameters. The comparison of the KG representation learning models is shown in <xref ref-type="table" rid="table1">Table 1</xref>.</p>
          <table-wrap position="float" id="table1">
            <label>Table 1</label>
            <caption>
              <p>Comparison of baseline KGE<sup>a</sup> models.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="30"/>
              <col width="370"/>
              <col width="0"/>
              <col width="300"/>
              <col width="0"/>
              <col width="300"/>
              <thead>
                <tr valign="top">
                  <td colspan="3">Model</td>
                  <td colspan="2">Scoring function f<sub>r</sub>(h,t)</td>
                  <td>Entity and relation embedding</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td colspan="6">
                    <bold>Translational model</bold>
                  </td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>TransE [<xref ref-type="bibr" rid="ref13">13</xref>]</td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig8.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig9.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>TransH [<xref ref-type="bibr" rid="ref14">14</xref>]</td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig10.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig11.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>TransR [<xref ref-type="bibr" rid="ref15">15</xref>]</td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig12.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig13.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>TransD [<xref ref-type="bibr" rid="ref16">16</xref>]</td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig14.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig15.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                </tr>
                <tr valign="top">
                  <td colspan="6">
                    <bold>Linear/bilinear model</bold>
                  </td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>SimplE [<xref ref-type="bibr" rid="ref22">22</xref>]</td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig16.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig17.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>HolE [<xref ref-type="bibr" rid="ref20">20</xref>]</td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig18.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig19.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                </tr>
                <tr valign="top">
                  <td colspan="6">
                    <bold>Rotational model</bold>
                  </td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>QuatE [<xref ref-type="bibr" rid="ref30">30</xref>]</td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig20.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig21.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>RotatE [<xref ref-type="bibr" rid="ref17">17</xref>]</td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig22.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig23.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                </tr>
                <tr valign="top">
                  <td colspan="6">
                    <bold>Convolutional neural network</bold>
                  </td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>ConvE [<xref ref-type="bibr" rid="ref23">23</xref>]</td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig24.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig25.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>ConvKB [<xref ref-type="bibr" rid="ref24">24</xref>]</td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig26.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig27.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                </tr>
                <tr valign="top">
                  <td colspan="6">
                    <bold>GNN<sup>b</sup></bold>
                  </td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>KBAT<sup>c</sup> [<xref ref-type="bibr" rid="ref26">26</xref>]</td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig28.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig29.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                </tr>
                <tr valign="top">
                  <td colspan="6">
                    <bold>Neural network transformer</bold>
                  </td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>CoKE<sup>d</sup> [<xref ref-type="bibr" rid="ref29">29</xref>]</td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig30.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                  <td colspan="2">
                    <disp-formula>
                      <graphic xlink:href="medinform_v10i9e38414_fig31.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
                    </disp-formula>
                  </td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table1fn1">
                <p><sup>a</sup>KGE: knowledge graph embedding.</p>
              </fn>
              <fn id="table1fn2">
                <p><sup>b</sup>GNN: graph neural network.</p>
              </fn>
              <fn id="table1fn3">
                <p><sup>c</sup>KBAT: knowledge base attention.</p>
              </fn>
              <fn id="table1fn4">
                <p><sup>d</sup>CoKE: Contextualized Knowledge Graph Embedding.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
        </sec>
        <sec>
          <title>Application of Medical Knowledge Graphs</title>
          <p>The hot topics related to the application of medical KGs are KG-fused multimodal information diagnosis, KGE-based text classification, and knowledge inference–based medical question answering and assisted diagnosis. Shen et al [<xref ref-type="bibr" rid="ref31">31</xref>] reused the existing knowledge base to build a high-quality KG and designed a prediction model to explore pharmacology and KG features. The model allowed the user to gain a better understanding of the drug properties from a drug similarity perspective and insights that were not easily observed in individual drugs. Zheng et al [<xref ref-type="bibr" rid="ref32">32</xref>] took advantage of 4 kinds of modality data (X-ray images, computed tomography [CT] images, ultrasound images, and text descriptions of diagnoses) to construct a KG. The model leveraged multimodal KG attention embedding for diagnosis of COVID-19. The experimental results demonstrated that it was essential to capture and join the importance of single- and multilevel modality information in a multimodal KG. Li et al [<xref ref-type="bibr" rid="ref33">33</xref>] designed an AI-powered voice assistant by constructing a comprehensive knowledge base with ontologies of defined Alzheimer disease and related dementia (ADRD) diet care and user profiles. They extended the model with external KGs, such as FoodData Central and DrugBank, which personalized ADRD diet services provided through a semantics-based KG search and reasoning engine.</p>
          <p>With the development of deep learning methods, diagnostic decisions have become interpretable. Theoretically, rule-based engines may infinitely approximate the performance of nonlinear classifiers by mining the expanded knowledge. In other words, through the integration of interpretable knowledge rules, rule-based engines may approximate the performance of deep learning models. Through deep mining of rules, the clinical assisted decision-making system may be able to perform multiple rounds of rule expansion under dynamic thresholds and further extend the capability of decision-making based on existing knowledge.</p>
        </sec>
      </sec>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <sec>
        <title>TCM Knowledge Graphs</title>
        <p>To construct a TCM KG (<xref ref-type="table" rid="table2">Table 2</xref>) for ordinary usage, such as disease diagnosis and treatment assistance, we cleaned the EMR data set of diagnosis and treatment of TCM diseases and represented the relations of entities in triples. For instance, given a description text of insulin resistance as a mechanism of type 2 diabetes, the entities and relations in the sentence were extracted and organized into a disease mechanism triple of (insulin resistance, mechanism=&#62;disease, diabetes). A KG was defined as G=(E,R,S), where entities, relations, and triples are <inline-graphic xlink:href="medinform_v10i9e38414_fig32.png" xlink:type="simple" mimetype="image"/>, respectively, and &#124;E&#124; and &#124;R&#124; are the counts of entities and relations. The triples consisted of entities, relations, describing concepts, or attributes.</p>
        <p>Traditional KGE models are designed to learn static representations of entities and relations. The features of graph contexts are obtained by representing neighbor entities and relations. Different meanings are expressed by entities and relations in diverse contexts, as words appear in different textual contexts. Multihop relations (ie, paths between entities) can provide rich contextual features for reasoning in KG [<xref ref-type="bibr" rid="ref29">29</xref>]. Existing work [<xref ref-type="bibr" rid="ref34">34</xref>] shows that multihop relation paths contain rich inference patterns between entities. Since not all relation paths are reliable, we designed a causal-constraint algorithm to filter the reliability of relation paths. Relation paths were represented via semantic composition of relation embeddings. The screened multihop relations were extended to triple alternative combinations.</p>
        <p>The rules for screening potential multihop causal relations are shown in <xref rid="figure1" ref-type="fig">Figure 1</xref>. For example, there exist triples (<italic>insulin resistance</italic>, <italic>treat</italic>, <italic>diabetes mellitus</italic>) and (<italic>metformin</italic>, <italic>mechanism</italic>, <italic>insulin resistance</italic>) in a clinical KG describing the relations between clinical mechanism and disease (or drug) as a positive example in the figure. The relations can be inferred as the causal multihop relation between a drug and a disease by the rules drug=&#62;mechanism and mechanism=&#62;disease, indicating that metformin can treat insulin-resistant diabetes. The triples (<italic>dyslipidemia</italic>, <italic>symptom</italic>, <italic>diabetes mellitus</italic>) and (<italic>dyslipidemia</italic>, <italic>symptom</italic>, <italic>CKD</italic> [where CKD refers to chronic kidney disease]) co-occurred and thus could not reflect the causal relation between diabetes mellitus and CKD or dyslipidemia. Such negative triples were screened according to the rules.</p>
        <p>An example of a casual multihop relation of TCM disease (abdominal mass)=&#62;mechanism (<italic>phlegm dampness</italic>, <italic>toxin</italic>, <italic>blood stasis</italic>)–mechanism=&#62;mechanism (<italic>clearing heat-toxin</italic>, <italic>eliminating dampness</italic>)–disease=&#62;drug (<italic>root of Chinese</italic> Pulsatilla) can be inferred according to the rules (<italic>abdominal mass</italic>, <italic>disease=&#62;drug</italic>, <italic>phlegm dampness</italic>, <italic>toxin</italic>, <italic>blood stasis</italic>), (<italic>phlegm dampness</italic>, <italic>toxin</italic>, <italic>blood stasis</italic>, <italic>mechanism=&#62;mechanism</italic>, <italic>clearing heat-toxin</italic>, <italic>eliminating dampness</italic>), and (<italic>abdominal mass</italic>, <italic>disease=&#62;drug</italic>, <italic>root of Chinese</italic> Pulsatilla). In other words, casual multihop relations of TCM can be inferred, which conform to the cognition of diseases–syndrome–principle–method–recipe–medicines of TCM, including the aforementioned path disease=&#62;mechanism=&#62;treatment=&#62;drug.</p>
        <p>The semantics of the entities <italic>diabetes mellitus</italic> and <italic>metformin</italic> were enriched by the embeddings of the 2-hop path inferred by triples (<italic>metformin</italic>, <italic>mechanism</italic>, <italic>insulin resistance</italic>) and (<italic>insulin resistance</italic>, <italic>treat</italic>, <italic>diabetes mellitus</italic>). To represent multihop relations, given the 2-hop path from the entity <italic>metformin</italic> to <italic>diabetes</italic> <italic>mellitus</italic>, triple forms (<italic>metformin</italic>, <italic>mechanism-treat</italic>, <italic>diabetes mellitus</italic>) were used for consistency. Since the multihop features were integrated, the representations of entities and relations tended to have strong inference capability, which facilitated entity link prediction. The KG was represented as textual triples that described multihop relations of entities.</p>
        <table-wrap position="float" id="table2">
          <label>Table 2</label>
          <caption>
            <p>Overview of the TCM<sup>a</sup> KG<sup>b</sup>.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="500"/>
            <col width="170"/>
            <col width="170"/>
            <col width="160"/>
            <thead>
              <tr valign="top">
                <td>Relation name</td>
                <td>Heads, n</td>
                <td>Tails, n</td>
                <td>Triples, n</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>symptom=&#62;symptom</td>
                <td>8101</td>
                <td>8544</td>
                <td>51,345</td>
              </tr>
              <tr valign="top">
                <td>disease=&#62;symptom</td>
                <td>12,225</td>
                <td>15,071</td>
                <td>133,648</td>
              </tr>
              <tr valign="top">
                <td>disease=&#62;drug</td>
                <td>12,650</td>
                <td>11,526</td>
                <td>84,524</td>
              </tr>
              <tr valign="top">
                <td>mechanism=&#62;mechanism</td>
                <td>527</td>
                <td>51</td>
                <td>590</td>
              </tr>
              <tr valign="top">
                <td>symptom=&#62;drug</td>
                <td>3941</td>
                <td>6145</td>
                <td>24,724</td>
              </tr>
              <tr valign="top">
                <td>symptom=&#62;mechanism</td>
                <td>6544</td>
                <td>1096</td>
                <td>10,906</td>
              </tr>
              <tr valign="top">
                <td>symptom=&#62;disease</td>
                <td>8101</td>
                <td>10,391</td>
                <td>87,651</td>
              </tr>
              <tr valign="top">
                <td>mechanism=&#62;department</td>
                <td>1908</td>
                <td>65</td>
                <td>4408</td>
              </tr>
              <tr valign="top">
                <td>symptom=&#62;body parts</td>
                <td>318</td>
                <td>85</td>
                <td>548</td>
              </tr>
              <tr valign="top">
                <td>mechanism=&#62;body parts</td>
                <td>2217</td>
                <td>72</td>
                <td>3221</td>
              </tr>
              <tr valign="top">
                <td>mechanism=&#62;symptom</td>
                <td>2147</td>
                <td>4191</td>
                <td>16,377</td>
              </tr>
              <tr valign="top">
                <td>symptom=&#62;department</td>
                <td>10,157</td>
                <td>178</td>
                <td>24,870</td>
              </tr>
              <tr valign="top">
                <td>disease=&#62;mechanism</td>
                <td>7774</td>
                <td>5304</td>
                <td>46,425</td>
              </tr>
              <tr valign="top">
                <td>disease=&#62;body parts</td>
                <td>7607</td>
                <td>110</td>
                <td>13,505</td>
              </tr>
              <tr valign="top">
                <td>disease=&#62;department</td>
                <td>14,484</td>
                <td>284</td>
                <td>40,762</td>
              </tr>
              <tr valign="top">
                <td>disease=&#62;disease</td>
                <td>9728</td>
                <td>10,545</td>
                <td>40,575</td>
              </tr>
              <tr valign="top">
                <td>mechanism =&#62;disease</td>
                <td>2228</td>
                <td>5443</td>
                <td>20,621</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table2fn1">
              <p><sup>a</sup>TCM: traditional Chinese medicine.</p>
            </fn>
            <fn id="table2fn2">
              <p><sup>b</sup>KG: knowledge graph.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <fig id="figure1" position="float">
          <label>Figure 1</label>
          <caption>
            <p>Positive and negative examples of multihop relation filtering and generation. CKD: chronic kidney disease; T2DM: type 2 diabetes mellitus.</p>
          </caption>
          <graphic xlink:href="medinform_v10i9e38414_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Knowledge Graph Representation Framework</title>
        <p>After preprocessing of the TCM KGs, we applied a CoKE-based KG representation learning model based on a diagnosis and treatment KG of Chinese and Western medicine and proposed a new KG representation framework. Compared with popular knowledge representation learning models, such as TransE and KBAT, our framework features the fusion of CoKE and multihop relations. The framework was verified with downstream applications, such as assisted decision-making and question answering, as shown in <xref rid="figure2" ref-type="fig">Figure 2</xref>.</p>
        <fig id="figure2" position="float">
          <label>Figure 2</label>
          <caption>
            <p>Proposed framework of TCM KG representation learning. CoKE: Contextualized Knowledge Graph Embedding; KG: knowledge graph; TCM: traditional Chinese medicine.</p>
          </caption>
          <graphic xlink:href="medinform_v10i9e38414_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Entity Link Prediction</title>
        <p>The CoKE model was leveraged as the base model in this paper. The BERT model was leveraged to learn contextualized embeddings of entities and relations in CoKE. The input sequence X = (x<sub>1</sub>, x<sub>2</sub>, …, x<sub>n</sub>) consisted of the embeddings of a head entity x<sub>1</sub> and a tail entity x<sub>n</sub>, while the embeddings of relations were denoted as x<sub>2</sub> from x<sub>n–1</sub>. Given x<sub>i</sub>from the input sequence, the hidden representation h<sub>i</sub> was expressed as Equation 1:</p>
        <disp-formula>
          <inline-graphic xlink:href="medinform_v10i9e38414_fig33.png" xlink:type="simple" mimetype="image"/>
        </disp-formula>
        <p>where <inline-graphic xlink:href="medinform_v10i9e38414_fig34.png" xlink:type="simple" mimetype="image"/> is the embedding of an element and <inline-graphic xlink:href="medinform_v10i9e38414_fig35.png" xlink:type="simple" mimetype="image"/> is the positional embedding of an element. The former was used to identify the current entities or relations in <inline-graphic xlink:href="medinform_v10i9e38414_fig36.png" xlink:type="simple" mimetype="image"/>, and the latter presented the positional features of the element in the sequence. The constructed hidden representations were fed into transformer encoders of L layers as Equation 2:</p>
        <disp-formula>
          <inline-graphic xlink:href="medinform_v10i9e38414_fig37.png" xlink:type="simple" mimetype="image"/>
        </disp-formula>
        <p>where <inline-graphic xlink:href="medinform_v10i9e38414_fig38.png" xlink:type="simple" mimetype="image"/> is the hidden representation of x<sub>i</sub> at the l-th layer of the encoder. A multihead self-attention mechanism was leveraged by the transformer, which allowed each element to attend to other elements in the sequence effectively for contextual feature modeling. As the use of transformers has become ubiquitous recently, we omitted a detailed description of the transformer. The final hidden representations <inline-graphic xlink:href="medinform_v10i9e38414_fig39.png" xlink:type="simple" mimetype="image"/> are representations for entities and relations within the sequence X. The learned representations were naturally contextualized and automatically adaptive to the input.</p>
      </sec>
      <sec>
        <title>Multihop Relational Representation Learning</title>
        <p>Given a triple (s,r,o) in a KG, the contexts between a head and a tail entity can be described as an edge and a path. An edge s→r→o is presented as a sequence that can be viewed as a triple. For instance, an edge <italic>metformin</italic>→<italic>mechanism</italic>→<italic>insulin resistance</italic> can form a triple (<italic>metformin</italic>, <italic>mechanism</italic>, <italic>insulin resistance</italic>) equivalently. As the basic unit of a KG, an edge (or a triple) is the simplest form of a graph context describing an entity. Another context is a path s→r<sub>1</sub>→…→r<sub>k</sub>→o as a sequence consisted of head and tail entities and a list of linked relations between them. For instance, the path <inline-graphic xlink:href="medinform_v10i9e38414_fig40.png" xlink:type="simple" mimetype="image"/> describes multihop relations between the head entity <italic>metformin</italic> and the tail entity <italic>diabetes</italic> <italic>mellitus</italic>, where <italic>insulin resistance</italic> is the intermediate entity in the path, while <italic>mechanism</italic> and <italic>treat</italic> are the relations. The path can be expressed as a triple (<italic>metformin</italic>, <italic>mechanism-treat</italic>, <italic>diabetes mellitus</italic>). Consisting of contextual features of entities, the multihop path representation can be leveraged for reasoning in a KG.</p>
        <p>To verify the effectiveness of the model, experiments of entity link prediction in knowledge graph completion (KGC) [<xref ref-type="bibr" rid="ref35">35</xref>] and multihop relation representation learning were conducted. Entity link prediction refers to a task that predicts missing target entities of triples (h, r, ?) and (?, r, t) with a candidate entity set by semantic constraints of KGE models. PathQuery answering [<xref ref-type="bibr" rid="ref36">36</xref>] was utilized in the experiments of multihop relation representation learning. Given a source entity s and a relation path p, a set of target entities that were inferred from the source entity s via the path p was predicted.</p>
        <p>In entity link prediction, our model was trained to predict missing target entities, given a context in the KG, answering 1-hop or multihop queries. Different strategies were considered to train our model with respect to the cases of edges and paths. Each edge s→r→o is associated with 2 instances ?→r→o and s→r→?, which can be regarded as 1-hop query answering. For instance, <italic>metformin</italic>→<italic>mechanism</italic>→<italic>?</italic> is to answer the query, What is the mechanism of m<italic>etformin</italic>? Similarly, each path s→r<sub>1</sub>→…→r<sub>k</sub>→o is also associated with 2 instances, one to predict s and the other to predict o, which can be viewed as multihop question answering. For instance, <inline-graphic xlink:href="medinform_v10i9e38414_fig41.png" xlink:type="simple" mimetype="image"/> is to answer the query, What disease can be treated by the mechanism of <italic>metformin</italic>?</p>
        <p>In the training procedure, edges or paths were unified as an input sequence X = (x<sub>1</sub>, x<sub>2</sub>, …, x<sub>n</sub>). Two instances were created by replacing x<sub>1</sub> with a special token [MASK] for s prediction and by replacing x<sub>n</sub> with [MASK] for o prediction. The masked sequence was fed into the transformer encoding blocks to obtain the final hidden representation for target entity prediction.</p>
        <p>As in the BERT model, the representations of the masked entities were fed into a feedforward neural network and a standard Softmax layer was leveraged for classification (Equation 3):</p>
        <disp-formula>
          <inline-graphic xlink:href="medinform_v10i9e38414_fig42.png" xlink:type="simple" mimetype="image"/>
        </disp-formula>
        <p>where z<sub>1</sub> and z<sub>n</sub> are the representations of h<sup>L</sup><sub>1</sub> and h<sup>L</sup><sub>n</sub> produced by the feedforward layer, while <inline-graphic xlink:href="medinform_v10i9e38414_fig43.png" xlink:type="simple" mimetype="image"/> is a matrix shared with the input element embedding matrix for classification. D is the hidden size, V is the size of the entity vocabulary, and p<sub>1</sub> and p<sub>n</sub> are the predicted distributions of target entities s and o. Cross-entropy loss was leveraged as the loss function for classification (Equation 4):</p>
        <disp-formula>
          <inline-graphic xlink:href="medinform_v10i9e38414_fig44.png" xlink:type="simple" mimetype="image"/>
        </disp-formula>
        <p>where y<sub>t</sub> and p<sub>t</sub> are the t-th components of the 1-hot label vector y and the distribution vector p, respectively. A label-smoothing strategy was leveraged to lessen the restriction of 1-hot labels. In other words, the value of the target entity was set to ε, while y<sub>t</sub> = (1 – ε)/(V – 1) for incorrect entities in the candidate entity set.</p>
      </sec>
      <sec>
        <title>Knowledge Distillation</title>
        <p>Inspired by the idea of TinyBERT [<xref ref-type="bibr" rid="ref37">37</xref>] for knowledge distillation, our model CoKE-distillation contains a teacher and a student model for knowledge distillation, as shown in <xref rid="figure3" ref-type="fig">Figure 3</xref>.</p>
        <fig id="figure3" position="float">
          <label>Figure 3</label>
          <caption>
            <p>Architecture of CoKE-distillation. CoKE: Contextualized Knowledge Graph Embedding.</p>
          </caption>
          <graphic xlink:href="medinform_v10i9e38414_fig3.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <p>Our proposed CoKE-distillation model consists of 3 levels of distillation: embedding layer distillation, transformer -layer distillation, and prediction layer distillation. At the embedding layer distillation level, the embedding matrices of the student and teacher model are constrained by the mean-square error (MSE) loss (Equation 5):</p>
        <disp-formula>
          <inline-graphic xlink:href="medinform_v10i9e38414_fig45.png" xlink:type="simple" mimetype="image"/>
        </disp-formula>
        <p>where <inline-graphic xlink:href="medinform_v10i9e38414_fig46.png" xlink:type="simple" mimetype="image"/> is a trainable linear transformation matrix to project the embedding of the student model into the semantic space of the teacher model. The embedding matrices of the student and teacher models are denoted by <inline-graphic xlink:href="medinform_v10i9e38414_fig47.png" xlink:type="simple" mimetype="image"/>, where l is length of the sequence, d<sub>0</sub> is the size of the embeddings of the teacher model, and d is the size of the embeddings of the student model.</p>
        <p>At the level of transformer layer distillation, the CoKE-distillation model distills knowledge in k-layer intervals. For instance, if the student model has 4 layers, a transformer loss is calculated every 3 layers, since the teacher model has 12 layers. The first layer of the student model corresponds to the third layer of the teacher model, while the second layer of the student model corresponds to the sixth layer of the teacher model and so on. The transformer loss of each layer is divided into 2 parts, attention-based knowledge distillation and implicit state–based knowledge distillation. The loss of each layer consists of an attention-based knowledge distillation loss and a hidden state-based knowledge distillation loss.</p>
        <p>The attention-based knowledge distillation loss is expressed as Equation 6:</p>
        <disp-formula>
          <inline-graphic xlink:href="medinform_v10i9e38414_fig48.png" xlink:type="simple" mimetype="image"/>
        </disp-formula>
        <p>where h is the number of attention heads, <inline-graphic xlink:href="medinform_v10i9e38414_fig49.png" xlink:type="simple" mimetype="image"/> refers to the attention matrix corresponding to the i-th head of the teacher or the student, and l is the length of the input text.</p>
        <p>The hidden state-based knowledge distillation loss is expressed as Equation 7:</p>
        <disp-formula>
          <inline-graphic xlink:href="medinform_v10i9e38414_fig50.png" xlink:type="simple" mimetype="image"/>
        </disp-formula>
        <p>where the matrices <inline-graphic xlink:href="medinform_v10i9e38414_fig51.png" xlink:type="simple" mimetype="image"/> refer to the hidden representations of student and teacher models, respectively. At the level of prediction layer distillation, prediction loss is shown as Equation 8:</p>
        <disp-formula>
          <inline-graphic xlink:href="medinform_v10i9e38414_fig52.png" xlink:type="simple" mimetype="image"/>
        </disp-formula>
        <p>where z<sup>T</sup> and z<sup>S</sup> are the logit vectors predicted by the student and the teacher respectively, CE means the cross-entropy loss, and t means the temperature value. In our experiment, t was set to .</p>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>Data Set</title>
        <p>To evaluate the proposed model, a widely used standard data set FB15k-237 [<xref ref-type="bibr" rid="ref38">38</xref>] was used, which is a subset of the Freebase knowledge base [<xref ref-type="bibr" rid="ref6">6</xref>] with 14,541 entities and 237 relations. Due to redundant relations existing in the FB15k data set, FB15K-237 removes the inverse relations, preventing models from directly inferring target entities by inverse relations. The FB15k-237 data set is randomly divided into 3 sets (training, validation, and test sets), with 272,115 triples in the training set, 17,535 triples in the validation set, and 20,466 triples in the test set.</p>
        <p>We constructed a medical diagnosis and treatment data set of TCM, called TCMdt, consisting of entities and relations as triples. The data set contained 17 kinds of relations, 59,882 entities, and 604,700 triples without repetitive and inverse relations. There were 3811 kinds of N–1 relations, such as relation combinations <italic>mechanism-body parts</italic> and <italic>mechanism-mechanism</italic>. The rest of the relations were N–N relations, 600,868 in total. There were no 1–1 and 1–N relations in the data set. The data set was divided into a training, a validation and a test set, containing 59,882 entities and 17 relations in total. The details of the FB15k-237 and TCMdt data sets are shown as <xref ref-type="table" rid="table3">Table 3</xref>.</p>
        <p>The hypertension data set (<xref ref-type="table" rid="table4">Table 4</xref>) in TCM for the multilabel modeling task was used in our experiment to evaluate the effectiveness of KGE learning. TCM has been used for the diagnosis of hypertension and has significant advantages. Symptom analysis and modeling of TCM provide a way for clinicians to accurately and efficiently diagnose hypertension. In this study, the initial data were collected from trained practitioners and clinical practitioners. Details of 928 cases of hypertension were collected from the clinical departments of the Guangdong Provincial Hospital, with both inpatient and outpatient medical records from the Liwan district [<xref ref-type="bibr" rid="ref39">39</xref>]. All cases with incomplete information were removed from the data set, and the remaining 886 (95.47%) cases were used for analysis in this study.</p>
        <p>Each case in the data set had 129 dimensions of TCM symptom features and syndrome diagnosis labels in 1-hot format. Each case had 2-5 labels of TCM syndrome diagnosis reidentified by trained clinicians. The KGE of the syndrome entities and the symptom vectors and matrix were constructed from the aforementioned TCMdt data set.</p>
        <table-wrap position="float" id="table3">
          <label>Table 3</label>
          <caption>
            <p>Statistics of the FB15k-237 data set and the constructed TCMdt data set.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="130"/>
            <col width="140"/>
            <col width="140"/>
            <col width="200"/>
            <col width="220"/>
            <col width="170"/>
            <thead>
              <tr valign="top">
                <td>Data set</td>
                <td>Entities, n</td>
                <td>Relations, n</td>
                <td>Triples in the training set, n</td>
                <td>Triples in the validation set, n</td>
                <td>Triples in the test set, n</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>FB15k-237</td>
                <td>14,541</td>
                <td>237</td>
                <td>272,115</td>
                <td>17,535</td>
                <td>20,446</td>
              </tr>
              <tr valign="top">
                <td>TCMdt</td>
                <td>59,882</td>
                <td>17</td>
                <td>544,230‬</td>
                <td>30,235</td>
                <td>30,235</td>
              </tr>
            </tbody>
          </table>
        </table-wrap>
        <table-wrap position="float" id="table4">
          <label>Table 4</label>
          <caption>
            <p>Statistics of the hypertension data set in TCM<sup>a</sup>.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="250"/>
            <col width="250"/>
            <col width="250"/>
            <col width="250"/>
            <thead>
              <tr valign="top">
                <td>Features, n</td>
                <td>Classes, n</td>
                <td>Total cases, N</td>
                <td>Validation</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>121</td>
                <td>8</td>
                <td>886</td>
                <td>10-fold cross-validation</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table4fn1">
              <p><sup>a</sup>TCM: traditional Chinese medicine.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Baselines</title>
        <p>Baseline methods were used for comparison in the experiments, including translational models, bilinear models, a rotational model, a GNN, and a transformer-based model. The details of the models and their types are shown in <xref ref-type="table" rid="table5">Table 5</xref>.</p>
        <table-wrap position="float" id="table5">
          <label>Table 5</label>
          <caption>
            <p>Baseline methods for KG<sup>a</sup> representation learning.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="500"/>
            <col width="500"/>
            <thead>
              <tr valign="top">
                <td>Type of model</td>
                <td>Models</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Translational model</td>
                <td>TransE [<xref ref-type="bibr" rid="ref13">13</xref>], TranH [<xref ref-type="bibr" rid="ref14">14</xref>], TransR [<xref ref-type="bibr" rid="ref15">15</xref>], TransD [<xref ref-type="bibr" rid="ref16">16</xref>]</td>
              </tr>
              <tr valign="top">
                <td>Linear/bilinear model</td>
                <td>ComplEx [<xref ref-type="bibr" rid="ref21">21</xref>], DistMult [<xref ref-type="bibr" rid="ref19">19</xref>], SimplE [<xref ref-type="bibr" rid="ref22">22</xref>]</td>
              </tr>
              <tr valign="top">
                <td>Rotational model</td>
                <td>RotatE [<xref ref-type="bibr" rid="ref17">17</xref>]</td>
              </tr>
              <tr valign="top">
                <td>GNN<sup>b</sup></td>
                <td>KBAT<sup>c</sup> [<xref ref-type="bibr" rid="ref26">26</xref>]</td>
              </tr>
              <tr valign="top">
                <td>Transformer-based model</td>
                <td>CoKE<sup>d</sup> [<xref ref-type="bibr" rid="ref29">29</xref>]</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table5fn1">
              <p><sup>a</sup>KGE: knowledge graph.</p>
            </fn>
            <fn id="table5fn2">
              <p><sup>b</sup>GNN: graph neural network.</p>
            </fn>
            <fn id="table5fn3">
              <p><sup>c</sup>KBAT: knowledge base attention.</p>
            </fn>
            <fn id="table5fn4">
              <p><sup>d</sup>CoKE: Contextualized Knowledge Graph Embedding.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Evaluation Metrics</title>
        <p>With respect to the evaluation metrics, Sun et al [<xref ref-type="bibr" rid="ref40">40</xref>] found that some high performance can be attributed to the inappropriate evaluation protocols and proposed an evaluation protocol to address this problem. The proposed protocol was more robust to handle bias in the model, which could substantially affect the final results. Ruffinelli et al [<xref ref-type="bibr" rid="ref41">41</xref>] conducted systematic experiments on the training methods used in various KGE models and found that some early models (eg, RESCAL) can outperform the state-of-the-art models, after adjusting the training strategies and exploring a larger search space of hyperparameters. This indicated that the performance improvement of the models might not reflect their advantage, since the training strategies might play a critical role. Therefore, we established a unified evaluation standard to mine the valuable ideas and superiority of the models.</p>
        <p>We used the mean reciprocal rank (MRR) and Hits@N, which are frequently used evaluation metrics for link prediction task in KGs (Equations 9 and 10). Applying the filtered settings given by Wang et al [<xref ref-type="bibr" rid="ref14">14</xref>], the rank of the head or tail entities in a test triple (e<sub>i</sub>, r<sub>k</sub>, e<sub>j</sub>) was computed within a filtered entity set. The filtered entity set contained entities that could be used to generate valid triples without valid head or tail entities in the training set. A large value of the MRR indicates that the KGE model have the capability of precise entity representation, while Hits@N denotes a rate of head and tail entities that rank within N (1, 3, or 10) empirically.</p>
        <disp-formula>
          <inline-graphic xlink:href="medinform_v10i9e38414_fig53.png" xlink:type="simple" mimetype="image"/>
        </disp-formula>
        <disp-formula>
          <inline-graphic xlink:href="medinform_v10i9e38414_fig54.png" xlink:type="simple" mimetype="image"/>
        </disp-formula>
        <p>In the equations, &#124;Γ<sub>t</sub>&#124; is the size of testing triple set Γ<sub>t</sub> and I(·) is an indicator function, while <inline-graphic xlink:href="medinform_v10i9e38414_fig55.png" xlink:type="simple" mimetype="image"/> denote values of ranks for a head and a tail entity e<sub>i</sub> and e<sub>j</sub>, respectively.</p>
      </sec>
      <sec>
        <title>Model Performances</title>
        <p>During the comparison, we evaluated the models with embedding vectors of 256, 512, 1024, and 2048 dimensions and sufficient iterations to ensure the obtained embeddings were qualified for the sake of the downstream task. The results are shown in <xref ref-type="table" rid="table6">Tables 6</xref> and <xref ref-type="table" rid="table7">7</xref>. Compared with the baseline models, the CoKE model showed a competitive performance on both the standard data set and the constructed TCMdt data set. The CoKE model had the highest MRR and CoKE-multihop model had the best Hits@10. The CoKE-multihop-distillation model still showed a competitive performance on the MRR and HIT@10 compared to the CoKE model.</p>
        <p>To evaluate the effectiveness of the KGE learning, 10-fold cross-validation was used in the multilabel modeling task experiments. Compared with typical models multilabel k nearest neighbors (MLKNN), RandomForest-RAkEL (where RAkEL refers to random k-labelsets), LogisticRegression-RAkEL, and deep neural network (DNN) [<xref ref-type="bibr" rid="ref42">42</xref>], the proposed model outperformed the baseline models on metrics precision, recall, and the F1 score, as shown in <xref ref-type="table" rid="table8">Table 8</xref>. In addition, multilabel models with KGE had better performance than those without KGE. The results demonstrate that learned KGE is capable of improving the performance of deep learning models.</p>
        <p>As shown in <xref rid="figure4" ref-type="fig">Figure 4</xref>, the DNN+BILSTM-KGE (where BILSTM refers to bidirectional long short-term memory) outperformed the DNN on evaluation metrics (eg, precision and F1 score) in the training procedure. Compared with the DNN, the average precision and F1 score of DNN+BILSTM-KGE showed improvement, with the Hamming loss significantly decreasing for the first 50 iterations.</p>
        <table-wrap position="float" id="table6">
          <label>Table 6</label>
          <caption>
            <p>Performance comparison of link prediction on the FB15k-237 data set.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="400"/>
            <col width="150"/>
            <col width="150"/>
            <col width="150"/>
            <col width="150"/>
            <thead>
              <tr valign="top">
                <td>Models</td>
                <td>MRR<sup>a</sup></td>
                <td colspan="3">Hits@N</td>
              </tr>
              <tr valign="top">
                <td colspan="2">
                  <break/>
                </td>
                <td>@10</td>
                <td>@3</td>
                <td>@1</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>TransE</td>
                <td>0.296</td>
                <td>0.499</td>
                <td>0.330</td>
                <td>0.196</td>
              </tr>
              <tr valign="top">
                <td>SimplE</td>
                <td>0.306</td>
                <td>0.496</td>
                <td>0.341</td>
                <td>0.212</td>
              </tr>
              <tr valign="top">
                <td>RotatE</td>
                <td>0.314</td>
                <td>0.505</td>
                <td>0.347</td>
                <td>0.221</td>
              </tr>
              <tr valign="top">
                <td>ComplEx</td>
                <td>0.296</td>
                <td>0.489</td>
                <td>0.333</td>
                <td>0.200</td>
              </tr>
              <tr valign="top">
                <td>DistMult</td>
                <td>0.309</td>
                <td>0.506</td>
                <td>0.346</td>
                <td>0.211</td>
              </tr>
              <tr valign="top">
                <td>KBAT<sup>b</sup></td>
                <td>0.103</td>
                <td>0.337</td>
                <td>0.248</td>
                <td>0.103</td>
              </tr>
              <tr valign="top">
                <td>ConvKB</td>
                <td>0.407</td>
                <td>0.527</td>
                <td>0.333</td>
                <td>0.200</td>
              </tr>
              <tr valign="top">
                <td>CoKE<sup>c</sup></td>
                <td>0.362</td>
                <td>0.550</td>
                <td>0.400</td>
                <td>0.269</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table6fn1">
              <p><sup>a</sup>MSE: mean-square error.</p>
            </fn>
            <fn id="table6fn2">
              <p><sup>b</sup>KBAT: knowledge base attention.</p>
            </fn>
            <fn id="table6fn3">
              <p><sup>c</sup>CoKE: Contextualized Knowledge Graph Embedding.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <table-wrap position="float" id="table7">
          <label>Table 7</label>
          <caption>
            <p>Performance comparison of link prediction on the TCMdt data set.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="400"/>
            <col width="150"/>
            <col width="150"/>
            <col width="150"/>
            <col width="150"/>
            <thead>
              <tr valign="top">
                <td>Models</td>
                <td>MRR<sup>a</sup></td>
                <td colspan="3">Hits@N</td>
              </tr>
              <tr valign="top">
                <td colspan="2">
                  <break/>
                </td>
                <td>@10</td>
                <td>@3</td>
                <td>@1</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>TransE</td>
                <td>0.243</td>
                <td>0.428</td>
                <td>0.279</td>
                <td>0.150</td>
              </tr>
              <tr valign="top">
                <td>SimplE</td>
                <td>0.162</td>
                <td>0.436</td>
                <td>0.222</td>
                <td>0.113</td>
              </tr>
              <tr valign="top">
                <td>RotatE</td>
                <td>0.146</td>
                <td>0.424</td>
                <td>0.193</td>
                <td>0.090</td>
              </tr>
              <tr valign="top">
                <td>ComplEx</td>
                <td>0.137</td>
                <td>0.411</td>
                <td>0.177</td>
                <td>0.080</td>
              </tr>
              <tr valign="top">
                <td>DistMult</td>
                <td>0.164</td>
                <td>0.438</td>
                <td>0.223</td>
                <td>0.117</td>
              </tr>
              <tr valign="top">
                <td>ConvKB</td>
                <td>0.271</td>
                <td>0.464</td>
                <td>0.302</td>
                <td>0.192</td>
              </tr>
              <tr valign="top">
                <td>CoKE<sup>b</sup></td>
                <td>0.332</td>
                <td>0.491</td>
                <td>0.365</td>
                <td>0.250</td>
              </tr>
              <tr valign="top">
                <td>KBAT<sup>c</sup></td>
                <td>0.129</td>
                <td>0.369</td>
                <td>0.178</td>
                <td>0.088</td>
              </tr>
              <tr valign="top">
                <td>CoKE-multihop</td>
                <td>0.251</td>
                <td>0.515</td>
                <td>0.278</td>
                <td>0.261</td>
              </tr>
              <tr valign="top">
                <td>CoKE-multihop-distillation</td>
                <td>0.32</td>
                <td>0.483</td>
                <td>0.374</td>
                <td>0.260</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table7fn1">
              <p><sup>a</sup>MSE: mean-square error.</p>
            </fn>
            <fn id="table7fn2">
              <p><sup>b</sup>KBAT: knowledge base attention.</p>
            </fn>
            <fn id="table7fn3">
              <p><sup>c</sup>CoKE: Contextualized Knowledge Graph Embedding.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <table-wrap position="float" id="table8">
          <label>Table 8</label>
          <caption>
            <p>Results of 10-fold cross-validation of deep learning multilabel models.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="30"/>
            <col width="470"/>
            <col width="170"/>
            <col width="170"/>
            <col width="160"/>
            <thead>
              <tr valign="top">
                <td colspan="2">Index</td>
                <td>Precision</td>
                <td>Recall</td>
                <td>F1 score</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td colspan="5">
                  <bold>MLKNN<sup>a</sup> (Hamming loss=0.186; best parameter: K=26)</bold>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Micro-avg</td>
                <td>0.810</td>
                <td>0.710</td>
                <td>0.760</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Macro-avg</td>
                <td>0.800</td>
                <td>0.610</td>
                <td>0.660</td>
              </tr>
              <tr valign="top">
                <td colspan="5">
                  <bold>RandomForest-RAkEL<sup>b</sup> (Hamming loss=0.186; best parameter: n_estimators=800)</bold>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Micro-avg</td>
                <td>0.790</td>
                <td>0.740</td>
                <td>0.760</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Macro-avg</td>
                <td>0.760</td>
                <td>0.640</td>
                <td>0.670</td>
              </tr>
              <tr valign="top">
                <td colspan="5">
                  <bold>LogisticRegression-RAkEL (Hamming loss=0.173; best parameter: C=0.5)</bold>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Micro-avg</td>
                <td>0.810</td>
                <td>0.750</td>
                <td>0.780</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Macro-avg</td>
                <td>0.760</td>
                <td>0.660</td>
                <td>0.700</td>
              </tr>
              <tr valign="top">
                <td colspan="5">
                  <bold>DNN<sup>c</sup> (Hamming loss=0.186; best parameters: hidden=500, layer=3)</bold>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Micro-avg</td>
                <td>0.790</td>
                <td>0.740</td>
                <td>0.760</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Macro-avg</td>
                <td>0.750</td>
                <td>0.670</td>
                <td>0.700</td>
              </tr>
              <tr valign="top">
                <td colspan="5">
                  <bold>DNN+LSTM<sup>d</sup>-KGE<sup>e</sup> (Hamming loss=0.167; best parameters: hidden=500, layer=3, LSTM=128)</bold>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Micro-avg</td>
                <td>0.800</td>
                <td>0.790</td>
                <td>0.790</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Macro-avg</td>
                <td>0.740</td>
                <td>0.740</td>
                <td>0.740</td>
              </tr>
              <tr valign="top">
                <td colspan="5">
                  <bold>DNN+BILSTM<sup>f</sup>-KGE (Hamming loss=0.127; best parameter: LSTM=128)</bold>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Micro-avg</td>
                <td>0.860</td>
                <td>0.820</td>
                <td>0.840</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Macro-avg</td>
                <td>0.810</td>
                <td>0.770</td>
                <td>0.790</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table8fn1">
              <p><sup>a</sup>MLKNN: multilabel k nearest neighbors.</p>
            </fn>
            <fn id="table8fn2">
              <p><sup>b</sup>RAkEL: random k-labelsets.</p>
            </fn>
            <fn id="table8fn3">
              <p><sup>c</sup>DNN: deep neural network.</p>
            </fn>
            <fn id="table8fn4">
              <p><sup>d</sup>LSTM: long short-term memory.</p>
            </fn>
            <fn id="table8fn5">
              <p><sup>e</sup>KGE: knowledge graph embedding.</p>
            </fn>
            <fn id="table8fn6">
              <p><sup>f</sup>BILSTM: bidirectional long short-term memory.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <fig id="figure4" position="float">
          <label>Figure 4</label>
          <caption>
            <p>Performances of DNN and DNN+BILSTM-KGE. BILSTM: bidirectional long short-term memory; DNN: deep neural network; KGE: knowledge graph embedding.</p>
          </caption>
          <graphic xlink:href="medinform_v10i9e38414_fig4.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <p>Learned representations of entities were visualized by t-SNE, as shown in <xref rid="figure5" ref-type="fig">Figure 5</xref>. Symptoms and TCM syndrome elements are denoted by ○ and X, respectively. The representation distribution conformed to theoretical common sense in TCM with obvious boundaries (ie, silhouette score&#62;0.44) between different classes of TCM syndromes. Intuitively, the learned representations preserved the semantic information about TCM syndromes by using the proposed KGE learning methods. In addition, the relation between entities <italic>Yang hyperactivity</italic> and <italic>dizziness</italic> was similar to the relation between entities <italic>liver depression</italic> and <italic>stringy pulse</italic>, indicating that the semantic constraint of translational distance is preserved after training. The results show that representations learned by the proposed KGE learning method are capable of providing semantic information in TCM.</p>
        <fig id="figure5" position="float">
          <label>Figure 5</label>
          <caption>
            <p>Learned representations of entity visualization.</p>
          </caption>
          <graphic xlink:href="medinform_v10i9e38414_fig5.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <sec>
        <title>Principal Findings</title>
        <p>The experiments show that the CoKE model has a more stable performance and can be used for improving downstream tasks. We assume that downstream tasks may be improved by KGE learning, since semantic information provided by KGE is preserved in learned representations of missing entities and relations in a KGC task. KGE is suitable to be applied in scenarios that suffer from incompleteness issues, including knowledge discovery for diagnosis and treatment and assisted decision-making in TCM. Based on the clinical KGE model, we automatically extracted the information about dominant diseases treated by Chinese physicians, evidence, symptoms, theories, treatment methods, prescriptions, medicines, and concept mappings according to the definition of clinical knowledge ontology by the physicians. Inspired by Luo et al [<xref ref-type="bibr" rid="ref43">43</xref>] and Jin et al [<xref ref-type="bibr" rid="ref44">44</xref>], the triples in a clinical KG are used to learn a personalized KGE model of Chinese physicians.</p>
        <p>The problem of incompleteness of a KG is alleviated by entity link prediction of the personalized KGE model. Through the visualization of the KG, our system assists experts in identifying and expanding the potential relations and neighbors of entities in order to obtain explicitness of the implicit knowledge. Through multiple iterations of embedded learning, the KGE model is suitable for treatment decision-making of Chinese physicians. The theories, treatment methods, prescriptions, capability of cause-effect reasoning, and interpretability are enhanced.</p>
        <p>Consisting of theories, treatment methods, prescriptions, and medicines of endometriosis (EM) in TCM, the visualization of our KG is shown in <xref rid="figure6" ref-type="fig">Figure 6</xref>. A personalized KG for gynecology is constructed to assist experts in knowledge discovery and decision-making. The thickness of the arrows represents the strength of the potential causality, and the size of the nodes represents their importance in the KG of EM in gynecology. Our system clusters the nodes and represents them with different colors of the clusters. Different shapes of nodes represent different entity types.</p>
        <p>We referred to a large amount of ancient and modern literature and the diagnosis and treatment data of Chinese and Western medicine, combined with the techniques of entity extraction and causality extraction in natural language processing. According to the definition of domain knowledge by Chinese physicians, valid entities and relations from real cases include the names of TCM diseases, Chinese medicines and prescriptions, tests and examinations, names of Western medicines and diseases, TCM symptoms, and hospital departments. In the training procedure, the weights of the CoKE model were updated until convergence in order to generate embedding vectors that captured semantic features for clinical interpretability. The proposed model can be applied for personalized recommendations of Chinese physicians, question answering, and optimization of diagnostic models.</p>
        <p>Inspired by the heterogeneous network representation learning model [<xref ref-type="bibr" rid="ref45">45</xref>], a framework for knowledge discovery and decision-making in TCM was proposed, as shown in <xref rid="figure7" ref-type="fig">Figure 7</xref>.</p>
        <fig id="figure6" position="float">
          <label>Figure 6</label>
          <caption>
            <p>Visualization of a personalized KG that consists of theories, treatment methods, prescriptions, and medicines of EM in TCM. EM: endometriosis; KG: knowledge graph; TCM: traditional Chinese medicine.</p>
          </caption>
          <graphic xlink:href="medinform_v10i9e38414_fig6.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure7" position="float">
          <label>Figure 7</label>
          <caption>
            <p>Application of the framework to knowledge discovery and decision-making in TCM. CKG: collaborative knowledge graph; TCM: traditional Chinese medicine; QA: question and answer.</p>
          </caption>
          <graphic xlink:href="medinform_v10i9e38414_fig7.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <p>For medical recommendation and assisted decision-making, the first step is to collect objective information about the four diagnostic methods. The clinical KG incorporates multimodal information recognized from tongue and facial diagnosis equipment, which can be used to improve the performance of models, even in few-shot learning scenarios. KGs can be used to effectively solve the problems of sparsity and cold start in recommendation systems. Integrating KGs into recommendation systems as external information facilitates the systems with common-sense reasoning capability. Based on the powerful capability of information aggregation and the inference of GNNs, we designed a recommendation system to recommend symptoms, diseases, and Chinese physicians, which effectively improves the performance of recommendations. In addition, the information propagation and inference capability of GNNs also provide interpretability for the results of recommendations.</p>
        <p>The model can be used for high-quality assisted decision-making in diagnosis and treatment based on multimodal information and specialty questionnaires. Our system helps practitioners and patients efficiently build online profiles, which enhances the research value of clinical cases. Constructed from natural language, KGs have a strong connection to text mining. KGE can be used to boost the performance of models for text classification and generation. For example, KGE can be leveraged for entity disambiguation when answering the question of what glucose-lowering drug is better for obese diabetics. Similar to link prediction, knowledge inference in question answering infers new relations between entities, given a KG, which is often a multihop relation inference process. For instance, the question can be viewed as a query <inline-graphic xlink:href="medinform_v10i9e38414_fig56.png" xlink:type="simple" mimetype="image"/> which can be predicted by PathQuery answering of CoKE for medicine recommendation to obtain related medicines, including <italic>metformin</italic> [<xref ref-type="bibr" rid="ref46">46</xref>-<xref ref-type="bibr" rid="ref49">49</xref>].</p>
      </sec>
      <sec>
        <title>Conclusion</title>
        <p>In this paper, a KG-fused multihop relational adaptive CoKE framework was proposed for screening enhancement, knowledge complement, knowledge inference, and knowledge distillation. The superiority of the model in knowledge discovery and assisted decision-making in TCM was shown in experiments and clinical practice. TCM is a systematic discipline focusing on inheritance and practice. A large amount of knowledge is hidden in the ancient literature and experimental cases of Chinese physicians, which can be mined by researchers. In the future, we aim to improve the quality of the intelligent system of human-machine collaborative KGs in TCM. More in-depth research will be conducted on the knowledge fusion of heterogeneous GNNs, complex inference of KGs with GNNs, and interpretable learning of GNNs.</p>
      </sec>
    </sec>
  </body>
  <back>
    <app-group/>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">ADRD</term>
          <def>
            <p>Alzheimer disease and related dementia</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">AI</term>
          <def>
            <p>artificial intelligence</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">BILSTM</term>
          <def>
            <p>bidirectional long short-term memory</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">CKD</term>
          <def>
            <p>chronic kidney disease</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">CoKE</term>
          <def>
            <p>Contextualized Knowledge Graph Embedding</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">DNN</term>
          <def>
            <p>deep neural network</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb7">EM</term>
          <def>
            <p>endometriosis</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb8">EMR</term>
          <def>
            <p>electronic medical record</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb9">GNN</term>
          <def>
            <p>graph neural network</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb10">KBAT</term>
          <def>
            <p>knowledge base attention</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb11">KG</term>
          <def>
            <p>knowledge graph</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb12">KGC</term>
          <def>
            <p>knowledge graph completion</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb13">KGE</term>
          <def>
            <p>knowledge graph embedding</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb14">MLKNN</term>
          <def>
            <p>multilabel k nearest neighbors</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb15">MRR</term>
          <def>
            <p>mean reciprocal rank</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb16">MSE</term>
          <def>
            <p>mean-square error</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb17">RAkEL</term>
          <def>
            <p>random k-labelsets</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb18">TCM</term>
          <def>
            <p>traditional Chinese medicine</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb19">TCMLS</term>
          <def>
            <p>Traditional Chinese Medicine Language System</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <ack>
      <p>The work was supported by grants from the National Natural Science Foundation of China (#61871141) research and development plan projects in key areas of Guangdong Province (#2021A1111120008), the Collaborative Innovation Team of Guangzhou University of Traditional Chinese Medicine (#2021XK08), and the State Key Laboratory of Dampness Syndrome of Chinese Medicine (#SZ2021ZZ3004, SZ2021ZZ01).</p>
    </ack>
    <fn-group>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Du</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Shi</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>The advantages of TCM in treating chronic diseases and the inspiration of TCM to modern medical treatment model</article-title>
          <source>Beijing J Tradit Chin Med (in Chinese)</source>
          <year>2010</year>
          <volume>29</volume>
          <issue>4</issue>
          <fpage>3</fpage>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Weng</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Yan</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>A framework for automated knowledge graph construction towards traditional chinese medicine</article-title>
          <source>Health Info Sci</source>
          <year>2017</year>
          <fpage>170</fpage>
          <pub-id pub-id-type="doi">10.1007/978-3-319-69182-4_18</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="web">
          <source>Google Knowledge Graph Search API</source>
          <access-date>2022-08-09</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://developers.google.com/knowledge-graph">https://developers.google.com/knowledge-graph</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Paulheim</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>Data-driven joint debugging of the dbpedia mappings and ontology</article-title>
          <year>2017</year>
          <conf-name>14th International European Semantic Web Conference</conf-name>
          <conf-date>May 28 to June 1, 2017</conf-date>
          <conf-loc>Portorož, Slovenia</conf-loc>
          <fpage>404</fpage>
          <lpage>418</lpage>
          <pub-id pub-id-type="doi">10.1007/978-3-319-58068-5_25</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Suchanek</surname>
              <given-names>FM</given-names>
            </name>
            <name name-style="western">
              <surname>Kasneci</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Weikum</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>Yago: a core of semantic knowledge</article-title>
          <year>2007</year>
          <conf-name>16th International Conference on World Wide Web</conf-name>
          <conf-date>May 8-12, 2007</conf-date>
          <conf-loc>Banff, Alberta, Canada</conf-loc>
          <fpage>697</fpage>
          <lpage>706</lpage>
          <pub-id pub-id-type="doi">10.1145/1242572.1242667</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bollacker</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Evans</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Paritosh</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>Freebase: a collaboratively created graph database for structuring human knowledge</article-title>
          <year>2018</year>
          <conf-name>SIGMOD/PODS '08: International Conference on Management of Data</conf-name>
          <conf-date>June 9-12, 2008</conf-date>
          <conf-loc>Vancouver Canada</conf-loc>
          <fpage>1247</fpage>
          <lpage>1250</lpage>
          <pub-id pub-id-type="doi">10.1145/1376616.1376746</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Cui</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <source>Big Data Intelligence: Machine Learning and Natural Language Processing in the Internet Age</source>
          <year>2016</year>
          <publisher-loc>Beijing</publisher-loc>
          <publisher-name>Publishing House of Electronics Industry</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Cheng</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Jin</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Guo</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>Survey on big data system and analytic technology</article-title>
          <source>J Softw</source>
          <year>2014</year>
          <issue>9</issue>
          <fpage>1889</fpage>
          <lpage>1908</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhou</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Cao</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <source>Medical Knowledge Acquisition: An Ontology - Based Approach</source>
          <year>2003</year>
          <publisher-loc>China</publisher-loc>
          <publisher-name>Key Laboratory of Intelligent Information Processing, Institute of Computing Technology, Chinese Academy of Sciences</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jia</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Yu</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Gao</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>Construction of traditional chinese medicine knowledge graph</article-title>
          <source>J Med Inform</source>
          <year>2015</year>
          <fpage>51</fpage>
          <lpage>53</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jia</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Dong</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Study and establishment of appraisal system for traditional chinese medicine language system</article-title>
          <source>Chin Digit Med</source>
          <year>2012</year>
          <volume>07</volume>
          <issue>010</issue>
          <fpage>13</fpage>
          <lpage>16</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Rotmensch</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Halpern</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Tlimat</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Horng</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Sontag</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Learning a health knowledge graph from electronic medical records</article-title>
          <source>Sci Rep</source>
          <year>2017</year>
          <month>07</month>
          <day>20</day>
          <volume>7</volume>
          <issue>1</issue>
          <fpage>5994</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41598-017-05778-z"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41598-017-05778-z</pub-id>
          <pub-id pub-id-type="medline">28729710</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41598-017-05778-z</pub-id>
          <pub-id pub-id-type="pmcid">PMC5519723</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bordes</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Usunier</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Garcia-Duran</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Translating embeddings for modeling multi-relational data</article-title>
          <source>Adv Neural Info Proc Syst</source>
          <year>2013</year>
          <fpage>26</fpage>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Feng</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>Z</given-names>
            </name>
          </person-group>
          <article-title>Knowledge graph embedding by translating on hyperplanes</article-title>
          <source>AAAI</source>
          <year>2014</year>
          <month>06</month>
          <day>21</day>
          <volume>28</volume>
          <issue>1</issue>
          <fpage>1112</fpage>
          <lpage>1119</lpage>
          <pub-id pub-id-type="doi">10.1609/aaai.v28i1.8870</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>X</given-names>
            </name>
          </person-group>
          <article-title>Learning entity and relation embeddings for knowledge graph completion</article-title>
          <source>AAAI</source>
          <year>2015</year>
          <month>02</month>
          <day>19</day>
          <volume>29</volume>
          <issue>1</issue>
          <fpage>2181</fpage>
          <lpage>2187</lpage>
          <pub-id pub-id-type="doi">10.1609/aaai.v29i1.9491</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Xiao</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Huang</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>X</given-names>
            </name>
          </person-group>
          <article-title>From one point to a manifold: knowledge graph embedding for precise link prediction</article-title>
          <year>2015</year>
          <conf-name>25th International Joint Conference on Artificial Intelligence</conf-name>
          <conf-date>July 9-15, 2016</conf-date>
          <conf-loc>New York, NY</conf-loc>
          <fpage>1315</fpage>
          <lpage>1321</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ji</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>He</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Knowledge graph embedding via dynamic mapping matrix</article-title>
          <year>2015</year>
          <conf-name>53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing</conf-name>
          <conf-date>July 2015</conf-date>
          <conf-loc>Beijing, China</conf-loc>
          <fpage>687</fpage>
          <lpage>397</lpage>
          <pub-id pub-id-type="doi">10.3115/v1/p15-1067</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Nickel</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Tresp</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Kriegel</surname>
              <given-names>HP</given-names>
            </name>
          </person-group>
          <article-title>A three-way model for collective learning on multi-relational data</article-title>
          <year>2011</year>
          <conf-name>28th International Conference on Machine Learning</conf-name>
          <conf-date>June 28 to July 2, 2011</conf-date>
          <conf-loc>Bellevue, WA</conf-loc>
          <fpage>809</fpage>
          <lpage>816</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Yih</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>He</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Gao</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Deng</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>Embedding entities and relations for learning and inference in knowledge bases</article-title>
          <source>ICLR</source>
          <year>2015</year>
          <fpage>13</fpage>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Nickel</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Rosasco</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Poggio</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Holographic embeddings of knowledge graphs</article-title>
          <source>AAAI</source>
          <year>2016</year>
          <month>03</month>
          <day>02</day>
          <volume>30</volume>
          <issue>1</issue>
          <fpage>1955</fpage>
          <lpage>1961</lpage>
          <pub-id pub-id-type="doi">10.1609/aaai.v30i1.10314</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Trouillon</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Welbl</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Riedel</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Gaussier</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Bouchard</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>Complex embeddings for simple link prediction</article-title>
          <source>ICML</source>
          <year>2016</year>
          <fpage>2071</fpage>
          <lpage>2080</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kazemi</surname>
              <given-names>SM</given-names>
            </name>
            <name name-style="western">
              <surname>Poole</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Simple embedding for link prediction in knowledge graphs</article-title>
          <year>2018</year>
          <conf-name>NeurIPS 2018: Annual Conference on Neural Information Processing Systems</conf-name>
          <conf-date>December 3-8, 2018</conf-date>
          <conf-loc>Montréal, Canada</conf-loc>
          <fpage>4284</fpage>
          <lpage>4295</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Dettmers</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Minervini</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Stenetorp</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Riedel</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Convolutional 2D knowledge graph embeddings</article-title>
          <year>2018</year>
          <month>04</month>
          <day>25</day>
          <conf-name>AAAI-18: Thirty-Second AAAI Conference on Artificial Intelligence</conf-name>
          <conf-date>February 2-7, 2018</conf-date>
          <conf-loc>New Orleans, LA</conf-loc>
          <fpage>1811</fpage>
          <lpage>1818</lpage>
          <pub-id pub-id-type="doi">10.1609/aaai.v32i1.11573</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Nguyen</surname>
              <given-names>DQ</given-names>
            </name>
            <name name-style="western">
              <surname>Nguyen</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>NguyenPhung</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>A novel embedding model for knowledge base completion based on convolutional neural network</article-title>
          <year>2018</year>
          <conf-name>2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</conf-name>
          <conf-date>June 1-6, 2018</conf-date>
          <conf-loc>New Orleans, LA</conf-loc>
          <fpage>327</fpage>
          <lpage>333</lpage>
          <pub-id pub-id-type="doi">10.18653/v1/n18-2053</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Vu</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Nguyen</surname>
              <given-names>TD</given-names>
            </name>
            <name name-style="western">
              <surname>Nguyen</surname>
              <given-names>DQ</given-names>
            </name>
          </person-group>
          <article-title>A capsule network-based embedding model for knowledge graph completion and search personalization</article-title>
          <year>2019</year>
          <conf-name>2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</conf-name>
          <conf-date>June 1-8, 2019</conf-date>
          <conf-loc>Minneapolis, MN</conf-loc>
          <fpage>2180</fpage>
          <lpage>2189</lpage>
          <pub-id pub-id-type="doi">10.18653/v1/n19-1226</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Nathani</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Chauhan</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Sharma</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Learning attention-based embeddings for relation prediction in knowledge graphs</article-title>
          <year>2019</year>
          <conf-name>57th Annual Meeting of the Association for Computational Linguistics</conf-name>
          <conf-date>July 28 to August 2, 2019</conf-date>
          <conf-loc>Florence, Italy</conf-loc>
          <fpage>4710</fpage>
          <lpage>4723</lpage>
          <pub-id pub-id-type="doi">10.18653/v1/p19-1466</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref27">
        <label>27</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Devlin</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Toutanova</surname>
              <given-names>LK</given-names>
            </name>
          </person-group>
          <article-title>BERT: pre-training of deep bidirectional transformers for language understanding</article-title>
          <year>2019</year>
          <conf-name>2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</conf-name>
          <conf-date>June 1-8, 2019</conf-date>
          <conf-loc>Minneapolis, MN</conf-loc>
          <fpage>4171</fpage>
          <lpage>4186</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref28">
        <label>28</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Vaswani</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Shazeer</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Parmar</surname>
              <given-names>N</given-names>
            </name>
          </person-group>
          <article-title>Attention is all you need</article-title>
          <year>2017</year>
          <conf-name>31st International Conference on Neural Information Processing Systems</conf-name>
          <conf-date>December 4-9, 2017</conf-date>
          <conf-loc>Long Beach, CA</conf-loc>
          <fpage>6000</fpage>
          <lpage>6010</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref29">
        <label>29</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Huang</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>Coke: Contextualized knowledge graph embedding</article-title>
          <source>arXiv</source>
          <year>2019</year>
          <fpage>2168</fpage>
          <pub-id pub-id-type="doi">10.1090/mbk/121/79</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref30">
        <label>30</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Qian</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Fu</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Cai</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>He</surname>
              <given-names>X</given-names>
            </name>
          </person-group>
          <article-title>Translating embeddings for knowledge graph completion with relation attention mechanism</article-title>
          <year>2018</year>
          <conf-name>Twenty-Seventh International Joint Conference on Artificial Intelligence</conf-name>
          <conf-date>July 13-19, 2018</conf-date>
          <conf-loc>Stockholm, Sweden</conf-loc>
          <fpage>4286</fpage>
          <lpage>4292</lpage>
          <pub-id pub-id-type="doi">10.24963/ijcai.2018/596</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref31">
        <label>31</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Shen</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Yuan</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Dai</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Tang</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>Min</given-names>
            </name>
            <name name-style="western">
              <surname>Lei</surname>
              <given-names>Kai</given-names>
            </name>
          </person-group>
          <article-title>KGDDS: a system for drug-drug similarity measure in therapeutic substitution based on knowledge graph curation</article-title>
          <source>J Med Syst</source>
          <year>2019</year>
          <month>03</month>
          <day>05</day>
          <volume>43</volume>
          <issue>4</issue>
          <fpage>92</fpage>
          <pub-id pub-id-type="doi">10.1007/s10916-019-1182-z</pub-id>
          <pub-id pub-id-type="medline">30834481</pub-id>
          <pub-id pub-id-type="pii">10.1007/s10916-019-1182-z</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref32">
        <label>32</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zheng</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Yan</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Gou</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>Z-C</given-names>
            </name>
            <name name-style="western">
              <surname>Jason Zhang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Hu</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>F-Y</given-names>
            </name>
          </person-group>
          <article-title>Pay attention to doctor-patient dialogues: multi-modal knowledge graph attention image-text embedding for COVID-19 diagnosis</article-title>
          <source>Inf Fusion</source>
          <year>2021</year>
          <month>11</month>
          <volume>75</volume>
          <fpage>168</fpage>
          <lpage>185</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/34093095"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.inffus.2021.05.015</pub-id>
          <pub-id pub-id-type="medline">34093095</pub-id>
          <pub-id pub-id-type="pii">S1566-2535(21)00115-9</pub-id>
          <pub-id pub-id-type="pmcid">PMC8168340</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref33">
        <label>33</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Li</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Maharjan</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Xie</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Tao</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>A personalized voice-based diet assistant for caregivers of Alzheimer disease and related dementias: system development and validation</article-title>
          <source>J Med Internet Res</source>
          <year>2020</year>
          <month>09</month>
          <day>21</day>
          <volume>22</volume>
          <issue>9</issue>
          <fpage>e19897</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmir.org/2020/9/e19897/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/19897</pub-id>
          <pub-id pub-id-type="medline">32955452</pub-id>
          <pub-id pub-id-type="pii">v22i9e19897</pub-id>
          <pub-id pub-id-type="pmcid">PMC7536606</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref34">
        <label>34</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Luan</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>Modeling relation paths for representation learning of knowledge bases</article-title>
          <year>2015</year>
          <conf-name>Conference on Empirical Methods in Natural Language Processing</conf-name>
          <conf-date>September 17-21, 2015</conf-date>
          <conf-loc>Lisbon, Portugal</conf-loc>
          <fpage>705</fpage>
          <lpage>714</lpage>
          <pub-id pub-id-type="doi">10.18653/v1/d15-1082</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref35">
        <label>35</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kadlec</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Bajgar</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Kleindienst</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Knowledge base completion: baselines strike back</article-title>
          <year>2017</year>
          <conf-name>2nd Workshop on Representation Learning for NLP</conf-name>
          <conf-date>August 2017</conf-date>
          <conf-loc>Vancouver, Canada</conf-loc>
          <fpage>69</fpage>
          <lpage>74</lpage>
          <pub-id pub-id-type="doi">10.18653/v1/w17-2609</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref36">
        <label>36</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Guu</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Miller</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Liang</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>Traversing knowledge graphs in vector space</article-title>
          <year>2015</year>
          <conf-name>Conference on Empirical Methods in Natural Language Processing</conf-name>
          <conf-date>September 2015</conf-date>
          <conf-loc>Lisbon, Portugal</conf-loc>
          <fpage>318</fpage>
          <lpage>327</lpage>
          <pub-id pub-id-type="doi">10.18653/v1/d15-1038</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref37">
        <label>37</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jiao</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Yin</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Shang</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Jiang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Q</given-names>
            </name>
          </person-group>
          <article-title>TinyBERT: distilling BERT for natural language understanding</article-title>
          <source>Findings of the Association for Computational Linguistics</source>
          <year>2020</year>
          <publisher-loc>Stroudsburg, PA</publisher-loc>
          <publisher-name>Association for Computational Linguistics</publisher-name>
          <fpage>4163</fpage>
          <lpage>4174</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref38">
        <label>38</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Toutanova</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Observed versus latent features for knowledge base and text inference</article-title>
          <year>2015</year>
          <conf-name>3rd Workshop on Continuous Vector Space Models and their Compositionality</conf-name>
          <conf-date>2015</conf-date>
          <conf-loc>Beijing, China</conf-loc>
          <fpage>57</fpage>
          <lpage>66</lpage>
          <pub-id pub-id-type="doi">10.18653/v1/w15-4007</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref39">
        <label>39</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ou</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>LEVIS: a hypertension dataset in traditional Chinese medicine</article-title>
          <year>2013</year>
          <conf-name>IEEE International Conference on Bioinformatics and Biomedicine</conf-name>
          <conf-date>December 18-21, 2013</conf-date>
          <conf-loc>Shanghai, China</conf-loc>
          <fpage>192</fpage>
          <lpage>197</lpage>
          <pub-id pub-id-type="doi">10.1109/bibm.2013.6732672</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref40">
        <label>40</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Vashishth</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Sanyal</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>A re-evaluation of knowledge graph completion methods</article-title>
          <year>2020</year>
          <conf-name>58th Annual Meeting of the Association for Computational Linguistics</conf-name>
          <conf-date>July 5-10, 2020</conf-date>
          <conf-loc>Seattle, WA</conf-loc>
          <fpage>5516</fpage>
          <lpage>5522</lpage>
          <pub-id pub-id-type="doi">10.18653/v1/2020.acl-main.489</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref41">
        <label>41</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ruffinelli</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Broscheit</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Gemulla</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>You can teach an old dog new tricks! on training knowledge graph embeddings</article-title>
          <year>2019</year>
          <conf-name>7th International Conference on Learning Representations</conf-name>
          <conf-date>May 6-9, 2019</conf-date>
          <conf-loc>New Orleans, LA</conf-loc>
          <pub-id pub-id-type="doi">10.4324/9781315657691-66</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref42">
        <label>42</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Maxwell</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Weng</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Ou</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Hong</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Zhou</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Gong</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Deep learning architectures for multi-label classification of intelligent health risk prediction</article-title>
          <source>BMC Bioinform</source>
          <year>2017</year>
          <month>12</month>
          <day>28</day>
          <volume>18</volume>
          <issue>Suppl 14</issue>
          <fpage>523</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://bmcbioinformatics.biomedcentral.com/articles/10.1186/s12859-017-1898-z"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/s12859-017-1898-z</pub-id>
          <pub-id pub-id-type="medline">29297288</pub-id>
          <pub-id pub-id-type="pii">10.1186/s12859-017-1898-z</pub-id>
          <pub-id pub-id-type="pmcid">PMC5751777</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref43">
        <label>43</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Luo</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Hou</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Analysis of the law of Professor Yang Nizhi for diabetic kidney disease based on knowledge graph experimental mining</article-title>
          <source>Modernizat Tradit Chin Med Materia Med-World Sci Technol</source>
          <year>2020</year>
          <volume>22</volume>
          <issue>5</issue>
          <fpage>1464</fpage>
          <lpage>1471</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref44">
        <label>44</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jin</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>He</surname>
              <given-names>W</given-names>
            </name>
          </person-group>
          <article-title>An analysis of clinical characteristics and prescription patterns of Professor Zhang Zhongde</article-title>
          <source>Modernizat Tradit Chin Med Materia Med-World Sci Technol</source>
          <year>2021</year>
          <fpage>1</fpage>
          <lpage>11</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref45">
        <label>45</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Xiao</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Han</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Heterogeneous Network Representation Learning: A Unified Framework with Survey and Benchmark</article-title>
          <source>IEEE Transactions on Knowledge &#38; Data Engineering</source>
          <year>2020</year>
          <volume>01</volume>
          <fpage>1</fpage>
          <lpage>1</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref46">
        <label>46</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hu</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Shi</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Research progress of knowledge graph based on graph neural network</article-title>
          <source>Commun CCF</source>
          <year>2020</year>
          <volume>016</volume>
          <issue>008</issue>
          <fpage>38</fpage>
        </nlm-citation>
      </ref>
      <ref id="ref47">
        <label>47</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Qiu</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Shao</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Dai</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Huang</surname>
              <given-names>X</given-names>
            </name>
          </person-group>
          <article-title>Pre-trained models for natural language processing: a survey</article-title>
          <source>Sci Chin Technol Sci</source>
          <year>2020</year>
          <month>09</month>
          <day>15</day>
          <volume>63</volume>
          <issue>10</issue>
          <fpage>1872</fpage>
          <lpage>1897</lpage>
          <pub-id pub-id-type="doi">10.1007/s11431-020-1647-3</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref48">
        <label>48</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Du</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Wan</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Ji</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>A review of knowledge graph techniques from the view of geometric deep learning</article-title>
          <source>Aero Weaponry (in Chinese)</source>
          <year>2020</year>
          <volume>27</volume>
          <issue>3</issue>
          <fpage>1</fpage>
          <lpage>10</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref49">
        <label>49</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Guan</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Jin</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Jia</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Cheng</surname>
              <given-names>X</given-names>
            </name>
          </person-group>
          <article-title>Knowledge reasoning over knowledge graph: a survey</article-title>
          <source>j Softw</source>
          <year>2018</year>
          <volume>29</volume>
          <issue>10</issue>
          <fpage>2966</fpage>
          <lpage>2994</lpage>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
