Published on in Vol 8, No 8 (2020): August

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/16948, first published .
Semantic Deep Learning: Prior Knowledge and a Type of Four-Term Embedding Analogy to Acquire Treatments for Well-Known Diseases

Semantic Deep Learning: Prior Knowledge and a Type of Four-Term Embedding Analogy to Acquire Treatments for Well-Known Diseases

Semantic Deep Learning: Prior Knowledge and a Type of Four-Term Embedding Analogy to Acquire Treatments for Well-Known Diseases

Original Paper

1Department of Computer Science, University of Manchester, Manchester, United Kingdom

2Hospital do Salnés, Villagarcía de Arousa, Spain

3Departamento de Lingüística Aplicada a la Ciencia y a la Tecnología, Universidad Politécnica de Madrid, Madrid, Spain

4Salford Languages, University of Salford, Salford, United Kingdom

5School of Social Sciences, University of Manchester, Manchester, United Kingdom

6BMJ, London, United Kingdom

7Mid Cheshire Hospital Foundation Trust, NHS England, Crewe, United Kingdom

*all authors contributed equally

Corresponding Author:

Robert Stevens, DPhil

Department of Computer Science

University of Manchester

Kilburn Building, Oxford Road, M13 9PL

Manchester,

United Kingdom

Phone: 44 161 275 6251

Email: robert.stevens@manchester.ac.uk


Background: How to treat a disease remains to be the most common type of clinical question. Obtaining evidence-based answers from biomedical literature is difficult. Analogical reasoning with embeddings from deep learning (embedding analogies) may extract such biomedical facts, although the state-of-the-art focuses on pair-based proportional (pairwise) analogies such as man:woman::king:queen (“queen = −man +king +woman”).

Objective: This study aimed to systematically extract disease treatment statements with a Semantic Deep Learning (SemDeep) approach underpinned by prior knowledge and another type of 4-term analogy (other than pairwise).

Methods: As preliminaries, we investigated Continuous Bag-of-Words (CBOW) embedding analogies in a common-English corpus with five lines of text and observed a type of 4-term analogy (not pairwise) applying the 3CosAdd formula and relating the semantic fields person and death: “dagger = −Romeo +die +died” (search query: −Romeo +die +died). Our SemDeep approach worked with pre-existing items of knowledge (what is known) to make inferences sanctioned by a 4-term analogy (search query −x +z1 +z2) from CBOW and Skip-gram embeddings created with a PubMed systematic reviews subset (PMSB dataset). Stage1: Knowledge acquisition. Obtaining a set of terms, candidate y, from embeddings using vector arithmetic. Some n-gram pairs from the cosine and validated with evidence (prior knowledge) are the input for the 3cosAdd, seeking a type of 4-term analogy relating the semantic fields disease and treatment. Stage 2: Knowledge organization. Identification of candidates sanctioned by the analogy belonging to the semantic field treatment and mapping these candidates to unified medical language system Metathesaurus concepts with MetaMap. A concept pair is a brief disease treatment statement (biomedical fact). Stage 3: Knowledge validation. An evidence-based evaluation followed by human validation of biomedical facts potentially useful for clinicians.

Results: We obtained 5352 n-gram pairs from 446 search queries by applying the 3CosAdd. The microaveraging performance of MetaMap for candidate y belonging to the semantic field treatment was F-measure=80.00% (precision=77.00%, recall=83.25%). We developed an empirical heuristic with some predictive power for clinical winners, that is, search queries bringing candidate y with evidence of a therapeutic intent for target disease x. The search queries -asthma +inhaled_corticosteroids +inhaled_corticosteroid and -epilepsy +valproate +antiepileptic_drug were clinical winners, finding eight evidence-based beneficial treatments.

Conclusions: Extracting treatments with therapeutic intent by analogical reasoning from embeddings (423K n-grams from the PMSB dataset) is an ambitious goal. Our SemDeep approach is knowledge-based, underpinned by embedding analogies that exploit prior knowledge. Biomedical facts from embedding analogies (4-term type, not pairwise) are potentially useful for clinicians. The heuristic offers a practical way to discover beneficial treatments for well-known diseases. Learning from deep learning models does not require a massive amount of data. Embedding analogies are not limited to pairwise analogies; hence, analogical reasoning with embeddings is underexploited.

JMIR Med Inform 2020;8(8):e16948

doi:10.2196/16948

Keywords



How to treat a disease or condition remains to be the most common type of clinical question [1]. It is difficult for clinicians to obtain comprehensive information on the clinical (and economic) worth of alternative drug choices for a given condition [2]. Evidence-based biomedical literature, although available in electronic form, primarily remains to be expert-to-expert communication—natural language statements intended for human consumption.

Analogical reasoning is basic relational reasoning without explicit representations of relations [3]. An acknowledged semantic property of embeddings (ie, vectors representing terms) from deep learning [4] is “their ability to capture relational meanings” [5], the so-called analogies [6]. Current efforts in analogical reasoning with embeddings focus on pair-based proportional analogies [5,7,8]. This is a type of “the four-term analogy” [6], also known as the cross-mapping analogy [6]. An example is queen = −man +king +woman [9], also represented as man:woman::king:queen [10], and read as “man is to king as woman is to queen” [11]. Examples for health care include the following:

  • 'acetaminophen' is as type of 'drug' as 'diabetes' is as type of ‘disease’” [12].
  • (furosemide - kidney) + heart ~ fosinopril” [13].

This study aimed to investigate embedding analogies (analogical reasoning with embeddings) [5] that are not pair-based proportional (pairwise for short) analogies. This study began by observing senior clinicians performing an analogical reasoning for sepsis (a major life-threatening condition) with embeddings and posing search queries such as −sepsis +serum_albumin +fluid_therapy to discover treatments with therapeutic intent. The clinical rationale behind this query is that “current evidence suggests that resuscitation using albumin-containing solutions is safe” [14], where serum_albumin is a shortened form of “human serum albumin supplementation” (extensively debated for sepsis [15]). We viewed this as another type of the four-term analogy, which is not pairwise.

This paper presents a semiautomatic approach to extract meaning (semantics) from the unstructured free text of biomedical literature (ie, PubMed systematic reviews [16]). The disease treatment statements systematically acquired from analogical reasoning are biomedical facts validated with evidence first and human audit afterward. The approach presented belongs to Semantic Deep Learning (SemDeep) [17], as we used embedding analogies (other than pairwise) and semantic knowledge representation paradigms [18] to provide meaning for the same.

Analogical Reasoning

Humans possess the ability to reason by analogy using abstract semantic relations such as synonyms or category membership [3]. For example, common cold and influenza are both types of illnesses with some common symptoms such as runny nose, sore throat, cough, and headache. As they share some key characteristics, we can possibly say they are near-synonyms, although they cannot be used interchangeably (as synonyms would) because of key medical differences. Our SemDeep approach acquires terms about treatments for a well-known disease using analogical reasoning that is underpinned by Aristotle’s theory [19]:

  • The strength of an analogy depends upon the number of similarities” [19]. For example, “intravenous antibiotics” and “intravenous fluid resuscitation” are basic therapies that improve outcomes in patients with sepsis [14], that is, both are treatments with a therapeutic intent for sepsis. However, we cannot say that they are similar as “intravenous fluid resuscitation” is a procedure whereas an “intravenous antibiotic” is a substance, although both are “intravenous.”
  • Similarity reduces to identical properties and relations” [19]. For example, “benzyl penicillin,” “cefotaxime,” or “amoxicillin/clavulanate” is similar as they belong to the same category, “antipseudomonal beta-lactam antibiotics” [14].
  • "Good analogies derive from underlying common causes or general laws” [19]. This study investigated the systematic acquisition of treatments for a disease using the simple generic 3CosAdd formula [20,21].

The 3CosAdd Formula

Our work relied on vector semantics [5] and used the neural language models, Continuous Bag-of-Words (CBOW) and Skip-gram by Mikolov et al [20], from deep learning to create embeddings. Embeddings with vector semantics such as cosine or 3CosAdd can acquire a list of strings of characters (eg, n-grams), although they lack explicit semantic meaning. Until now, the 3CosAdd formula 1 [20,21] has been applied to analogies between 2 pairs of words a:a*::b:b* [7], where b* is the unknown (hidden) word.

We used the 3CosAdd formula as in Levy and Goldberg [21] with the rewording “find the term y, which is similar to the term z1 and the term z2, while different from the term x”, where the target term x provides the semantic context and similar refers to the terms sharing “commonalities in structural features.” In this study, a semantic field is a set of terms that “belong together under the same conceptual heading” [22] and is a form of knowledge representation that provides meaning to those terms. We rewrote the 3CosAdd formula as formula 2, with the search query -x +z1 +z2, and the type of 4-term analogy we sought had the following:

  • The target term x belonging to the semantic field disease and representing a medical diagnosis mappable to a “type of” of systematized nomenclature of medicine - clinical terms (SNOMED CT) [23] concept called disorder.
  • The 3 terms {z1; z2; y} belonging to the semantic field treatment(Tx for short), where Tx encompasses 3 textual definitions from Hart et al [24]. The candidate term y is the unknown.

The Research Questions

We adopted the view by Hill et al [25] by considering “relatedness” as “association” and synonymy as the strongest similarity. In this study, the association relationship of interest is “correlation,” as defined in the semantic science integrated ontology (SIO) [26].

As preliminaries, we asked 2 research questions not specific to the health care domain:

  • Q1: Can “good” embeddings be created with a small corpus?
  • Q2: If the simple generic 3CosAdd formula [20,21] can capture a type of 4-term analogy as read in formula 2, can they be observed in embeddings created with a small corpus?

Our third research question (Q3) asked whether the 4-term type of analogy discovered in a small common-English corpus can also be discovered in a larger-scale biomedical corpus. To provide proof of such a generalization, we performed a real-world test with embeddings created with free text from PubMed systematic reviews [16]. We postulated that candidate inferences can be validated using evidence-based information resources. This study investigated the discovery of clinical winners, that is, search queries -x +z1 +z2 bringing candidate treatments y with evidence of a therapeutic intent for target disease x; thus, enabling the most common type of clinical question, “how to treat a disease or condition” [1], to be answered.

Our final research question (Q4) asked for some predictive power over the clinical winners obtained (ie, an empirical heuristic) if our SemDeep approach worked, that is the type of analogy proposed finds disease treatment statements from PubMed systematic reviews (ie, a larger-scale biomedical corpus). This last question pursued a tacit preference and referred to the final characteristic of analogy: systematicity [6]. However, challenges have been acknowledged “for any vector space model that aims to make predictions about relational similarity” [27].

Between the semantic field disease and the semantic field treatment, “few maximal structurally consistent interpretations (ie, mappings displaying one-to-one correspondences and parallel connectivity)” [6] are to be expected. For example, aspirin treatment does not have a one-to-one correspondence with a disease as it can treat headache (common knowledge) and acute myocardial infarction [1]. In this study, “spontaneous unplanned inferences” [6] were also expected, and this propensity was captured with the notion of incremental mappings [6].


Overview

Our SemDeep approach answered Q3 and comprised the 3 stages depicted in Figure 1. The software package word2vec [28] implements the CBOW and Skip-gram algorithms along with the cosine and 3CosAdd formulas. The terms in this study are n-grams.

Stage 1 used prior knowledge (open-access reusable datasets [29]) consisting of n-gram pairs obtained by applying the cosine to embeddings, then mapped to the Unified Medical Language System (UMLS) Metathesaurus [30] concept pairs, and finally validated with evidence from biomedical literature using the British Medical Journal (BMJ) Best Practice [31] as the main information source.

BMJ Best Practice is separate from PubMed/MEDLINE [32] and is acknowledged for its editorial quality and evidence-based methodology [33]. In the United Kingdom, BMJ Best Practice is provided (free access) to all National Health Service (NHS) health care professionals in England, Scotland, and Wales [34]. BMJ Best Practice provides advice on symptom evaluation, tests to order, and treatment approach structured around the patient consultation.

We started by investigating embedding analogies in a small common-English corpus to answer Q1 and Q2.

Figure 1. Overview of our SemDeep approach.
View this figure

Preliminaries: Analogies for Shakespeare’s Romeo in a Small Common-English Corpus

Topic models are related to semantic fields [5]. There are many small corpora and tutorials illustrating the inner workings of topic models, such as the spatially motivated Latent Semantic Analysis (LSA) method [35] and the probabilistic method latent Dirichlet allocation (LDA) [36]. We used a small common-English corpus appearing in an LSA tutorial [37]. Textbox 1 shows the corpus used to answer Q1 and Q2.

A small common-English corpus consisting of 5 lines of text.

Romeo and Juliet

Juliet: Oh happy dagger!

Romeo died by dagger.

“Live free or die”, that’s the New-Hampshire’s motto

Did you know New-Hampshire is in New-England?

Textbox 1. A small common-English corpus consisting of 5 lines of text.

In common English, punctuation marks can change the meaning of a sentence. For example, “prevail, not perish” versus “prevail not, perish.” We did not transform routine letters into lowercase letters and did not remove punctuation marks, with the only exception of double quotations. Multimedia Appendix 1 contains the input text and the hyperparameter configuration for the CBOW model with word2vec [28].

Below, we summarize the answers to Q1 and Q2 (Multimedia Appendix 1):

  • Answer to Q1: A “good” vector semantic model should find a candidate y that is “semantically similar” to the target x = Romeo. The candidate y with the highest cosine for the CBOW model is you: The terms you and Romeo are near-synonyms, that is “interchangeable in some contexts” [38]. Hence, the answer to Q1 is “yes.”
  • Answer to Q2: We applied the 3CosAdd formula 2, where the target x = Romeo provides the semantic context. The terms die = z1 and died = z2 from the corpus are representative of inflectional morphology infinitive:past. The search query –x +z1 +z2 is posed to the CBOW model, that is “find the term y, which is similar to die and died, while different from Romeo”. Candidate y with the highest 3CosAdd is “dagger.” The term dagger belongs to the semantic field death as “dagger is an instrument that causes death”; thus, the candidate inference is true. Hence, the answer to Q2 is also “yes.”
Stage 1: Knowledge Acquisition (Acquisition of Domain-Specific Terms)

The PubMed systematic reviews (in Figure 1) [16] is an evidence-based searching filter “AND (systematic [sb])”, intended for retrieving “best evidence” information sources from PubMed/MEDLINE [32] such as Cochrane systematic reviews [39]. Health care–related institutions such as the World Health Organization promote PubMed searches with this filter (examples in Prevention and Control of Noncommunicable Diseases: Guidelines for Primary Health Care in Low Resource Settings [40]).

This study used a subset of PubMed systematic reviews [16] of 301,201 PubMed/MEDLINE publications (titles and available abstracts), called the PubMed systematic reviews subset (PMSB dataset). The preprocessing of the input text for the PMSB dataset and the hyperparameter configuration for Skip-gram and CBOW are identical to those in our previous study [41] and detailed in the study by Arguello Casteleiro et al [42].

From the PMSB dataset, a total of 423K n-grams with a frequency count >5 have vector representations in both models, that is CBOW and Skip-gram. We considered “good” the Skip-gram and CBOW embeddings created in our previous study [41] as they both perform well (using conventional evaluation measure precision [43]) in semantic similarity and relatedness tasks with the cosine formula. The n-gram z reused in this study (ie, z1 and z2) is from our previous study [41].

Applying the 3CosAdd Formula to Acquire the Top 12 Ranked Term Pairs (x,y): A 4-Term Analogy

To address Q3 and apply the 3CosAdd formula 2, 2 n-gram pairs (disease x,treatment z) from our previous study (prior knowledge) [41] were needed. We kept only the 12 top-ranked candidate n-grams y for the 3CosAdd formula, that is, the 12-candidate y with CBOW and Skip-gram embeddings yielding the highest 3CosAdd values. We limited the list of candidates to 12, similar to Arguello Casteleiro et al [42], and following cognitive theories like Novak JD and Cañas AJ [44].

Stage 2: Knowledge Organization (Explicit Conceptualization of the Meaning of Terms)

This stage accomplishes a named entity recognition (NER) [45] task involving 3 domain experts (2 biomedical terminologists and 1 medical consultant who performs clinical coding). Every UMLS Metathesaurus concept has a concept unique identifier (CUI) and at least one UMLS Semantic Type (broad category) [30] assigned. The NER task consists of 3 sequential subtasks (Multimedia Appendix 1):

  • First, disambiguation of n-grams y is difficult to interpret for being truncated strings of characters or containing short forms (eg, abbreviations or acronyms). String searches in the PMSB dataset and the web search the sense inventory, Allie [46], enabling disambiguation.
  • Second, the manual binary classification of candidate n-gram y as to whether it belongs to the semantic field Tx (ie, yTx). Following Artstein R and Poesio M [47], we reported the interrater agreement with a Krippendorff alpha [48].
  • Third, entity normalization (grounding) [49] with MetaMap [50], where 3 domain experts apply the NER guidelines for MetaMap's output [51] and together judge the automatic mapping of n-grams yTx to UMLS Metathesaurus concepts YTx. MetaMap performance is calculated using precision, recall, and F-measure [43,52].

We took n1 as the number of different UMLS Metathesaurus concepts (represented as Z1 and Z2) mapped as z1 and z2 in the search query. Once the NER task was completed, we obtained the NER winners. An NER winner was a search query -x +z1 +z2 with the maximum observed number for n2 or n3:

  • n2 is the number of different 12 top-ranked candidate n-grams y belonging to Tx, that is, the number of yTx.
  • n3 is the number of different UMLS Metathesaurus concepts YTx excluding Z1 and Z2.
Stage 3: Knowledge Validation (Validating Statements)

We sought evidence for the Metathesaurus concept pairs (X,YTx) acquired previously to determine the therapeutic intent of candidate YTx for target disease X, where X was the UMLS Metathesaurus concept mapped to n-gram x.

The same 3 domain experts from Stage 2 triaged the results of manual literature searches considering the following:

  1. The type of evidence-based information sources, seeking the “best evidence.” Evidence-based medicine [53] categorizes and ranks different types of clinical evidence [1]. For example, the Cochrane systematic reviews are at the forefront of “best evidence” [1], whereas studies of the physiological functions and clinicians’ observations are considered evidence of least value [1].
  2. The publication date, seeking the “most recent papers published.”

The 3 domain experts introduced 6 evidence-based categories to further refine the correlations between the semantic field disease and treatment (Tx). Table 1 illustrates them with examples of evidence (quoted text) and references for the UMLS Metathesaurus concepts YTx related to the target concept disease X=“C024302|Sepsis” with CUI=C024302. The rationale for the 7 evidence-based categories introduced is as follows:

  • The name of 4 of the evidence-based categories (top rows in Table 1) resembles the categories “beneficial, likely beneficial, no known benefit, harmful” for health care interventions from the decommissioned BMJ Clinical Evidence (predecessor of BMJ Best Practice [31]).
  • The evidence-based category “Tx ingredient” acknowledges that a complex treatment may have parts, that is “partitive relationships” [54].
  • The evidence-based category “correlation” captures “spontaneous unplanned inferences” [6].
  • The evidence-based category “general medical term” includes broad concepts of little value for clinicians that do not need further evidence (quotes and references).

This study distinguishes between NER winners (maximum observed number for n2 or n3 in Stage 2) and clinical winners. A clinical winner is a search query -x +z1 +z2 (a type of 4-term analogy) for target disease x with a maximum observed value for n4, that is, the number of different concepts YTx (excluding Z1 and Z2) assigned to the evidence-based category “Tx with therapeutic effect.”

To audit the evidence-based categories assigned along with the evidence collected (quotes and references) for the concept pairs (X,YTx), 2 more observers (O1 a medical consultant and O2 a BMJ health informatician who works with BMJ Best Practice content and has a junior doctor background) were asked to express agreement or disagreement with the evidence for the concept pairs (X,YTx). Multimedia Appendix 1 has the evaluation guidelines given to the observers. Cohen kappa [55] was used to measure interobserver agreement.

Table 1. Evidence from the literature searches, that is quoted text and reference, for unified medical language system Metathesaurus concept pairs (X, YTx) with X=C0243026|Sepsis.
Candidate concept YTx; UMLS CUI|Concept nameaEvidence-based categories for concept YTx correlated with concept XEvidence (quoted text) [evidence source] [citation]
C0056562|crystalloid solutionsTx with therapeutic effect“Step-by-step treatment approach: ... Administer 30 mL/kg crystalloid for hypotension or lactate ≥4 mmol/L (≥36 mg/dL)” [BMJ BP topic: 245] [14]
C0001617|Adrenal Cortex HormonesTx with uncertain therapeutic effect“Step-by-step treatment approach: Adjunctive therapies ... evidence for giving corticosteroids to patients with sepsis or septic shock is mixed.” [BMJ BP topic: 245] [14]
C0020352|HetastarchTx with unwanted or adverse effects (ie, nontherapeutic)“Step-by-step treatment approach: Fluid resuscitation ... HES solutions for infusion have been significantly restricted across the European Union and are contraindicated in critically ill patients and those with sepsis or renal impairment.” [BMJ BP topic: 245] [14]
C0677850|Adjuvant therapyPotential Tx (under research and development)“Adjuvant immune therapy to manipulate the hyper-inflammatory and/or immune-suppressive phase of sepsis is an attractive therapeutic option, which may improve outcome and ease the burden of antimicrobial resistance. However, before this can become a clinical reality, we must recognise that sepsis is a clinical syndrome, where significant heterogeneity exists.” [PMID: 30515242] [56]
C3273371|CD4 Positive Memory T-LymphocyteTx ingredient“Administration of immune-modulatory therapy is a promising treatment approach for treating sepsis survivors. … these therapies can improve pathogen clearance, increase CD4 T cell responsiveness, and promote survival in sepsis.” [PMID: 24791959] [57]
C0745442|Intravenous CathetersTx ingredient“Recommendations: Monitoring ... Central venous catheters will be required to ensure reliable delivery of vasoactive medication.” [BMJ BP topic: 245] [14]
C0812144|Medication administration: epiduralCorrelation (epidural → potential sites of infection: epidural sites → sepsis: investigations)“Investigations to identify causative organisms: ... If no localising signs are present, examination and culture of all potential sites of infection including wounds, catheters, prosthetic implants, epidural sites, and pleural or peritoneal fluid, as indicated by the clinical presentation and history, is required.” [BMJ BP topic: 245] [14]
C0013227|Pharmaceutical PreparationsGeneral medical termb

aThe references shown are either the PubMed identifier (PMID) or the topic number in BMJ Best Practice (“BMJ BP topic” for short).

bThe evidence-based category “general medical term” has no evidence (quoted text).


We obtained 5352 n-gram pairs from 446 search queries by applying the 3CosAdd formula and taking the top 12 values. These are presented in Multimedia Appendix 2 (worksheet Stage 1). These n-gram pairs are enriched with domain knowledge meaning (Stage 2) and the biomedical evidence found from literature searches is ratified with an audit (Stage 3).

Stage 1: Knowledge Acquisition (Acquisition of Domain-Specific Terms)

To apply the 3CosAdd formula (and systematic creation of search queries), we reused 63 unique n-gram pairs (x,z) from our previous study [41] (open-access [29]). Every reused n-gram z was mapped to the UMLS concept Z with the UMLS Semantic Type “T061|Therapeutic or Preventive Procedure” or “T121|Pharmacologic Substance.” Multimedia Appendix 1 has the UMLS CUI pairs (X,Z).

Applying the 3CosAdd Formula to Acquire the Top 12 Ranked Term Pairs (x,y): A 4-Term Analogy

With 63 n-gram pairs (x,z), we built 223 search queries -x +z1 +z2 for the 3CosAdd formula. Multimedia Appendix 2 (worksheet Stage 1) contains the 223 search queries and the 5352 (x,y) n-gram pairs for 10 target diseases x, that is, the 12 top-ranked n-grams (highest 3CosAdd value) obtained per search query from the CBOW and Skip-gram embeddings. An n-gram pair with y as a non-ASCII character was discarded.

Stage 2: Knowledge Organization (Explicit Conceptualization of the Meaning of Terms)

Different search queries brought the same (target x,candidate y) n-gram pairs from applying the 3CosAdd formula. Multimedia Appendix 2 (worksheet Stage 2) has 1935 unique (x,y) n-gram pairs from the 5352 n-gram pairs. Among the 1935 unique (x,y) n-gram pairs, there were 954 n-gram pairs (x,yTx) with candidate y belonging to Tx. The Krippendorff alpha [48] was 0.86 for the 3 domain experts for the binary classification (Tx or non-Tx). Considering all candidates yTx mapped to YTx for the 10 diseases (microaveraging) [43], MetaMap had an F-measure=80.00% with precision=77.00% and recall=83.25%. Multimedia Appendix 1 has the detailed results for NER subtasks, including an investigation of the UMLS semantic types for YTx.

Table 2 contains the NER winners, that is, the search query -x +z1 +z2 for the 3CosAdd formula per model and disease target x having the maximum observed values for n2 or n3.

  • The maximum observed value for n2 was the highest possible value, that is, n2=12, for both CBOW and Skip-gram.
  • The maximum observed value for n3 was for the search query, −epilepsy +valproate +AED. However, the number of different YTx (excluding Z1 and Z2) differed, that is, n3=11 for Skip-gram and n3=10 for CBOW.

Stage 3: Knowledge Validation (Validating Statements)

Multimedia Appendix 2 (worksheet Stage 3) has the 569 unique UMLS Metathesaurus concept pairs (X,YTx) mapped to the unique 954 n-gram pairs (x,yTx). Although the UMLS related concepts table (file=MRREL) [58] contains relationships asserted by source vocabularies between CUI pairs, only 68 of the 569 CUI pairs appeared within the MRREL table of 2019AA UMLS release.

Manual searches in the literature proved to be time-consuming and labor-intensive; thus, not all the concept pairs for the target disease anemia and hypertension had evidence. Hence, we limited the study to 408 UMLS CUI pairs (Multimedia Appendix 1), and only 59 of these were within the MRREL table (column J of Multimedia Appendix 2 worksheet Stage 3).

Table 2. NER winners per target disease x (search query -x +z1 +z2) for the 3CosAdd formula, that is, the highest value for n2 or n3 per model and per disease target x.
Disease target xModelNER max (n2)NER max (n3)Treatment z1 search queryTreatment z2 search queryn1n2n3
heart_failureCBOWaYesN/Abangiotensin-converting_enzyme_(ACE)_inhibitorsaldosterone_antagonists2126
heart_failureCBOWN/AYescardiac_resynchronization_therapy_(CRT)aldosterone_antagonists2109
heart_failureSkip-gramYesYesbeta-blockersaldosterone_antagonists2128
glaucomaCBOWYesYestrabeculectomycataract_surgery255
glaucomaSkip-gramYesYestrabeculectomycataract_surgery2106
CKDcCBOWYesYesnot_requiring_dialysisdialysis197
CKDSkip-gramYesYesnot_requiring_dialysisdialysis185
diabetesCBOWYesYesglucose_variabilityglucagon-like_peptide-1_receptor_agonists2106
diabetesSkip-gramYesYesglucose_variabilityglucagon-like_peptide-1_receptor_agonists2105
asthmaCBOWYesN/Ainhaled_corticosteroidLABAsd2116
asthmaCBOWN/AYesinhaled_corticosteroidsinhaled_corticosteroid1108
asthmaSkip-gramYesYesanti-LTsLABAs2128
epilepsyCBOWYesYesvalproateAEDe21210
epilepsySkip-gramYesYesvalproateAED21211
arthritisCBOWYesYesplus_methotrexatemethotrexate1129
arthritisSkip-gramYesYesmethotrexateDMARDsf2116
osteoarthritisCBOWN/AYeshyaluronic_acidglucosamine289
osteoarthritisCBOWYesN/Aknee_arthroplastyhyaluronic_acid297
osteoarthritisCBOWN/AYesvs_acetaminophenglucosamine289
osteoarthritisSkip-gramYesYesvs_acetaminophenhyaluronic_acid2118
anaemiaCBOWYesYesironerythropoiesis-stimulating_agents2119
anaemiaSkip-gramYesN/Ablood_transfusionsESAsg2126
anaemiaSkip-gramN/AYesrecombinant_human_erythropoietiniron2118
hypertensionCBOWYesN/Aantihypertensive_drugsangiotensin_receptor_blockers2126
hypertensionCBOWN/AYesantihypertensive_therapyantihypertensive2118
hypertensionSkip-gramYesYesantihypertensive_drug_classesantihypertensive11210

aCBOW: Continuous Bag-of-Words.

bN/A: not applicable.

cCKD: chronic kidney disease.

dLABA: long-acting beta2-agonist.

eAED: antiepileptic drug.

fDMARD: disease-modifying antirheumatic drug.

gESA: erythropoiesis-stimulating agent.

Table 3 shows the 7 evidence-based categories assigned to the 408 UMLS CUI pairs investigated thoroughly. There are 19 concept pairs (X,YTx) with more than 1 evidence-based category, such as the concept pair (X=C0014544|Epilepsy,YTx=C0080356|Valproate). The evidence-based category “Tx with therapeutic effect” has the highest number of CUI pairs, with 190 pairs (X,YTx), where 117 pairs have evidence (quotes) taken from BMJ Best Practice. The evidence-based category “correlation” has the highest number of evidence-based information sources with 108 uniform resource identifiers of the total 238. Multimedia Appendix 1 has further details.

Table 4 shows the clinical winners, that is, search query -x +z1 +z2 (a type of 4-term analogy) with the maximum observed number for n4 per target disease x. Table 4 reveals that an NER winner is not necessarily a clinical winner, that is, the maximum observed value for n4 does not always correspond to the maximum observed value for n3 or n2.

Table 3. The 408 unified medical language system concept unique identifier pairs investigated thoroughly and their evidence-based information sources per evidence-based category.
Evidence-based categories for concept YTx correlated with concept XNumber of CUIa pairsNumber of evidence-based information sources (ie, URIsb) for CUI pairsNumber of CUI pairs with BMJ Best Practice as evidence source
Tx with therapeutic effect19073117
Tx with uncertain therapeutic effect382211
Tx with unwanted or adverse effects (ie, nontherapeutic)524117
Potential Tx (under research and development)550
Tx ingredient22216
General medical term2600
Correlation9410819

aCUI: concept unique identifier.

bURI: Universal Resource Identifier.

In Table 4, there are two rows that are not clinical winners according to the observer O2. All rows except two are clinical winners according to the 3 domain experts and both observers.

Considering the 408 concept pairs (X,YTx) with evidence, observer O1 disagrees with 25 of them, and observer O2 disagrees with 26 of them. The Cohen kappa of −0.023 is paradoxical [59-61], resolved in Multimedia Appendix 1 following Cicchetti DV and Feinstein AR [61].

Table 5 shows how the evidence-based category “Tx with therapeutic effect” assigned by an observer (when in disagreement) affects the clinical winners from Table 4. For observer O1, the only change was a decrease of n4 from 5 (Table 4) to 4 (Table 5) in the search query, −anaemia +recombinant_human_erythropoietin +iron, for Skip-gram. The observer O2 provided additional therapeutic evidence from BMJ Best Practice when in disagreement, typically increasing n4 or making “new” clinical winners (eg, search query, −epilepsy +valproate +levetiracetam).

Table 4. Clinical winners (highest value of n4) per model and disease target x considering the 3 domain experts.
Disease target xModelNER max (n2)NER max (n3)Treatment z1 search queryTreatment z2 search queryn1n2n3n4
heart_failureCBOWaYescardiac_resynchronization_therapy_(CRT)aldosterone_antagonists21096
heart_failureSkip-gramYesYesbeta-blockersaldosterone_antagonists21285
glaucomaCBOWYesYestrabeculectomycataract_surgery2553
glaucomaSkip-gramYesYestrabeculectomycataract_surgery21063
CKDbCBOWYesYesnot_requiring_dialysisdialysis1975
CKDSkip-gramYesYesnot_requiring_dialysisdialysis1855
diabetesCBOWYesYesglucose_variabilityglucagon-like_peptide-1_receptor_agonists21066
diabetesSkip-gramYesYesglucose_variabilityglucagon-like_peptide-1_receptor_agonists21054
asthmaCBOWYesinhaled_corticosteroidsinhaled_corticosteroid11088
asthmaSkip-graminhaled_corticosteroidsinhaled_corticosteroid11187
epilepsyCBOWvalproateantiepileptic_drug211108
epilepsyCBOWvalproateantiepileptic_drugs211108
epilepsycSkip-gramYesYesvalproateAEDd212117
arthritisCBOWYesYesplus_methotrexatemethotrexate11292
arthritiscSkip-gramplus_methotrexatemethotrexate1742
osteoarthritisCBOWYesknee_arthroplastyhyaluronic_acid2975
osteoarthritisSkip-gramvs_acetaminophenviscosupplementation2787
anaemiaCBOWYesYesironerythropoiesis-stimulating_agents21194
anaemiaSkip-gramYesrecombinant_human_erythropoietiniron21185
hypertensionCBOWYesantihypertensive_therapyantihypertensive21186
hypertensionSkip-gramYesYesantihypertensive_drug_classesantihypertensive112108

aCBOW: Continuous Bag-of-Words.

bCKD: chronic kidney disease.

cNot clinical winners according to O2.

dAED: antiepileptic drug.

Table 5. Changes in clinical winners (highest value of n4) per model and disease target x considering observer O1 and O2.
Disease target xModelDifferences in clinical winner max (n4) according to observersTreatment z1 search queryTreatment z2 search queryn1n2n3n4
epilepsySkip-gramObserver O2a: Newvalproatelevetiracetam21097
arthritisCBOWbObserver O2: n4 differentplus_methotrexatemethotrexate11296
arthritisCBOWObserver O2: NewmethotrexateDMARDsc21196
arthritisSkip-gramObserver O2: NewIACIdDMARDs2965
arthritisSkip-gramObserver O2: Newplus_methotrexateDMARDs21065
anaemiaSkip-gramObserver O1e: n4 differentrecombinant_human_erythropoietiniron21184

aO2: BMJ health informatician who works with BMJ Best Practice content and has a junior doctor background.

bCBOW: Continuous Bag-of-Words.

cDMARD: disease-modifying antirheumatic drug.

dIACI: intra-articular corticosteroid injection.

eO1: medical consultant.

Multimedia Appendix 1 has the best clinical winner, which is an NER winner. Table 6 shows the best clinical winner that is not an NER winner. Table 6 illustrates the enrichment of the candidate n-grams y with domain knowledge meaning (Stage 2 normalizes n-grams with UMLS CUIs) and biomedical evidence ratified with an audit (Stage 3). The evidence provided for the evidence-based categories (quotes with references from the biomedical literature) is presented in Multimedia Appendix 2 (worksheet Stage 3).

In conclusion, considering the clinical winners found (Table 4), the answer to Q3 is “yes,” that is, the 4-term type of analogies discovered in a small common-English corpus can also be discovered in a large-scale biomedical corpus.

Table 6. Illustration of a Best clinical winner with max (n4)=8 for CBOW and disease target x = epilepsy, which is not an NER winner.
RankaCandidate y3CosAddUMLS CUI for concept YTx mapped to candidate yTxEvidence-based categories for concept YTx correlated with concept X
1lamotrigine0.385201C0064636Tx with therapeutic effect
2carbamazepine0.345227C0006949Tx with unwanted or adverse effects (ie, nontherapeutic)
3low_propensity0.324285b
4clonazepam0.310706C0009011Tx with uncertain therapeutic effect
5topiramate0.308402C0076829Tx with therapeutic effect
6lithium_valproate0.308223C0023870|C0080356Tx with therapeutic effect|Tx with unwanted or adverse effects (ie, nontherapeutic)
7clobazam0.306901C0055891Tx with therapeutic effect
8sodium_valproate0.300513C0037567Tx with therapeutic effect
9lorazepam0.29562C0024002Tx with therapeutic effect
10lithium0.294804C0023870Tx with therapeutic effect|Tx with unwanted or adverse effects (ie nontherapeutic)
11gabapentin_pregabalin_topiramate0.291698C0657912|C0076829|C0060926Tx with therapeutic effect
12antiepileptic_drugs_other_than0.290046C0003299Tx with therapeutic effect

aThe search query −x +z1 +z2 is listed in Table 4, which is −epilepsy +valproate +antiepileptic_drug. The character “|” appears when there is more than 1 CUI or evidence-based category.

bThe candidate y = “low_propensity” does not belong to the semantic field Tx, and so, it has no UMLS CUI assigned.

Answer Q4: An Empirical Heuristic with Some Predictive Power for Clinical Winners

Multimedia Appendix 2 (worksheet Q4) has the 304 search queries of the total of 446 (223 for CBOW and 223 for Skip-gram) queries, where all the candidates yTx mapped to concepts YTx have at least one evidence-based category assigned. Textbox 2 summarizes the empirical heuristic developed by visual inspection, focusing on rows with the minimum (n4=0) and the maximum observed values of n4. The heuristic is programmatically implemented as a Boolean expression composed of 3 expressions with the Boolean AND.

An empirical heuristic developed by visual inspection with some predictive power for the clinical winners.
  1. Avoid n-grams z1 and z2 having short forms
  2. Favor n-grams z1 or z2 (or both) not appearing among the 20 top-ranked candidates for target x with the highest value for cosine with Skip-gram embeddings
  3. Favor n-gram z2 with frequency counts in the corpus >100
Textbox 2. An empirical heuristic developed by visual inspection with some predictive power for the clinical winners.

The heuristic selects 93 of the 304 search queries, which brings 126 of the 190 UMLS Metathesaurus concepts YTx with the evidence-based category “Tx with therapeutic effect,” that is, YTx with therapeutic intent.

Table 7 (source data in Multimedia Appendix 1) shows the performance of the heuristic considering (1) the values of n4 (the last 3 yellow columns in Multimedia Appendix 2 worksheet Q4), (2) the different thresholds for n4, and (3) precision and recall as metric.

Considering the precision and recall values for the empirical heuristic (Table 7), the answer to Q4 is also “yes,” that is, some predictive power over the clinical winners obtained is possible.

Table 7. Precision and recall for the empirical heuristic developed using Multimedia Appendix 2 (worksheet Q4).
ThresholdTrue positive (TP)False positive (FP)False negative (FN)Precisiona %Recallb %
n4>091218997.8532.5
n4>184915190.3235.74
n4>2732011178.4939.67
n4>348457651.6138.71
n4>428655230.1135
n4>514792015.0541.18
n4>698459.6864.29
n4>748904.3100

aPrecision: calculated as TP/(TP+FP).

bRecall: calculated as TP/(TP+FN).


Principal Findings

Humans can agree that the semantic field person {you; Romeo} is related to the semantic field death {die; died; dagger} in the context of Shakespeare’s Romeo. Hence, we answer Q1 and Q2 with a “yes”; therefore, analogical reasoning with CBOW embeddings seems feasible with a small common-English corpus. This challenges the current assumption that “learning in current deep learning models relies on massive data” [3].

We answered Q3 by demonstrating that there is proof of the generalization; thus, the 3CosAdd formula can discover another type of 4-term analogy that is not a pair-based proportional analogy. Furthermore, we have proven that the analogical inferences sanctioned by the 3CosAdd formula with embeddings could extract treatments with therapeutic intent from free text. Indeed, there were strong examples of analogical reasoning with abstract semantic relations between z1 and z2 among clinical winners (Table 4):

  • Antonym. The search query, −CKD +not_requiring_dialysis +dialysis, with n4=5 for CBOW and Skip-gram.
  • Synonym. The search query, −asthma +inhaled_corticosteroids +inhaled_corticosteroid, with n4=8 for CBOW (the best clinical winner) and n4=7 for Skip-gram, where the relation between z2 and z1 was inflectional morphology singular:plural. This query resembled the search query, −Romeo +die +died.
  • Category membership. The search query, −epilepsy +valproate +antiepileptic_drug, with n4=8 for CBOW. The search query, −hypertension +antihypertensive_drug_classes +antihypertensive, with n4=8 for Skip-gram. Both search queries were the best clinical winners (maximum observed value for n4).
  • Commonalities in structural features. All search queries focused on the therapeutic intent of z1 and z2 for target disease x. However, some queries did not have the above abstract semantic relationships between z1 and z2. For example, the search queries −osteoarthritis +knee_arthroplasty +hyaluronic_acid with n4=5 for CBOW and −heart_failure +beta-blockers +aldosterone_antagonists with n4=5 for Skip-gram.

We answered Q4 by demonstrating that it is feasible to gain some predictive power for the clinical winners; therefore, a tactic preference was latent promising systematicity [6]. Textbox 3 highlights the precision and recall values for 3 n4 thresholds of the overall performance of the empirical heuristic developed by visual inspection.

Empirical heuristic performance for 304 search queries with all candidate concepts YTx with evidence.
  • With a threshold n4 >7, the recall is 100%. All search queries with n4 >7 (the best clinical winners) are selected by the heuristic. The precision was 4.30% (the lowest value).
  • With a threshold n4 >0 (at least one YTx with therapeutic intent), the precision was 97.85% (the highest value) and the recall was 32.50%.
  • With a threshold n4 >2, where 3 was the lowest value among the clinical winners (Tables 3 and 4), the precision was 78.49% and the recall was 39.67%.
Textbox 3. Empirical heuristic performance for 304 search queries with all candidate concepts YTx with evidence.

Limitations

Our work relies on semantic fields and has 2 main limitations [62]: (1) there are overlaps of meaning and (2) there are gaps in meaning. This has 2 clear implications for the lists of concepts YTx per disease x:

  • The lists may not comprise mutually exclusive concepts in meaning. For example, “C0060657|formoterol” and “C1276807|Budesonide/formoterol” are both treatments with evidence of therapeutic intent for asthma [63].
  • The lists were incomplete. For example, “C0772501|Levalbuterol” and “C0907850|ciclesonide” are both treatments with evidence of therapeutic intent for asthma [63] and not among the YTx for asthma.

We did not use Skip-gram with negative sampling (also known as SGNS); therefore, it can be argued that we did not use the best configuration of a word2vec model [21]. The effect of hyperparameter configurations appeared in studies by Levy et al and Chiu et al [64,65], and Allen and Hospedales [66] reviewed mathematical proofs and equations with an emphasis on SGNS for pair-based proportional analogies.

For Stage 1, the 3CosAdd formula needed at least two n-gram pairs (disease x, treatment z) [29]. Only one search query could be made for the target disease, chronic kidney disease and diabetes, and none for obesity. Other studies that replicated the application to the 3CosAdd formula for target disease x could suffer the same limitation. For example, in Appendix B in the study by Pakhomov et al [67], among the 100 top-ranked candidate terms (highest cosine value) “semantically similar or related” to target disease “heart failure”, there were no treatments (ie, Tx encompassing 3 textual definitions from Hart et al [24]).

For Stage 2, the MetaMap version was 2016v2 (with a 2016 UMLS release), and few n-grams were considered as clear terminological gaps. The n-gram “anti-VEGF_agents” was manually mapped to CUI=C4727875, which exists in the 2019AA UMLS release. Five n-grams were mapped to very broad CUIs as they had the character “*” in Multimedia Appendix 2 (worksheet Stage 3).

The NER task (Stage 2) and the searchers in the literature seeking evidence for concept pairs (Stage 3) were time-consuming and required highly trained domain experts. The appraisal of the literature was not performed by a review team as proficient as the ones conducting Cochrane systematic reviews.

The heuristic developed by visual inspection lacked finesse, and its improvement calls for further investigation.

Comparison With Prior Work

The UMLS CUIs were mapped to SNOMED CT identifiers [30]. From a “digital health care” perspective [68], the UK NHS is moving toward the adoption of SNOMED CT as the only terminology for all care settings [69]. A subset of SNOMED CT concepts under worldwide adoption is the CORE Problem List Subset of SNOMED CT [70], and the UK NHS has developed 2 human-readable SNOMED CT subsets [71]: UK Clinical Extension and UK Drug Extension. However, SNOMED CT lacks statements representing the treatments that can be considered for a disease (eg, inhaled corticosteroid treats asthma) and, to the best of the authors’ knowledge, there are no SNOMED CT subsets for well-known diseases.

There are reusable datasets for evaluating relatedness made of UMLS CUI pairs:

  • Medical coders set [72]: 101 CUI pairs mapped to terms, typically multiple words. Only 29 pairs have a high interrater agreement.
  • Medical Residents Relatedness Set [73]: 588 CUI pairs mapped to terms, typically single words. Using single words is a severe limitation as “most medical terms consist of more than one word” [67].
  • UMLS MRREL table [58]: It has relationships asserted by source vocabularies between CUI pairs. Among the relationship attributes appear the following: “may_prevent”, “may_treat”, and “has_contraindicated_drug”.

All reusable datasets mentioned above lack evidence (quotes with references) from the biomedical literature. Multimedia Appendix 1 cross-compares these reusable datasets and the 408 UMLS CUI pairs investigated thoroughly in this study.

Conclusions

Extracting clinically useful information automatically from free text in PubMed/MEDLINE may require a natural language understanding of statements containing relevant relations for health care. Hence, extracting treatments with therapeutic intent by analogical reasoning from embeddings (423K n-grams from the PMSB dataset) is an ambitious goal. Our SemDeep approach is knowledge-based, underpinned by embedding analogies that exploit prior knowledge. Biomedical facts from embedding analogies (a 4-term type, not pairwise) are potentially useful for clinicians. The heuristic offers a practical way to discover beneficial treatments for well-known diseases.

Learning from deep learning models does not require a massive amount of data. Embedding analogies are not limited to pairwise analogies; hence, analogical reasoning with embeddings is underexploited.

Acknowledgments

The authors thank Tim Furmston (IT Research, University of Manchester, United Kingdom) for help with the software and electronic infrastructure.

Conflicts of Interest

CW works for BMJ which produces the clinical decision support tool BMJ Best Practice. All other authors have no conflict of interest to declare.

Multimedia Appendix 1

Additional material.

PDF File (Adobe PDF File), 378 KB

Multimedia Appendix 2

Data to replicate the results reported.

XLS File (Microsoft Excel File), 708 KB

  1. Masic I, Miokovic M, Muhamedagic B. Evidence based medicine - new approaches and challenges. Acta Inform Med 2008;16(4):219-225 [FREE Full text] [CrossRef] [Medline]
  2. Avorn J. The psychology of clinical decision making - implications for medication use. N Engl J Med 2018 Mar 22;378(8):689-691. [CrossRef] [Medline]
  3. Lu H, Wu YN, Holyoak KJ. Emergence of analogy from relation learning. Proc Natl Acad Sci U S A 2019 Mar 5;116(10):4176-4181 [FREE Full text] [CrossRef] [Medline]
  4. LeCun Y, Bengio Y, Hinton G. Deep learning. Nature 2015 May 28;521(7553):436-444. [CrossRef] [Medline]
  5. Jurafsky D, Martin J. Stanford University. 2019. Speech and Language Processing   URL: https://web.stanford.edu/~jurafsky/slp3/ [accessed 2020-06-01]
  6. Gentner D, Markman AB. Structure mapping in analogy and similarity. Am Psychol 1997;52(1):45-56. [CrossRef]
  7. ACL Anthology. Analogy (State of the Art)   URL: https://aclweb.org/aclwiki/Analogy_(State_of_the_art) [accessed 2020-06-01]
  8. Schnabel T, Labutov I, Mimno D, Joachims T. Evaluation Methods for Unsupervised Word Embeddings. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing. 2015 Presented at: EMNLP'15; September 17-21, 2015; Lisbon, Portugal p. 298-307. [CrossRef]
  9. Mikolov T, Yih W, Zweig G. Linguistic Regularities in Continuous Space Word Representations. In: Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 2013 Presented at: NAACL'13; June 9-14, 2013; Atlanta, Georgia, US p. 746-751   URL: https://www.aclweb.org/anthology/N13-1090/
  10. Drozd A, Gladkova A, Matsuoka S. Word Embeddings, Analogies, and Machine Learning: Beyond King - Man + Woman = Queen. In: Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers. 2016 Presented at: COLING'16; December 11-16, 2016; Osaka, Japan p. 3519-3530   URL: https://www.aclweb.org/anthology/C16-1332/
  11. Liu H, Wu Y, Yang Y. Analogical inference for multi-relational embeddings. Arxiv 2017:- epub ahead of print (1705.02426) [FREE Full text]
  12. Rather NN, Patel CO, Khan SA. Using deep learning towards biomedical knowledge discovery. Int J Math Sci Comput 2017 Apr 8;3(2):1-10. [CrossRef]
  13. Dynomant E, Lelong R, Dahamna B, Massonnaud C, Kerdelhué G, Grosjean J, et al. Word embedding for the French natural language in health care: comparative study. JMIR Med Inform 2019 Jul 29;7(3):e12310 [FREE Full text] [CrossRef] [Medline]
  14. BMJ Best Practice. 2018. Sepsis in Adults   URL: http://bestpractice.bmj.com/topics/en-gb/245 [accessed 2020-06-01]
  15. Gatta A, Verardo A, Bolognesi M. Hypoalbuminemia. Intern Emerg Med 2012 Oct;7(Suppl 3):S193-S199. [CrossRef] [Medline]
  16. National Library of Medicine - National Institutes of Health - NIH. Search Strategy Used to Create the PubMed Systematic Reviews Filter   URL: https://www.nlm.nih.gov/bsd/pubmed_subsets/sysreviews_strategy.html [accessed 2020-06-01]
  17. Special Issue on Semantic Deep Learning.   URL: http://www.semantic-web-journal.net/content/special-issue-semantic-deep-learning [accessed 2020-06-01]
  18. Stevens R, Goble CA, Bechhofer S. Ontology-based knowledge representation for bioinformatics. Brief Bioinform 2000 Dec;1(4):398-414. [CrossRef] [Medline]
  19. Bartha P. Stanford Encyclopedia of Philosophy. 2019. Analogy and Analogical Reasoning   URL: https://plato.stanford.edu/archives/spr2019/entries/reasoning-analogy/ [accessed 2020-06-01]
  20. Mikolov T, Chen K, Corrado G, Dean J. Efficient Estimation of Word Representations in Vector Space. In: Proceedings of 1st International Conference on Learning Representations. 2013 Presented at: ICLR'13; May 2-4, 2013; Scottsdale, Arizona, US   URL: https://arxiv.org/abs/1301.3781
  21. Levy O, Goldberg Y. Linguistic Regularities in Sparse and Explicit Word Representations. In: Proceedings of the Eighteenth Conference on Computational Natural Language Learning. 2014 Presented at: ACL'14; June 26-27, 2014; Baltimore, Maryland, USA p. 171-180. [CrossRef]
  22. Nerlich B, Clarke DD. Semantic fields and frames: historical explorations of the interface between language, action, and cognition. J Pragmat 2000 Jan;32(2):125-150. [CrossRef]
  23. SNOMED Confluence. Technical Implementation Guide   URL: http://snomed.org/tig [accessed 2020-06-01]
  24. Hart T, Tsaousides T, Zanca JM, Whyte J, Packel A, Ferraro M, et al. Toward a theory-driven classification of rehabilitation treatments. Arch Phys Med Rehabil 2014 Jan;95(1 Suppl):S33-44.e2. [CrossRef] [Medline]
  25. Hill F, Reichart R, Korhonen A. SimLex-999: evaluating semantic models with (genuine) similarity estimation. Comput Linguist 2015 Dec;41(4):665-695. [CrossRef]
  26. NCBO BioPortal. 2020. Semanticscience Integrated Ontology   URL: https://bioportal.bioontology.org/ontologies/SIO [accessed 2020-06-01]
  27. Chen D, Peterson J, Griffiths T. Evaluating Vector-Space Models of Analogy. In: Proceedings of the 39th Annual Meeting of the Cognitive Science Society. 2017 Presented at: CogSci'17; July 26-29, 2017; London, UK.
  28. Google Code. Word2vec - Default   URL: https://code.google.com/archive/p/word2vec/source/default/source [accessed 2020-06-01]
  29. Arguello-Casteleiro M, Stevens R, Des-Diz J, Wroe C, Fernandez-Prieto MJ, Maroto N, et al. Exploring semantic deep learning for building reliable and reusable one health knowledge from PubMed systematic reviews and veterinary clinical notes. J Biomed Semantics 2019 Nov 12;10(Suppl 1):22 [FREE Full text] [CrossRef] [Medline]
  30. Bodenreider O. The unified medical language system (UMLS): integrating biomedical terminology. Nucleic Acids Res 2004 Jan 1;32(Database issue):D267-D270 [FREE Full text] [CrossRef] [Medline]
  31. BMJ Best Practice.   URL: https://bestpractice.bmj.com/ [accessed 2020-06-01]
  32. PubMed/MEDLINE.   URL: https://www.ncbi.nlm.nih.gov/pubmed/ [accessed 2020-06-10]
  33. Kwag KH, González-Lorenzo M, Banzi R, Bonovas S, Moja L. Providing doctors with high-quality information: an updated evaluation of web-based point-of-care information summaries. J Med Internet Res 2016 Jan 19;18(1):e15 [FREE Full text] [CrossRef] [Medline]
  34. BMJ Best Practice. Do You Work for the NHS in England, Scotland or Wales?   URL: https://bestpractice.bmj.com/info/bmagp/ [accessed 2020-06-01]
  35. Landauer TK, Dumais ST. A solution to Plato's problem: the latent semantic analysis theory of acquisition, induction, and representation of knowledge. Psychol Rev 1997;104(2):211-240. [CrossRef]
  36. Blei DM, Ng AY, Jordan MI. Latent dirichlet allocation. J Mach Learn Res 2003;3(4-5):993-1002. [CrossRef]
  37. Thomo A. Engineering and Computer Science - University of Victoria. Latent Semantic Analysis   URL: https://www.engr.uvic.ca/~seng474/svd.pdf [accessed 2020-06-01]
  38. Ardila J. Meaning in language. An introduction to semantics and pragmatics. J Pragma 2011 Aug;43(10):2670-2672. [CrossRef]
  39. Cochrane Library: Cochrane Reviews. Cochrane Database of Systematic Reviews   URL: https://www.cochranelibrary.com/cdsr/about-cdsr [accessed 2020-06-01]
  40. World Health Organization. Prevention and Control of Noncommunicable Diseases: Guidelines for Primary Health Care in Low Resource Settings. Geneva, Switzerland: World Health Organization; 2012.
  41. Arguello-Casteleiro M, Stevens R, des-Diz J, Wroe C, Fernandez-Prieto MJ, Maroto N, et al. Exploring semantic deep learning for building reliable and reusable one health knowledge from PubMed systematic reviews and veterinary clinical notes. J Biomed Semantics 2019 Nov 12;10(Suppl 1):22 [FREE Full text] [CrossRef] [Medline]
  42. Casteleiro MA, Demetriou G, Read W, Prieto MJ, Maroto N, Fernandez DM, et al. Deep learning meets ontologies: experiments to anchor the cardiovascular disease ontology in the biomedical literature. J Biomed Semantics 2018 Apr 12;9(1):13 [FREE Full text] [CrossRef] [Medline]
  43. Manning C, Schütze H. Foundations of Statistical Natural Language Processing. New York, USA: MIT Press; 1999.
  44. Novak J, Cañas A. CMap. 2008. The Theory Underlying Concept Maps and How to Construct and Use Them   URL: http://cmap.ihmc.us/Publications/ResearchPapers/TheoryUnderlyingConceptMaps.pdf [accessed 2020-06-01]
  45. Nadkarni PM, Ohno-Machado L, Chapman WW. Natural language processing: an introduction. J Am Med Inform Assoc 2011;18(5):544-551 [FREE Full text] [CrossRef] [Medline]
  46. Yamamoto Y, Yamaguchi A, Bono H, Takagi T. Allie: a database and a search service of abbreviations and long forms. Database (Oxford) 2011;2011:bar013 [FREE Full text] [CrossRef] [Medline]
  47. Artstein R, Poesio M. Inter-coder agreement for computational linguistics. Comput Linguist 2008 Dec;34(4):555-596. [CrossRef]
  48. Shelley M, Krippendorff K. Content analysis: an introduction to its methodology. J Am Stat Assoc 1984 Mar;79(385):240. [CrossRef]
  49. Rebholz-Schuhmann D, Oellrich A, Hoehndorf R. Text-mining solutions for biomedical research: enabling integrative biology. Nat Rev Genet 2012 Dec;13(12):829-839. [CrossRef] [Medline]
  50. Aronson AR, Lang F. An overview of MetaMap: historical perspective and recent advances. J Am Med Inform Assoc 2010;17(3):229-236 [FREE Full text] [CrossRef] [Medline]
  51. Guidelines Developed for Step 4: Named Entity Recognition Task.   URL: https:/​/static-content.​springer.com/​esm/​art%3A10.1186%2Fs13326-019-0212-6/​MediaObjects/​13326_2019_212_MOESM2_ESM.​pdf [accessed 2020-06-01]
  52. Pratt W, Yetisgen-Yildiz M. A study of biomedical concept identification: MetaMap vs people. AMIA Annu Symp Proc 2003:529-533 [FREE Full text] [Medline]
  53. Sackett DL, Rosenberg WM, Gray JA, Haynes RB, Richardson WS. Evidence based medicine: what it is and what it isn't. Br Med J 1996 Jan 13;312(7023):71-72 [FREE Full text] [CrossRef] [Medline]
  54. Sager J. A Practical Course in Terminology Processing. Amsterdam, The Netherlands: John Benjamins Publishing; 1990.
  55. Cohen J. A coefficient of agreement for nominal scales. Educ Psychol Meas 1960 Jul 2;20(1):37-46. [CrossRef]
  56. Davies R, O'Dea K, Gordon A. Immune therapy in sepsis: are we ready to try again? J Intensive Care Soc 2018 Nov;19(4):326-344 [FREE Full text] [CrossRef] [Medline]
  57. Cabrera-Perez J, Condotta SA, Badovinac VP, Griffith TS. Impact of sepsis on CD4 T cell immunity. J Leukoc Biol 2014 Nov;96(5):767-777 [FREE Full text] [CrossRef] [Medline]
  58. NCBI Bookshelf. 2019. UMLS Reference Manual   URL: https://www.ncbi.nlm.nih.gov/books/NBK9685/table/ch03.T.related_concepts_file_mrrel_rrf/ [accessed 2020-06-10]
  59. Viera AJ, Garrett JM. Understanding interobserver agreement: the kappa statistic. Fam Med 2005 May;37(5):360-363 [FREE Full text] [Medline]
  60. Feinstein AR, Cicchetti DV. High agreement but low kappa: I. The problems of two paradoxes. J Clin Epidemiol 1990;43(6):543-549. [CrossRef] [Medline]
  61. Cicchetti DV, Feinstein AR. High agreement but low kappa: II. Resolving the paradoxes. J Clin Epidemiol 1990;43(6):551-558. [CrossRef] [Medline]
  62. Lehrer A. The Influence of Semantic Fields on Semantic Change. Berlin, Germany: De Gruyter Mouton; 1985.
  63. BMJ Best Practice. Asthma in Adults   URL: http://bestpractice.bmj.com/topics/en-gb/44 [accessed 2020-06-01]
  64. Levy O, Goldberg Y, Dagan I. Improving distributional similarity with lessons learned from word embeddings. Trans Assoc Comput Linguist 2015 Dec;3:211-225. [CrossRef]
  65. Chiu B, Crichton G, Korhonen A, Pyysalo S. How to Train good Word Embeddings for Biomedical NLP. In: Proceedings of the 15th Workshop on Biomedical Natural Language Processing. 2016 Presented at: ACL'16; August 12, 2016; Berlin, Germany. [CrossRef]
  66. Allen C, Hospedales T. Analogies Explained: Towards Understanding Word Embeddings. In: Proceedings of the 36th International Conference on Machine Learning. 2019 Presented at: ICML'19; June 9-15, 2019; Long Beach, California, USA p. 223-231. [CrossRef]
  67. Pakhomov SV, Finley G, McEwan R, Wang Y, Melton GB. Corpus domain effects on distributional semantic modeling of medical terms. Bioinformatics 2016 Dec 1;32(23):3635-3644 [FREE Full text] [CrossRef] [Medline]
  68. Tresp V, Overhage JM, Bundschus M, Rabizadeh S, Fasching PA, Yu S. Going digital: a survey on digitalization and large-scale data analytics in healthcare. Proc IEEE 2016 Nov;104(11):2180-2206. [CrossRef]
  69. Government of UK. Personalised Health and Care 2020   URL: https://www.gov.uk/government/publications/personalised-health-and-care-2020 [accessed 2020-06-01]
  70. National Library of Medicine - National Institutes of Health - NIH. The CORE Problem List Subset of SNOMED CT   URL: https://www.nlm.nih.gov/research/umls/Snomed/core_subset.html [accessed 2020-06-01]
  71. NHS Digital. SNOMED CT Subset Members in Spreadsheet View   URL: https://isd.digital.nhs.uk/trud3/user/guest/group/0/pack/40 [accessed 2020-06-01]
  72. Pedersen T, Pakhomov SV, Patwardhan S, Chute CG. Measures of semantic similarity and relatedness in the biomedical domain. J Biomed Inform 2007 Jul;40(3):288-299 [FREE Full text] [CrossRef] [Medline]
  73. Pakhomov S, McInnes B, Adam T, Liu Y, Pedersen T, Melton GB. Semantic similarity and relatedness between clinical terms: an experimental study. AMIA Annu Symp Proc 2010 Dec 13;2010:572-576 [FREE Full text] [Medline]


BMJ: British Medical Journal
CBOW: Continuous Bag-of-Words
CUI: concept unique identifier
MRREL: the UMLS related concepts table (file=MRREL)
NER: named entity recognition
NHS: National Health Service
PMSB: PubMed systematic reviews subset
SemDeep: Semantic Deep Learning
SGNS: skip-gram with negative sampling
SNOMED CT: systematized nomenclature of medicine - clinical terms
Tx: treatment
UMLS: unified medical language system
URIs: Universal Resource Identifiers


Edited by G Eysenbach; submitted 06.11.19; peer-reviewed by Z He; comments to author 28.11.19; revised version received 27.02.20; accepted 27.02.20; published 06.08.20

Copyright

©Mercedes Arguello Casteleiro, Julio Des Diz, Nava Maroto, Maria Jesus Fernandez Prieto, Simon Peters, Chris Wroe, Carlos Sevillano Torrado, Diego Maseda Fernandez, Robert Stevens. Originally published in JMIR Medical Informatics (http://medinform.jmir.org), 06.08.2020.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on http://medinform.jmir.org/, as well as this copyright and license information must be included.