Published on in Vol 13 (2025)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/76912, first published .
Named Entity Recognition for Chinese Cancer Electronic Health Records—Development and Evaluation of a Domain-Specific BERT Model: Quantitative Study

Named Entity Recognition for Chinese Cancer Electronic Health Records—Development and Evaluation of a Domain-Specific BERT Model: Quantitative Study

Named Entity Recognition for Chinese Cancer Electronic Health Records—Development and Evaluation of a Domain-Specific BERT Model: Quantitative Study

1School of Management, Beijing University of Chinese Medicine, No. 11, North Third Ring Road East, Chaoyang District, Beijing, China

2Department of Information Management, Peking University, Beijing, China

3Information Center, Dongfang Hospital, Beijing, China

Corresponding Author:

Fengying Guo, MS


Background: The unstructured data of Chinese cancer electronic health records (EHRs) contains valuable medical expertise. Accurate medical entity recognition is crucial for building a medical-assisted decision system. Named entity recognition (NER) in cancer EHRs typically uses general models designed for English medical records. There is a lack of specialized handling for cancer-specific records and limited application to Chinese medical records.

Objective: This study aims to propose a specific NER model to enhance the recognition of medical entities in Chinese cancer EHRs.

Methods: Desensitized inpatient EHRs related to breast cancer were collected from a leading hospital in Beijing. Building upon the MC Bidirectional Encoder Representations from Transformers (BERT) foundation, the study further incorporated a Chinese cancer corpus for pretraining, resulting in the construction of the ChCancerBERT pretrained model. In conjunction with dilated-gated convolutional neural networks, bidirectional long short-term memory, multihead attention mechanism, and a conditional random field, this model forms a multimodel, multilevel integrated NER approach.

Results: This approach effectively extracts medical entity features related to symptoms, signs, tests, treatments, and time in Chinese breast cancer EHRs. The entity recognition performance of the proposed model surpasses that of the baseline model and other models compared in the experiment. The F1-score reached 86.93%, precision reached 87.24%, and recall reached 86.61%. The model introduced in this study demonstrates exceptional performance on the CCKS2019 dataset, attaining a precision rate of 87.26%, a recall rate of 87.27%, and an impressive F1-score of 87.26%, surpassing that of existing models.

Conclusions: The experiments demonstrate that the approach proposed in this study exhibits excellent performance in NER within breast cancer EHRs. This advancement will further contribute to clinical decision support for cancer treatment and research. In addition, the study reveals that incorporating domain-specific corpora in clinical NER tasks can further enhance the performance of BERT models in specialized domains.

JMIR Med Inform 2025;13:e76912

doi:10.2196/76912

Keywords



Background

Cancer is one of the leading causes of mortality [1], imposing a significant psychological burden on patients and potentially triggering mental health disorders such as anxiety and depression [2]. The proliferation of electronic health records (EHRs) has provided a crucial source of demographic information, medical history, diagnostic tests, and clinical treatment data for cancer research, aiding in better diagnosis, prognosis, and treatment of the disease [3]. However, the large amount of unstructured text in EHRs presents significant challenges for clinical research and analysis. Structuring this unstructured information is essential for subsequent data analysis and mining to support clinical decision-making. Named entity recognition (NER) is a critical component of natural language processing (NLP), plays a key role in this process by identifying and classifying entities, allowing for transforming raw clinical texts into structured data, thereby enabling large-scale data mining and supporting intelligent health care applications [4].

In recent years, many studies have explored different NER methods to structure EHRs. The widespread application of pretrained models such as Bidirectional Encoder Representations from Transformers (BERT) has significantly enhanced NER performance in the medical field [5,6]. These improvements are especially valuable in the medical domain, where precise recognition of specialized entities directly impacts the reliability of downstream analyses including cohort selection, clinical decision support, and the construction of medical knowledge bases [7,8]. However, most of these studies focus on English texts, and the differences between English and Chinese in terms of language expression may limit the generalizability of English medical NER methods when directly applied to Chinese medical texts.

Compared to medical texts in common diseases, cancer EHRs contain richer and more complex information. These records span multiple fields, including medicine, biology, and pharmacology, and encompass specific medical terminologies such as cancer staging, treatment plans, and drug names. They also include diagnostic and treatment records like surgical notes and pathology results, with a prevalent occurrence of nested entities. In current NER tasks for the cancer domain, models such as BERT and other deep learning approaches based on general corpora are often used without fine-tuning for the numerous specialized terms and specific expressions in this field, which may result in errors or omissions. In addition, there is no publicly available dataset benchmark for the cancer domain. Therefore, NER tasks for Chinese cancer EHRs remain both challenging and necessary.

To address the aforementioned issues, this study contributes the following:

  1. We retrained the MC-BERT model using a corpus specific to the Chinese cancer domain, resulting in the ChCancerBERT pretrained model tailored for NER tasks in the Chinese cancer field. This approach effectively enhances the semantic representation capabilities of pretrained models in a specialized domain.
  2. In the Chinese medical domain, we built a hybrid model to capture temporal features and bidirectional semantic information. This combination enhances the model’s ability to perform multidimensional feature modeling.
  3. To validate the model’s effectiveness, we created a Chinese breast cancer EHR corpus with manual annotations. We applied various models to this corpus, and the experimental results demonstrated that our proposed model outperforms existing models, achieving superior performance. We also applied the model proposed in this study to the CCKS2019 dataset and compared the results with those from existing studies for evaluation.

Related Work

In the medical field, prior research has developed various NER methods for extracting information from unstructured data such as EHRs. Table 1 summarizes related studies and methods for clinical NER in the medical field.

NER techniques can be categorized into 3 main types: dictionary and rule-based methods, machine learning–based methods, and deep learning–based methods. Dictionary and rule-based NER methods rely primarily on researchers analyzing textual writing rules and establishing dictionaries, then using regular expression matching to achieve entity recognition and information extraction. For instance, Najafabadipour et al [9] extracted multiple cancer concepts such as tumor staging, mutation status, and patient presentation from lung cancer clinical records using regular expressions and the UMLS8 dictionary. Yim et al [10] used rule-based extraction to identify 3 liver cancer entities: tumor size, staging level, and percentage of tumors invading the liver. These methods are straightforward and efficient but lack generalizability and require significant human effort to build dictionaries [10]. Machine learning–based methods use sequence labeling and train models such as hidden Markov model, conditional random fields (CRFs), and maximum entropy Markov model on large manually annotated feature corpora to achieve NER. For example, Savova et al [11] developed the DeepPhe software, which uses a combination of rules, domain knowledge bases, and machine learning to extract cancer phenotypes from clinical records. Weegar et al [12] used a CRF model to extract cervical cancer symptom information.

In recent years, deep learning methods have made significant progress in medical NER tasks, demonstrating substantial advantages in NLP compared to traditional feature-based machine learning methods. These approaches typically predict the boundaries and types of entities by labeling each word, thereby capturing deeper and more abstract features. For instance, An et al [13] used a bidirectional long short-term memory (BiLSTM)-CRF model with a multihead attention (MHA) mechanism to perform Chinese clinical EHR NER. Kong et al [14] combined multilayer convolutional neural networks (CNNs) with an attention mechanism to improve NER in Chinese EHRs. The emergence of pretrained models like BERT has further enhanced the performance of NER. In the medical domain, existing research shows that BERT-based models can be effectively applied to medical information extraction tasks. Li et al [5] achieved excellent results using a BERT-BiLSTM-CRF model on the CCKS2018 and CCKS2019 datasets. Li et al [15] proposed an A Lite Bidirectional Encoder Representations from Transformers (ALBERT)-based model with a MHA mechanism for Chinese medical NER, validated on the CCKS2019 dataset. Chen et al [6] constructed a hybrid model combining MC-BERT, BiLSTM, CNN, MHA, and CRF to achieve NER in Chinese EHRs. Most of these studies primarily applied their deep learning models to publicly available datasets such as CCKS2017 and CCKS2019, without further testing them on specific medical departments or diseases.

Existing research indicates that using domain-specific text as training data, as opposed to general language models, can yield better performance. For the cancer field, Zhou et al [7] proposed CancerBERT, a pretrained model specifically designed for the English language in the cancer domain. Currently, there is no similar model for the Chinese cancer domain. Given the differences between Chinese and English in terms of vocabulary, sentence structure, grammatical rules, and semantic expression, directly applying models trained on English corpora may lead to suboptimal results.

Table 1. Summary of related work on clinical named entity recognition methods in the medical and cancer domains.
ProposalApproachMethodDomainDomain modelBenchmarkLanguage
Najabadipour et al 2018 [9]RulesRule-based methodsCancerYesNoSpanish
Yim et al.2016 [10]RulesRule-based methodsCancerYesNoEnglish
Savova et al 2017 [11]Rules and machine learningRule-based methodsCancerYesNoEnglish
Weegar et al 2015 [12]Machine learningCRFaCancerNoNoEnglish
An et al 2022 [13]Deep learningBiLSTMb-MHAc-CRFMedicalNoYesChinese
Kong et al [14]Deep learningCNNdMedicalNoYesChinese
Li et al 2022 [5]Deep learningALBERTe-IDCNNf-MHA-CRFMedicalNoYesChinese
Li et al 2020 [5]Deep learningBERTg-BiLSTM-CRFMedicalNoYesChinese
Chen et al 2022 [6]Deep learningMC-BERT-BiLSTM-CNN-MHA-CRFMedicalYesYesChinese
Li et al 2023 [16]Deep learning
  • MC-BERT
  • GCNh-CRF
MedicalYesYesChinese
Zhou et al 2022,2023 [7,8]Deep learningCancerBERTCancerYesNoEnglish

aCRF: conditional random field.

bBiLSTM: bidirectional long short-term memory.

cMHA: multihead attention

dCNN: convolutional neural network.

eALBERT: A Lite Bidirectional Encoder Representations from Transformers.

fIDCNN: iterated dilated convolutional neural network.

gBERT: Bidirectional Encoder Representations from Transformers.

hGCN: graph convolutional network.


Ethical Considerations

Based on Article 32 of the Measures for Ethical Review of Life Sciences and Medical Research Involving Human Subjects [17] research using human data or biological samples that does not cause harm to the human body, does not involve sensitive personal information or commercial interests, is eligible for exemption from ethical review. This exemption is intended to reduce unnecessary burdens on researchers and to facilitate the progress of life sciences and medical research. In our study, the dataset used has undergone thorough anonymization, with all personal patient information removed. In addition, the patients involved in the clinical activities at the hospital have already signed the relevant informed consent forms, confirming their agreement to the use of their data for research purposes. Therefore, as per the national regulations, our study did not require ethical review. The data were granted permission to be accessed by SD, who is the director of the Information Center of Dongfang Hospital Beijing University of Chinese Medicine, responsible for managing all the hospital's data, and he is listed as a contributing author.

Model Construction

The proposed model architecture in this study is illustrated in Figure 1.

Figure 1. The architecture of the proposed model, for ease of viewing and understanding, the input text examples in the figure are accompanied by corresponding English translations. BiLSTM: bidirectional long short-term memory; CRF: conditional random field; DGCNN: dilated-gated convolutional neural network; LSTM: long short-term memory.

The model comprises 4 parts: the vector representation layer, feature extraction layer, feature fusion layer, and CRF layer. The Chinese cancer EHR texts are input into the model, where the input text is segmented into individual characters to form the basic sequence X={x1,x2,,xi}, where xi is the i-th character. Initially, in the vector representation layer, the input sequences X={x1,x2,,xi} are represented as semantically rich character-level embedding vectors hi using the autonomously trained BERT model tailored to the Chinese cancer domain (ChCancerBERT), forming the embedding sequence H={h1,h,,hi}. Subsequently, the sequence H={h1,h,,hi} is fed into the BiLSTM and dilated-gated convolutional neural network (DGCNN) models within the feature extraction layer. The feature extraction layer uses BiLSTM to extract long-distance dependency associations and temporal features from the sequences, while DGCNN captures spatial features. In the feature fusion layer, a MHA mechanism with enhanced generalization capability dynamically fuses the features extracted by BiLSTM and DGCNN to enhance the model’s feature representation capability. The CRF layer imposes constraints on the output to obtain the optimal labeling results. Finally, the model outputs a set of label sequences Y={y1,y,,yi}, where each label yi corresponds to a character in the input text. For instance, “B-TIME” denotes the beginning of a TIME entity, while “I-TIME” denotes the inside of a TIME entity, with the label sequence indicating the model’s classification results for each character.

Chinese Cancer Specialized BERT

To better capture semantics in cancer electronic medical records, this study trained a specialized BERT model tailored for the Chinese cancer domain (ChCancerBERT) to enhance the performance of extracting cancer-related phenotypes. Figure 2 illustrates the training process of the ChCancerBERT model.

Figure 2. The training process of ChCancerBERT models. BERT: Bidirectional Encoder Representations from Transformers; CNKI: China National Knowledge Infrastructure; MC: ;MLM: masked language model; NSP: next sentence prediction

Pretrained models are typically based on large-scale general corpora for data pre-training. Depending on the specific needs, they can also incorporate domain-specific vocabulary to calculate the probability P(S) of all possible language sequences.

P(S)=p(ω1,ω2,,ωn)=i=1np(ω|ω1,ω2,,ωi1)(1)

The most commonly used pretrained model is the BERT model, a deep bidirectional language representation model that offers significant advantages in capturing textual semantic information and entity extraction [18-20]. Although BERT achieves good results in most domains, its generic nature may limit its ability to fully capture the specialized terminology and semantic nuances of the medical field [15]. MC-BERT fine-tuned on BERT by using different pretraining strategies to incorporate medical-specific and related vocabulary, enabling superior performance in medical NER tasks [6].

The ChCancerBERT model proposed in this study is further pretrained based on the MC-BERT model. The specific steps are as follows: firstly, a Chinese cancer corpus was constructed. Using keywords “cancer” and “tumor,” a search strategy retrieved relevant Chinese-language literature from the China National Knowledge Infrastructure (CNKI) database. The corpus included abstracts from the first 6000 articles, organized chronologically, resulting in a comprehensive dataset of 3,268,570 tokens. This dataset ensures that ChCancerBERT is exposed to terminology and context unique to the cancer domain, thus improving its capacity for semantic understanding. Second, the pretraining of ChCancerBERT used both masked language model and next sentence prediction tasks to enhance semantic representation. Based on the masked language model, the approach involves randomly masking certain characters within a sentence to learn the semantic information of each character in context. This encourages ChCancerBERT to learn nuanced semantic relationships, particularly for specialized terminology within cancer-specific contexts. In addition, the next sentence prediction model is used to learn the association between 2 consecutive sentences, enabling it to capture long-distance dependencies and contextual flows frequently present in medical records. This step is particularly beneficial for understanding the multisentence descriptions often found in medical documents.

BiLSTM Model

Long short-term memory (LSTM) is a specialized form of recurrent neural network that can capture larger and longer-distance information. By using a set of gate controllers, LSTM effectively addresses the gradient vanishing and exploding problems inherent in traditional recurrent neural networks [21]. An LSTM, the input gate it controls information into the current cell, the forget gate ft manages the forgetting of information, and the output gate ot controls the information output. ht records the output at the current time step, while ct preserves the current cell state and serves as input for the next cell. The sigmoid activation function is denoted by σ, the weight matrix by W, the bias term by b, and the current cell input by xt. st represents the updated state at the current time step, and the hyperbolic tangent activation function is denoted by tanh:

{it=σ(Wi[ht1,xt]+bi)ft=σ(Wf[ht1,xt]+bf)ot=σ(Wf[ht1,xt]+bo)st=tanh(Ws[ht1,xt]+bs)ct=ftct1+itstht=ottanh(ct)(2)

BiLSTM consists of a forward LSTM and a backward LSTM, which respectively capture forward and backward information from the text sequence [22]. These are then concatenated to obtain the final hidden layer feature representation. Compared to a single LSTM model, BiLSTM provides a more nuanced understanding of bidirectional semantic dependencies and allows for more precise semantic discrimination [23].

In our model, BiLSTM is applied immediately after the vector representation layer, where ChCancerBERT generates the character-level embeddings tailored to the cancer domain. These embeddings are then fed into the BiLSTM, which consists of a forward and backward LSTMs, capturing both forward and backward contextual information. By leveraging BiLSTM, the model can retain bidirectional semantic dependencies that are crucial in understanding complex medical texts.

DGCNN Model

DGCNN is a neural network model with residual connections that uses gated dilated convolutions to gather more information through stacked convolutional layers [4]. Its convolutional structure, which enables the model to analyze sequences in segments, examining local patterns that are essential to textual spatial characteristics, which are usually referred to the arrangement and relationships of words and phrases that form meaningful units [4]. These units include nearby dependencies and longer-range relationships, like multiword medical entities in EHRs. DGCNN’s use of dilated convolutions allows it to capture these features at different scales or distances, enabling it to handle both close and distant relationships within the text [14]. The convolution operations proceed as follows:

ct=Wck=0rxt±k(3)

In the model, Wc denotes a filter with window size r, and ct represents the output after convolution transformation. The symbol ⊕ indicates vector concatenation. Building on this, dilation is introduced for interval sampling of data, enabling the coverage of longer sentences. The calculation method for dilated convolution is as follows:

ct=Wck=0rxt±kδ(4)

δ represents the dilation rate, and Wc is a filter with window size r. ct denotes the output, while ⊕ signifies vector concatenation. A gating mechanism is also incorporated to control data flow and reduce the risk of gradient vanishing. The operations within each convolutional gated unit are as follows:

Y=X+Conv1D1(X)σ(Conv2D2(X))(5)

X represents the input. Conv1D1 and Conv2D2 are dilated convolutions with the same parameter structure but without sharing weights. σ denotes the sigmoid function, and represents the Hadamard product of vectors.

After the ChCancerBERT model generates the character-level embeddings, these embeddings are input independently into the DGCNN model, extracting spatial dependencies and local features, which are essential for identifying nested entities and capturing complex structures within cancer-related records.

MHA Mechanism

The attention mechanism simulates human attention by scoring and weighting important content to highlight significant features [24]. The calculation formula is as follows:

Attention(Q,K,V)=Softmax(QKTdk)V(6)

The AttentionQ,K,Vvalue represents the attention score, where Qs the query vector matrix, K is the key vector matrix, and V is the value vector matrix. dk is the word vector dimension which normalizes the similarity between the query and key vectors calculated via QKT to prevent overly large results. Softmax normalization is then applied, followed by multiplication with the value vector to obtain the final attention value. When vector dimensions are high, a single attention calculation may not capture all features of a word. Thus, the MHA mechanism is used, linearly mapping Q,K,V into multiple subspaces and combining the calculated results:

{headi=Attention(QWiQ,KWiK,VWiV)Attention(Q,K,V)=Softmax(QKTdk)VMultiHead(Q,K,V)=Concat(head1,head2,,headk)WO(7)

This study uses an MHA mechanism to dynamically fuse the BiLSTM and DGCNN models. First, the outputs of both models are mapped to separate feature spaces, generating 2 sets of Q,K,V matrices. Next, the attention values for each model are calculated, resulting in 2 attention representation matrices. Finally, these matrices are combined into a single matrix through matrix multiplication.

CRF Model

CRF is a sequence modeling framework that incorporates constraints between labels to assign higher probabilities to reasonable label sequences [20]. This approach effectively addresses the issue of label bias.

The probability formula for obtaining the predicted label sequence X=(x1,x2,,xn)given the input sequence Y=(y1,y2,,yn) is as follows:

P(Y|X)=eT(X,Y)YYXT(X,Y)(8)

Given the true label sequence Y`, and all possible label sequences YX, the likelihood function for predicting the label sequence Y is as follows:

ln(P(Y|X))=T(X,Y)ln(YYXT(X,Y))(9)

By maximizing the likelihood function, the most probable and reasonable label sequence is determined and output. The specific formula is as follows:

Y=argyYXmaxs(X,y)(10)

Experiment Setups

Dataset Construction

The study’s data were sourced from the inpatient EHRs at a renowned hospital in Beijing, China, spanning from 2013 to 2022. After excluding patients with unrelated admission reasons (ie, those who have previously had breast cancer but whose current reason for admission is not breast cancer), the illness histories from all patients’ admission records within this time period were included. Each patient had 1 record per visit, and we ultimately obtained 876 records of patients with breast cancer (total tokens: 528,925, average length: 659.67, SD 354.23). The data received full approval from the hospital. All procedures in this research adhere to ethical standards. Patient privacy data were deidentified before manipulation, ensuring anonymity and confidentiality of personal information.

Annotation Procedure for the Dataset

The annotation of data from 876 admission records of patients with breast cancer was a collaborative effort involving 7 annotators. All of these annotators were the students of the Management School of Beijing University of Chinese Medicine and possessed relevant medical and computer expertise. To enhance the precision of the labeling outcomes, 2 oncology clinicians from the Beijing University of Chinese Medicine were enlisted to train the annotators. They also facilitated content segmentation and offered guidance throughout the labeling process. Based on information extracted from the collected data, in conjunction with insights from pertinent clinical studies [25,26] and guided by categorizations and classifications in clinical literature, as well as referencing indicators present in public medical databases such as SEER, CCKS2017, CCKS2019, along with existing study labeling protocols [27], an index system (Table 2) and labeling criteria (Table 3) were formulated for this research. These encompassed symptoms, tests, treatments, and time.

Table 2. Index system of the annotated dataset for Chinese breast cancer electrical health records (EHRs).
ClassEntity
SymptomsNodule
Tests
  • Vascular tumor thrombus and nerve aggress
  • Incisal edge
  • Lymph node dissection
  • Birads
  • TNMa
  • ERb
  • PRc
  • HER2d
  • Ki67
Treatments
  • Surgery name
  • Chemotherapy
  • Radiotherapy
  • Targeted therapy
  • Endocrine therapy
TimeTime

aTNM: TNM classification of malignant tumors.

bER: Estrogen receptor.

cPR: progesterone receptor.

dHER2: human epidermal growth factor receptor-2.

Table 3. Entity labeling standards with examplesa.
EntityStandardExampleExample in English
TimeDate and time in various formats2011年2011
Surgery nameName of a patient undergoing surgery左侧乳腺癌根治术Left breast cancer radical surgery
NoduleExamination results and location of breast nodules右乳腺肿物Right breast mass
ChemotherapyPatient chemotherapy regimen, medication, cycle术后行辅助化疗6周期 (紫杉醇+环磷酰胺,具体剂量不详)After surgery, undergo adjuvant chemotherapy for 6 cycles (paclitaxel+ cyclophosphamide, specific dosages not specified)
RadiotherapyThe patient’s radiotherapy regimen and frequency放疗25次,具体剂量不详Undergo 25 sessions of radiation therapy, specific dosage not specified
Targeted therapyTargeted therapy, drugs后续贯赫赛汀靶向治疗18次Followed by 18 sessions of trastuzumab targeted therapy
Endocrine therapyEndocrine treatment options, drugs阿那曲唑1mgQd口服内分泌治疗Anastrozole 1mg once daily, oral administration for endocrine therapy
Incisal edgeWhether the incisal edge of the patient is cancerous各皮肤切缘未见癌侵润No cancer infiltration is observed at the margins of the skin
Vascular tumor thrombus and nerve aggressionResults of examination of vascular tumor thrombus and nerve aggression侵及周围脂肪组织及神经纤维束,脉管中见癌栓Involvement of surrounding adipose tissue and nerve fiber bundles, with cancer emboli seen in blood vessels
Lymph node dissectionThe patient’s lymph node dissection results淋巴结未见转移癌 (0/21),腋窝淋巴结 (0/18), (右腋窝淋巴结)0/3No metastatic cancer detected in lymph nodes (0/21), axillary lymph nodes (0/18), (right axillary lymph nodes) 0/3
Bi-RADSbBi-RADS classification of breast nodulesBI_RADS4b类bBI-RADS4bb
TNMcThe patients’ TNM stagespT2N0M0, IIa期pT2N0M0,IIa
ERdPatient pathology reported ER resultsER (+++)ER (+++)
PRePatient pathology reported PR resultsPR (+++)PR (+++)
HER2fHER2 results were reported pathologicallyHER2 (–)HER2 (–)
Ki67Patient pathology reported KI67 resultski-67index约20%Ki-67 index is approximately 20%

aAll texts to be annotated are in Chinese; for ease of viewing and understanding, the annotated examples in the table are accompanied by corresponding English translations, the patient information presented is simulated.

bBi-RADS: breast imaging reporting and data system.

cTNM: TNM classification of malignant tumors.

dER: estrogen receptor.

ePR: progesterone receptor.

fHER2: human epidermal growth factor receptor-2.

The “symptom” labels serve to indicate pathophysiological alterations and breast cancer symptoms, including details about breast nodules such as their location and dimensions. The “tests” label designates the array of tests undergone by patients with breast cancer during diagnosis and treatment, aiding physicians in precise assessments. These encompass parameters like resection margins, vascular thrombosis, nerve invasion, lymph node dissection, Bi-RADS grading, TNM staging, and immunohistochemical markers (ER, PR, HER2, and Ki-67). The “treatments” label encompasses therapies previously administered to patients with breast cancer, comprising chemotherapy, radiation, targeted therapy, endocrine therapy, and surgery. The “time” label incorporates details regarding each temporal point within the admission record of a patient with breast cancer.

Data annotation was performed using Label Studio software (Human Signal). A total of 7 trained annotators, each responsible for a distinct portion of the dataset, performed the initial labeling in strict accordance with the predefined annotation guidelines. The annotated corpus was then independently reviewed and revised by 2 experienced clinical oncologists, resulting in 2 expert-annotated versions. To assess interannotator agreement between the 2 experts, Cohen κ coefficient was calculated, which indicated almost perfect agreement (κ=0.82). Any discrepancies between the 2 expert versions were subsequently identified and resolved through discussion, leading to the final consensus gold-standard dataset. This finalized version was used for training and validation of the entity recognition model.

Data Partitioning

For evaluation, we use the hold-out method, partitioning the dataset randomly into training, validation, and test sets, distributed at a ratio of 6:2:2. Diverging from typical NER tasks, this model primarily deals with medical nouns or statements as named entities, with a particular focus on breast cancer–related terms. Table 4 illustrates the count of entities corresponding to each label within the training, validation, and test sets. Concurrently, the length of labels within each category was tabulated, and the outcomes are presented in Table 4.

Table 4. Distribution and length statistics of entities in training, validation, and test sets.
Class and entityTrain set, n (%)Development set, n (%)Test set, n (%)Entities length
MinimumMaximumMean (SD)
Symptoms
Nodule600 (61.66)192 (19.73)181 (18.60)312315.49 (17.84)
Tests
Vascular tumor thrombus and nerve aggress208 (56.68)76 (20.71)83 (22.62)44213.93 (7.79)
Incisal edge155 (58.49)53 (20)57 (21.51)47524.15 (17.06)
Lymph node dissection558 (62.21)137 (15.27)202 (22.52)413927.92 (24.67)
Bi-RADSa161 (54.39)58 (19.59)77 (26.01)88248.59 (22.54)
TNMb55 (57.89)15 (15.79)25 (26.32)51710.97 (3.76)
ERc438 (61.78)132 (18.62)139 (19.61)3459.7 (5.52)
PRd430 (61.69)131 (18.79)136 (19.51)2335.61 (4.88)
HER2e393 (61.50)117 (18.31)129 (20.19)58213.64 (10.36)
Ki67420 (62.22)132 (19.56)123 (18.22)75914.66 (7.06)
Treatments
Surgery name985 (60.39)304 (18.64)342 (20.97)37418.91 (9.06)
Chemotherapy1027 (61.72)319 (19.17)318 (19.11)2940101.43 (107.68)
Radiotherapy177 (62.32)56 (19.72)51 (17.96)214121.32 (21.42)
Targeted therapy155 (65.96)30 (12.77)50 (21.28)535268.83 (68.16)
Endocrine therapy229 (61.89)71 (19.19)70 (18.92)520232.55 (26.9)
Time
Time2765 (59.54)892 (19.21)987 (21.25)239556.41 (55.25)

aBi-RADS: breast imaging reporting and data system.

bTNM: TNM classification of malignant tumors.

cER: estrogen receptor.

dPR: progesterone receptor.

eHER2: human epidermal growth factor receptor-2.

Evaluation Metrics

During the experiment, the evaluation of the NER model used 3 metrics: precision (P), recall (R), and F1-score (F1). The formulas for calculating precision, recall, and F1-score are as follows:

{P=TPTP+FP×100%R=TPTP+FN×100%F1=2PRP+R×100%(11)

where TP represents the number of true positive entities, FP indicates the number of false positive entities, and FN denotes the number of false negative entities.

Implementation Details

The experiments were conducted using the PyTorch 1.6 framework in a Python (version 3.7; Python Software Foundation) environment. The CPU used is Intel(R) Core i5-12500H, and the GPU used is NVIDIA T4 GPU. The ChCancerBERT model parameters were configured identically to the BERT model, with 12 transformer layers, 12 attention heads, and 768 hidden units. The hyperparameters for the NER experiment are detailed in Table 5.

Table 5. Hyperparameter settings.
HyperparametersValue
Batch size8
Dropout0.5
Learning rate0.00003
Epoch64
Maximum length128
Embedding768

Using the partitioned dataset, we devised comparative experiments, ablation experiments, and separate pretrained model comparative experiments to facilitate a comprehensive assessment of the recognition efficacy across various mechanisms and models.

Results of the Comparative Experiment for NER Models

We applied our proposed model and the NER models mentioned in Table 1, which have been developed for medical and cancer fields in recent years, to our dataset to compare performance. Detailed results can be found in Table 6. We did not include Zhou et al CancerBERT model [7] in the comparison because it is not yet publicly available. Our results demonstrate that, for the Chinese breast cancer EHR dataset we constructed, our proposed model outperformed existing models, with F1-scores increasing by 1.40%, 2.57%, and 1.13%, respectively.

Table 6. Comparative performance of baseline and proposed models on the breast cancer dataset.
ModelPrecision (%)Recall (%)F1-score (%)
BERTa- BiLSTMb -CRF [5]85.7785.3885.58
ALBERTc-IDCNNd-MHA-CRF [15]87.7881.2884.41
MC-BERT-BiLSTM-CNNe-MHAf-CRFg [6]85.6486.0785.85
Our proposed model87.6386.3486.98

aBERT: Bidirectional Encoder Representations from Transformers.

bBiLSTM: bidirectional long short-term memory.

cALBERT: A Lite Bidirectional Encoder Representations from Transformers.

dIDCNN: iterated dilated convolutional neural network.

eCNN: convolutional neural network.

fMHA: multihead attention.

gCRF: conditional random field.

Results of the Ablation Experiment

We employ the fundamental BERT-BiLSTM-CRF model as a baseline to assess the impact of distinct mechanisms on the recognition of named entities within Chinese EHRs pertaining to breast cancer. Subsequently, BERT-DGCNN-CRF replaces the BiLSTM model within the baseline, while the baseline+DGCNN model merges the BiLSTM and DGCNN models. In addition, the baseline+DGCNN+MHA model introduces a MHA mechanism to the BERT-BiLSTM-DGCNN-CRF foundation. The corresponding experimental outcomes are detailed in Table 7.

Analysis of the ablation outcomes reveals that the substitution of the BiLSTM model with the DGCNN model results in a reduction of the precision and F1-score on the dataset of this study by 0.37% and 0.14%. However, the amalgamation of the DGCNN model and the BILSTM model yields F1-score increments of 0.35% and 0.49% compared to both the baseline and BERT-DGCNN-CRF models. This amalgamation also improves precision by 0.02% and 0.29%, and recall by 0.69%, compared to the baseline and BERT-DGCNN-CRF models, respectively. This observation underscores that, within the context of feature extraction, the standalone DGCNN model presents no notable superiority over the BiLSTM model. Conversely, the combined application of these 2 models augments recognition effectiveness. Building upon this foundation, the incorporation of the MHA serves to enhance the extraction of pertinent information within the data, thereby mitigating the impact of irrelevant data. Consequently, the F1-score experiences a supplementary enhancement of 0.17%, with a precision decrease of 0.46% offset by a substantial recall increase of 0.82%. This suggests that although the addition of the MHA slightly reduces precision, it compensates for this by capturing a broader range of relevant information, ultimately enhancing overall recognition effectiveness. Thus, based on the insights garnered from ablation experiments, it is ascertained that the optimal approach for this study involves the incorporation of the BiLSTM-DGCNN fusion model as the feature extraction layer, accompanied by the integration of the MHA mechanism.

Table 7. Results of ablation studies and pretrained model comparisons.
ModelPrecision (%)Recall (%)F1-score (%)
BERTa-BiLSTMb-CRFc (baseline)85.7785.3885.58
BERT-DGCNNd-CRF85.5085.3885.44
BERT-DGCNN-BiLSTM-CRF85.7986.0785.93
BERT-DGCNN-BiLSTM-MHAe-CRF85.3386.8986.10
ALBERTf-BiLSTM-CRF87.5077.3282.10
ALBERT-DGCNN-BiLSTM-MHA-CRF86.6184.9885.78
RoBERTag-BiLSTM-CRF86.1482.5184.29
RoBERTa-DGCNN-BiLSTM-MHA-CRF86.8685.7986.32
XLNeth-BiLSTM-CRF86.6678.5582.40
XLNet-DGCNN-BiLSTM-MHA-CRF87.5285.9386.72
MC-BERT-BiLSTM-CRF85.3685.2585.30
MC-BERT-DGCNN-BiLSTM-MHA-CRF86.9886.4886.73
ChCancerBERT-BiLSTM-CRF85.7586.6186.18
Our proposed model87.6386.3486.98

aBERT: Bidirectional Encoder Representations from Transformers.

bBiLSTM: bidirectional long short-term memory.

cCRF: conditional random field.

dDGCNN: dilated gated convolutional neural network.

eMHA: multihead attention.

fALBERT: A Lite Bidirectional Encoder Representations from Transformers.

gRoBERTa: robustly optimized Bidirectional Encoder Representations from Transformers pretraining approach.

hXLNet: generalized autoregressive pretraining.

Results of the Comparative Experiment for the Pretrained Models

Leveraging the model exhibiting the highest F1-score from the aforementioned experiments, we proceeded to perform comparative experiments by substituting various pretrained models. Comparison experiments were conducted by integrating different pretrained models with both BiLSTM+CRF and DGCNN-BiLSTM-MHA-CRF architectures. All used pretrained models are available for download from the “Models-Hugging Face” website. The corresponding outcomes are presented in Table 7.

The results indicate that using the proposed pretraining model, ChCancerBERT, for character-level embedding outperforms various pretrained models compared in the experiment. Specifically, in the comparison experiment combined with BiLSTM+CRF, the F1-scores of the ChCancerBERT model increased by 0.6%, 4.08%, 1.89%, 3.78%, and 0.88% compared to BERT, ALBERT, RoBERTa, XLNet, and MC-BERT, respectively. In the comparison experiment with the DGCNN-BiLSTM-MHA-CRF architecture, it achieved F1-score improvements of 0.88%, 1.20%, 0.66%, 0.26%, and 0.25% over BERT, ALBERT, RoBERTa, XLNet, and MC-BERT, respectively. In addition, regardless of the pretraining model used, the F1-scores obtained with the DGCNN-BiLSTM-MHA-CRF architecture were consistently higher than those obtained with the BiLSTM+CRF model alone, further demonstrating that the DGCNN-BiLSTM-MHA model, by integrating multidimensional text features, possesses a stronger semantic representation capability than the BiLSTM model alone.

After using ChCancerBERT for character-level embedding and integrating it with the DGCNN-BILSTM-MHA-CRF model to fuse multidimensional text features, the semantic representation capability is further enhanced, and the F1-score had a 1.4% improvement compared to the current mainstream BERT-BiLSTM-CRF (baseline) model. In addition, our model achieves a precision of 87.63% and a recall of 86.34%, which represents increases of 1.86% in precision and 0.96% in recall compared to the baseline BERT-BiLSTM-CRF model. This reflects both high detection accuracy and robustness in capturing target entities.

Error Analysis

We conducted a comprehensive analysis of the 4 main entity categories in this study, including their quantity, average length, as well as precision, recall, and F1-scores. The specific results are presented in Table 8.

The results presented in Table 8 indicate that TIME entities are the most frequently occurring, totaling 4644 instances. The model demonstrates the highest recognition effectiveness for these entities, achieving an F1-score of 92.79%. In contrast, although the treatments category also has a high number of entities, it shows the lowest recognition performance with an F1-score of 74.24%. This lower performance can be attributed to the complexity and variability of treatment-related terms, which is further compounded by the fact that the average length of treatment entities is 17.27 characters, significantly longer than other entity categories. Treatment entities often contain multiple details, including various drugs, dosages, treatment methods, and frequencies, such as in the example: “Chemotherapy: Cyclophosphamide 1000 mg on day 1, Epirubicin 50 mg on day 1, 21 days per cycle, 4 cycles.” However, the model only identifies the drugs and their dosages used in chemotherapy, while ignoring the subsequent frequency information. The increased length and complexity of these entities make it more challenging for the model to accurately identify and classify them, resulting in a decrease in overall recognition accuracy.

Table 8. Performance obtained for each entity.
ClassEntity numberAverage lengthPrecision (%)Recall (%)F1-score (%)
Symptoms9737.6282.8684.8983.86
Tests6407.8282.6181.4081.02
Treatments418417.2776.9072.3074.24
Time46448.0192.6292.9692.79

Experiment on CCKS2019

In order to comprehensively validate the efficacy of the model proposed in this study, we applied it to the CCKS2019 dataset and conducted a performance comparison with models recently introduced in the literature. The CCKS2019 dataset, publicly accessible and broadly used within China’s medical NLP research community, offers a standardized evaluation platform for diverse NER models (CCKS2019 [28]). In total, we compared 8 models, including 3 models in 5.1, and the other 5 models are as follows:

  • Embeddings from language models (ELMo)-lattice-LSTM-CRF model: this model uses the lattice LSTM-CRF architecture and integrates variant contextual character representations (Chinese Embeddings from Language Models [ELMo]) to enhance the use of character and word information in Chinese EHRs [29].
  • ELMo-encoder from transformer–CRF model: this model fine-tunes the ELMo model using domain-specific clinical records and uses transformer (encoder from transformer) as an encoder to address long-context dependencies [30].
  • All CNN model: this model incorporates Chinese radical characters to enhance the morphosemantic representation of Chinese characters [14].
  • Graph attention network–BiLSTM-CRF model: this model uses a graph neural network to model lexical dependency inertia as syntactic semantics and integrates semantic information at various levels to enhance contextual representation [31].
  • Chinese medical–NER model: this model enhances the MC-BERT model by incorporating graph convolutional networks and CRFs [16].

Table 9 presents a comparison of the performance of our proposed model on the CCKS2019 dataset with that of models introduced in recent literature. Our model achieves an F1-score of 87.26%, surpassing recent models by margins of 2.64%, 2.24%, 2.13%, 1.67%, 1.63%, 1.34%, 0.99%, and 0.97%. These results strongly affirm the effectiveness of our model. The experimental findings clearly demonstrate that our cancer-specific named NER model outperforms existing models on the CCKS2019 dataset, highlighting its potential for effectively handling larger medical datasets.

Table 9. Performance comparison of the proposed model and recent models on the CCKS2019 dataset.
ModelPrecision (%)Recall (%)F1-score (%)
BERTa- BiLSTMb -CRFc [5]82.0987.3284.62
ELMod-lattice-LSTMe-CRF [29]85.3584.6985.02
ACNNf [14]87.2983.0785.13
ELMo-ETg-CRF model [30]87.6183.6585.59
ALBERTh-IDCNNi-MHAj-CRF [15]84.8286.4685.63
GATk-BiLSTM-CRF [31]86.7485.1185.92
MC-BERT-BiLSTM-CNNl-MHA-CRF [6]84.9087.6786.27
CMm-NERn [16]86.4586.1386.29
Our proposed model87.2687.2787.26

aBERT: Bidirectional Encoder Representations from Transformers.

bBiLSTM: bidirectional long short-term memory.

cCRF: conditional random field.

dELMo: embeddings from language models.

eLSTM: long short-term memory.

fACNN: all convolutional neural network.

gET: encoder from transformer.

hALBERT: A Lite Bidirectional Encoder Representations from Transformers.

iIDCNN: iterated dilated convolutional neural network.

jMHA: multihead attention.

kGAT: graph attention network.

lCNN: convolutional neural network.

mCM: Chinese medical.

nNER: named entity recognition.


Principal Findings

Unstructured EHR data contains valuable patient information that can be used for clinical decision support and research [32]. In the Chinese breast cancer EHR corpus constructed in this study, there are 876 admission records, comprising a total of 528,925 tokens. Annotations were performed for 4 major categories and 17 subcategories of named entities, with each subcategory having more than 100 annotated instances. In the NER task, the model we propose outperforms other models currently applied in medical-related fields. The results of ablation experiments revealed that the fusion of the pretrained model with the DGCNN-BILSTM-MHA-CRF model effectively enhances entity recognition performance. Specifically, the introduction of the DGCNN structure on top of the baseline results in the maximum improvement of 0.35%. This enhancement is attributed to the structure’s ability to mitigate the risk of gradient vanishing and facilitate information transmission across multiple channels, capturing a wealth of feature information. In addition, the incorporation of the MHA mechanism on this foundation not only enhances the model’s generalization capability but also enables parallel attention computation, strengthening feature fusion capability [24]. This leads to improved use of valid information and reduced impact of irrelevant information, resulting in a further F1-score improvement of 0.17%.

Comparing the results of the pretrained model experiments, it is evident that the proposed ChCancerBERT pretrained model outperforms various state-of-the-art BERT models, with the medical domain-specific MC-BERT model ranking second. This is attributed to MC-BERT incorporating more medical corpus training into BERT, and ChCancerBERT further including the Chinese cancer corpus on top of MC-BERT. This augmentation effectively enhances the model’s semantic representation capability in the Chinese cancer domain, yielding superior results in the Chinese breast cancer corpus in this study. In addition, the study’s results indicate that pretrained BERT models based on specific domains can lead to improved performance in downstream tasks [7].

Among the 4 categories extracted in this study, the performance in extracting time entities stood out, with an impressive F1-score of 92.79%. This result can be attributed to the relatively high quantity of time entities in this dataset, amounting to 4644, and the fact that time entities exhibit a more standardized writing format compared to other entities, which, to some extent, reduces the difficulty of recognition and enhances recognition accuracy [33].

We scrutinize the entity recognition outcomes of the most proficient fusion model derived from our experiments. Categories with infrequent occurrences in the dataset, such as TNM, Bi-RADS, and some other entities in class “tests,” are prone to omission errors. These errors manifest as instances where entities that should have been identified remain undetected, suggesting a propensity of the model used in this study to overlook less common entity categories [34]. Such omissions may result in incomplete information extraction, which is crucial for disease diagnosis and the formulation of treatment plans. Therefore, downstream applications should incorporate clinician verification mechanisms or error correction modules to ensure safety in clinical use.

In this study, entities characterized by considerable length, such as chemotherapy, radiotherapy, and other in-class “treatments,” are typically classified correctly by the model. However, the model is susceptible to boundary recognition errors [7], leading to either the exclusion of words at the sentence’s periphery or the inclusion of extraneous words beyond the sentence, which may cause confusion in clinical documentation. Furthermore, on one hand, Chinese words frequently exhibit polysemy and ambiguity [33]; on the other hand, the dataset for this study contains instances where statements are nested within the entity categories of other entities. Such intricacies heighten the risk of erroneous entity category determinations.

Concurrently, we apply the model introduced in this research to the publicly accessible CCKS2019 dataset, and the findings illustrate its superior performance compared to prior research efforts, suggesting the ChCancerBERT framework has the potential to be extended to other cancer types and even broader clinical domains. Nevertheless, differences in documentation style and terminology across hospitals may affect generalization. Domain adaptation strategies such as continual pretraining on local EHR corpora and transfer learning can be explored to improve portability to other institutions.

Using our proposed model, vital clinical data encompassing timestamps, examination details, symptoms, and treatment modalities can be readily extracted from the EHRs of patients with cancer. The extracted entities provide important value for both clinical and research applications. Time, symptom, test, and treatment information can be structured into patient timelines to support disease monitoring, clinical decision support, population health analysis, and trial recruitment. These entities can also be mapped to standards such as International Classification of Diseases, 10th Revision, enabling interoperability and secondary use of structured cancer data.

Conclusions

This study designed and developed a pretrained model specifically tailored to the Chinese cancer domain, named ChCancerBERT. The model was then integrated with various NER models proposed in the field of computer science, resulting in the ChCancerBERT-DGCNN-BILSTM-MHA-CRF model. This integrated model was used to extract 4 major types of entities related to cancer from admission records in the oncology department of hospitals. We applied the model proposed in this study to Chinese breast cancer EHR data and CCKS2019 for experimentation and validation. The proposed model achieved an F1-score of 86.98%, surpassing other models compared in the experiment.

The primary limitation of this study is the restricted data accessibility and relatively small sample size. This limited sample size may affect the statistical power and robustness of the findings. Moreover, potential variations in clinical documentation practices across different regions or institutions in China could pose challenges to the generalizability of the model. Despite these limitations, it is worth noting that the model demonstrated promising transferable capability by achieving competitive performance on the CCKS2019 medical dataset, which suggests its potential applicability beyond breast cancer data. Therefore, future research should aim to acquire EHR data for other types of cancer to further validate the generalizability of the proposed model. Furthermore, a NER system tailored for clinical application is envisaged to be constructed based on this model. This system aims to assist clinical researchers in data organization and extraction. In addition, plans encompass the establishment of a clinical decision support system to facilitate more effective treatment for patients.

Acknowledgments

This work is supported by Beijing University of Chinese Medicine Independent Topic Selection Young Faculty Project (grant: 040201004002011), 2024 Independent Research Projects for Postgraduate Students of Beijing University of Chinese Medicine (grant: ZJKT2024033), 2022 Basic Research Business Fund ’Top-Down Leadership’ Project (grant: 2022JYBJBRW012), the Fundamental Research Funds for the Central Universities (grant: 2025JYBXJSJJ022), Beijing University of Chinese Medicine Educational and Teaching Reform Special Project (grant: XJYZ25010), Beijing Social Science Fund (grant: 24XCC022), Research Funding on TCM Monitoring and Statistics for 2025 of the Monitoring and Statistics Center of the National Administration of Traditional Chinese Medicine (grant: 2025JCTJE50), 'Top-Down Leadership’ Educational Reform Project of Beijing University of Chinese Medicine (grant: JXJBGS2501005), and the 2025 Basic Research Business Fund ’Top-Down Leadership’ Project (grant: 2025JYBJBGS009).

Data Availability

The annotated datasets generated and analyzed during the current study are not publicly available due to hospital policy and regulation, but the ChCancerBERT model is available from the corresponding author on reasonable request.

Authors' Contributions

JC and SD participated in data collection and collation. JC, BZ, XT, and ZZ participated in the method design, analyzed data, and drafted the initial manuscript. JC, BZ, XT, ZZ, and JW participated in text checking correction and helped to draft the manuscript. FG and RW oversaw and provided input on all aspects of manuscript writing and the final analytical plan. All the authors read and approved the final manuscript.

Conflicts of Interest

None declared.

  1. Deimazar G, Sheikhtaheri A. Machine learning models to detect and predict patient safety events using electronic health records: a systematic review. Int J Med Inform. Dec 2023;180:105246. [CrossRef] [Medline]
  2. Chen PE, Hung HH, Huang WT, et al. Chinese herbal medicine improved survival in stage IV breast cancer patients: data mining of the Incorporated Taiwan Cancer Registry Database and hospital database. Integr Cancer Ther. 2023;22:15347354231178898. [CrossRef] [Medline]
  3. Zhao X, Tong Z, Sun L, et al. Clinical characteristics, treatment patterns, and effectiveness in Chinese patients with angina pectoris using electronic patient-reported outcomes: protocol for a multicenter, prospective, cohort study (GREAT). Adv Ther. Apr 2023;40(4):1899-1912. [CrossRef] [Medline]
  4. Abdullah MHA, Aziz N, Abdulkadir SJ, Alhussian HSA, Talpur N. Systematic literature review of information extraction from textual data: recent methods, applications, trends, and challenges. IEEE Access. 2023;11:10535-10562. [CrossRef]
  5. Li X, Zhang H, Zhou XH. Chinese clinical named entity recognition with variant neural structures based on BERT methods. J Biomed Inform. Jul 2020;107:103422. [CrossRef] [Medline]
  6. Chen P, Zhang M, Yu X, Li S. Named entity recognition of Chinese electronic medical records based on a hybrid neural network and medical MC-BERT. BMC Med Inform Decis Mak. Dec 1, 2022;22(1):315. [CrossRef] [Medline]
  7. Zhou S, Wang N, Wang L, Liu H, Zhang R. CancerBERT: a cancer domain-specific language model for extracting breast cancer phenotypes from electronic health records. J Am Med Inform Assoc. Jun 14, 2022;29(7):1208-1216. [CrossRef] [Medline]
  8. Zhou S, Wang N, Wang L, et al. A cross-institutional evaluation on breast cancer phenotyping NLP algorithms on electronic health records. Comput Struct Biotechnol J. 2023;22:32-40. [CrossRef] [Medline]
  9. Najafabadipour M, Tunas JM, Rodriguez-Gonzalez A. Lung cancer concept annotation from Spanish clinical narratives. Presented at: 13th International Conference on Data Integration in the Life Sciences 2018 (DILS2018); Nov 20-21, 2018. [CrossRef]
  10. Yim WW, Kwan SW, Yetisgen M. Tumor reference resolution and characteristic extraction in radiology reports for liver cancer stage prediction. J Biomed Inform. Dec 2016;64:179-191. [CrossRef] [Medline]
  11. Savova GK, Tseytlin E, Finan S, et al. DeepPhe: A natural language processing system for extracting cancer phenotypes from clinical records. Cancer Res. Nov 1, 2017;77(21):e115-e118. [CrossRef] [Medline]
  12. Weegar R, Kvist M, Sundström K, Brunak S, Dalianis H. Finding cervical cancer symptoms in Swedish clinical text using a machine learning approach and NegEx. AMIA Annu Symp Proc. 2015;2015:1296-1305. URL: https://pmc.ncbi.nlm.nih.gov/articles/PMC4765575/ [Accessed 2025-11-03] [Medline]
  13. An Y, Xia X, Chen X, Wu FX, Wang J. Chinese clinical named entity recognition via multi-head self-attention based BiLSTM-CRF. Artif Intell Med. May 2022;127:102282. [CrossRef] [Medline]
  14. Kong J, Zhang L, Jiang M, Liu T. Incorporating multi-level CNN and attention mechanism for Chinese clinical named entity recognition. J Biomed Inform. Apr 2021;116:103737. [CrossRef] [Medline]
  15. Li D, Long J, Qu J, Zhang X. Chinese clinical named entity recognition with ALBERT and MHA mechanism. Evid Based Complement Alternat Med. 2022;2022(2056039):2056039. [CrossRef] [Medline]
  16. Li M, Gao CR, Zhang K, Zhou HJ, Ying J. A weakly supervised method for named entity recognition of Chinese electronic medical records. Med Biol Eng Comput. Oct 2023;61(10):2733-2743. [CrossRef] [Medline]
  17. Measures for ethical review of life sciences and medical research involving human subjects [Website in Chinese]. National Health Commission. 2023. URL: https://www.gov.cn/zhengce/zhengceku/2023-02/28/content_5743658.htm [Accessed 2025-10-30]
  18. Khan MA, Huang Y, Feng J, et al. A multi-attention approach using BERT and stacked bidirectional LSTM for improved dialogue state tracking. Appl Sci (Basel). 2023;13(3):1775. [CrossRef]
  19. Sun J, Liu Y, Cui J, He H. Deep learning–based methods for natural hazard named entity recognition. Sci Rep. Mar 17, 2022;12(1):4598. [CrossRef] [Medline]
  20. Li D, Yan L, Yang J, Ma Z. Dependency syntax guided BERT-BiLSTM-GAM-CRF for Chinese NER. Expert Syst Appl. Jun 2022;196:116682. [CrossRef]
  21. Runmei Z, Lulu L, Lei Y, et al. Chinese named entity recognition method combining ALBERT and a local adversarial training and adding attention mechanism. Int J Semant Web Inf Syst. Jan 2022;18(1):1-20. URL: https://services.igi-global.com/resolvedoi/resolve.aspx?doi=10.4018/IJSWIS.20220101 [CrossRef]
  22. Zhou F, Wang C, Wang J, Zhang J. Named entity recognition of ancient poems based on Albert‐BiLSTM‐MHA‐CRF model. Wirel Commun Mob Comput. Jan 2022;2022(1):1-11. URL: https://onlinelibrary.wiley.com/toc/6302/2022/1 [CrossRef]
  23. Xu K, Yang Z, Kang P, Wang Q, Liu W. Document-level attention-based BiLSTM-CRF incorporating disease dictionary for disease named entity recognition. Comput Biol Med. May 2019;108:122-132. [CrossRef] [Medline]
  24. Zhu ZC, Li JQ, Zhao Q, Akhtar F. A dictionary-guided attention network for biomedical named entity recognition in Chinese electronic medical records. Expert Syst Appl. Nov 2023;231:120709. [CrossRef]
  25. Frei J, Kramer F. Annotated dataset creation through large language models for non-English medical NLP. J Biomed Inform. Sep 2023;145:104478. [CrossRef] [Medline]
  26. Lingren T, Deleger L, Molnar K, et al. Evaluating the impact of pre-annotation on annotation speed and potential bias: natural language processing gold standard development for clinical named entity recognition in clinical trial announcements. J Am Med Inform Assoc. 2014;21(3):406-413. [CrossRef] [Medline]
  27. Uzuner O, Solti I, Xia F, Cadag E. Community annotation experiment for ground truth generation for the i2b2 medication challenge. J Am Med Inform Assoc. 2010;17(5):519-523. [CrossRef] [Medline]
  28. China conference on knowledge graph and semantic computing. URL: https://www.sigkg.cn/ccks2019/ [Accessed 2025-11-14]
  29. Li Y, Wang X, Hui L, et al. Chinese clinical named entity recognition in electronic medical records: development of a lattice long short-term memory model with contextualized character representations. JMIR Med Inform. Sep 4, 2020;8(9):e19848. [CrossRef] [Medline]
  30. Wan Q, Liu J, Wei LN, Ji B. A self-attention based neural architecture for Chinese medical named entity recognition. Math Biosci Eng. May 9, 2020;17(4):3498-3511. [CrossRef] [Medline]
  31. Shi J, Sun M, Sun Z, Li M, Gu Y, Zhang W. Multi-level semantic fusion network for Chinese medical named entity recognition. J Biomed Inform. Sep 2022;133:104144. [CrossRef] [Medline]
  32. Landolsi MY, Hlaoua L, Ben Romdhane L. Information extraction from electronic medical documents: state of the art and future research directions. Knowl Inf Syst. 2023;65(2):463-516. [CrossRef] [Medline]
  33. Qiu Q, Tian M, Huang Z, et al. Chinese engineering geological named entity recognition by fusing multi-features and data enhancement using deep learning. Expert Syst Appl. Mar 2024;238:121925. [CrossRef]
  34. Yu Y, Wang Y, Mu J, et al. Chinese mineral named entity recognition based on BERT model. Expert Syst Appl. Nov 2022;206:117727. [CrossRef]


BiLSTM: bidirectional long short-term memory.
DGCNN: dilated-gated convolutional neural network
EHR: electronic health record
ELMo: embeddings from language models
LSTM: long short-term memory
MHA: multihead attention
NER: named entity recognition
NLP: natural language processing


Edited by Arriel Benis; submitted 04.May.2025; peer-reviewed by Cheligeer Cheligeer, Dillon Chrimes, Praveen Tekkali; final revised version received 08.Oct.2025; accepted 13.Oct.2025; published 14.Nov.2025.

Copyright

© Junbai Chen, Butian Zhao, Xiaohan Tian, Zhengkai Zou, Ruojia Wang, Jiarui Wu, Songxing Du, Fengying Guo. Originally published in JMIR Medical Informatics (https://medinform.jmir.org), 14.Nov.2025.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on https://medinform.jmir.org/, as well as this copyright and license information must be included.