TY - JOUR AU - Zhang, Zhichang AU - Zhu, Lin AU - Yu, Peilin PY - 2020 DA - 2020/5/4 TI - Multi-Level Representation Learning for Chinese Medical Entity Recognition: Model Development and Validation JO - JMIR Med Inform SP - e17637 VL - 8 IS - 5 KW - medical entity recognition KW - multi-level representation learning KW - Chinese KW - natural language processing KW - electronic medical records KW - multi-head attention mechanism AB - Background: Medical entity recognition is a key technology that supports the development of smart medicine. Existing methods on English medical entity recognition have undergone great development, but their progress in the Chinese language has been slow. Because of limitations due to the complexity of the Chinese language and annotated corpora, these methods are based on simple neural networks, which cannot effectively extract the deep semantic representations of electronic medical records (EMRs) and be used on the scarce medical corpora. We thus developed a new Chinese EMR (CEMR) dataset with six types of entities and proposed a multi-level representation learning model based on Bidirectional Encoder Representation from Transformers (BERT) for Chinese medical entity recognition. Objective: This study aimed to improve the performance of the language model by having it learn multi-level representation and recognize Chinese medical entities. Methods: In this paper, the pretraining language representation model was investigated; utilizing information not only from the final layer but from intermediate layers was found to affect the performance of the Chinese medical entity recognition task. Therefore, we proposed a multi-level representation learning model for entity recognition in Chinese EMRs. Specifically, we first used the BERT language model to extract semantic representations. Then, the multi-head attention mechanism was leveraged to automatically extract deeper semantic information from each layer. Finally, semantic representations from multi-level representation extraction were utilized as the final semantic context embedding for each token and we used softmax to predict the entity tags. Results: The best F1 score reached by the experiment was 82.11% when using the CEMR dataset, and the F1 score when using the CCKS (China Conference on Knowledge Graph and Semantic Computing) 2018 benchmark dataset further increased to 83.18%. Various comparative experiments showed that our proposed method outperforms methods from previous work and performs as a new state-of-the-art method. Conclusions: The multi-level representation learning model is proposed as a method to perform the Chinese EMRs entity recognition task. Experiments on two clinical datasets demonstrate the usefulness of using the multi-head attention mechanism to extract multi-level representation as part of the language model. SN - 2291-9694 UR - https://medinform.jmir.org/2020/5/e17637 UR - https://doi.org/10.2196/17637 UR - http://www.ncbi.nlm.nih.gov/pubmed/32364514 DO - 10.2196/17637 ID - info:doi/10.2196/17637 ER -