Chinese Medical Named Entity Recognition Based on Context-Dependent Perception and Novel Memory Units
Abstract
:1. Introduction
- (1)
- New memory unit introduction: The model integrates novel memory units (GLMU and RMN) to enhance its capacity for capturing long-distance dependencies. These memory units more effectively preserve and utilize contextual information, particularly in the processing of lengthy texts and complex sentences, significantly improving NER accuracy.
- (2)
- Fully connected layer optimization and application: The fully connected layer optimizes feature representation and classification performance in the model. By integrating and mapping high-dimensional features from earlier layers, it enhances classification accuracy, allowing for more precise identification of named entities.
- (3)
- Complex named entity structure handling: The model effectively processes complex named entity structures. It accurately captures internal entity structures and relationships through the combined efforts of the memory units and fully connected layer, improving recognition accuracy and robustness.
- (4)
- Enhanced recognition of new words and domain-specific terms: Through the integration of BERT, RNNs, attention mechanisms, novel memory units, and fully connected layers, the model’s ability to recognize new words and domain-specific terms has been markedly enhanced.
2. Related Work
2.1. Rule- and Dictionary-Based Approaches
2.2. Statistical Machine Learning-Based Methods
2.3. Deep-Learning-Based Approaches
3. Method
3.1. Medical Text Encoder
3.2. Medical Named Entity Encoder
3.2.1. Character-Level Sequence Processing
3.2.2. Sentence-Level Relationship Processing
3.3. Novel Memory Unit
3.4. Full-Linkage Layer
3.5. Loss Function
4. Experiment and Result
4.1. Experimental Datasets
4.2. Parameterisation and Evaluation Criteria
4.3. Experimental Setup
- (1)
- BERT-base [23]: BERT is a pre-trained language model developed by the Google AI Language team based on Transformer architecture. It utilizes a bidirectional encoder to understand contextual relationships in text. BERT-base consists of 12 Transformer layers, each with 768 hidden units and 12 self-attention heads.
- (2)
- BERT-wwm [24]: BERT-wwm, developed by the Xunfei Joint Laboratory of HIT University, is a variant of BERT that employs the whole-word masking strategy during pre-training. This approach masks entire words rather than just parts of words.
- (3)
- MacBERT-base [25]: MacBERT, an improved version of BERT, incorporates several enhancements including a revised masking strategy and a denoising task to bolster the model’s robustness and generalization. MacBERT-base uses “word-level masking” and “sentence reconstruction” strategies to better learn contextual semantics during pre-training.
- (4)
- RoBERTa [26]: RoBERTa, developed by Facebook AI, is an enhanced version of BERT that optimizes the pre-training process. Improvements include utilizing a larger training dataset, extending training duration, eliminating the next-sentence prediction (NSP) task, and dynamically varying the masking pattern.
- (5)
- BERT-wwm-ext [27]: BERT-wwm-ext, developed by the Xunfei Joint Laboratory of HITC, is an extended version of BERT-wwm. It enhances BERT-wwm by utilizing a larger Chinese corpus for pre-training while continuing to apply the whole-word masking strategy.
5. Discussion
- (1)
- In the first ablation experiment, we removed the fully connected layer, retaining only the BERT-GLMU+RMN model with the CRF layer, to compare the performance of the model without the fully connected layer to that of the full model.
- (2)
- In the second ablation experiment, we retained the BERT-FC-CRF model and removed the novel memory unit (GLMU+RMN) to evaluate the contribution of the novel memory unit to overall model performance.
- (3)
- In the third ablation experiment, we retained the BERT-CRF model and removed both the novel memory unit (GLMU+RMN) and the fully connected layer (FC) to evaluate their contributions to the overall model performance.
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Jehangir, B.; Radhakrishnan, S.; Agarwal, R. A survey on Named Entity Recognition—Datasets, tools, and methodologies. Nat. Lang. Process. J. 2023, 3, 100017. [Google Scholar] [CrossRef]
- Soltau, H.; Shafran, I.; Wang, M.; El Shafey, L. RNN Transducers for Named Entity Recognition with constraints on alignment for understanding medical conversations. In Proceedings of the INTERSPEECH 2022, Incheon, Republic of Korea, 18–22 September 2022; pp. 1901–1905. [Google Scholar]
- Cahuantzi, R.; Chen, X.; Güttel, S. A comparison of LSTM and GRU networks for learning symbolic sequences. In Science and Information Conference; Springer Nature: Cham, Switzerland, 2023; pp. 771–785. [Google Scholar]
- Zhang, R.; Zhao, P.; Guo, W.; Wang, R.; Lu, W. Medical named entity recognition based on dilated convolutional neural network. Cogn. Robot. 2022, 2, 13–20. [Google Scholar] [CrossRef]
- Lu, Y.; Chen, H.; Zhang, Y.; Peng, J.; Xiang, D.; Zhang, J. Research on entity relation extraction for Chinese medical text. Health Inform. J. 2024, 30, 14604582241274762. [Google Scholar] [CrossRef] [PubMed]
- Zhu, Z.; Li, J.; Zhao, Q.; Akhtar, F. A dictionary-guided attention network for biomedical named entity recognition in Chinese electronic medical records. Expert Syst. Appl. 2023, 231, 120709. [Google Scholar] [CrossRef]
- Ke, J.; Wang, W.; Chen, X.; Gou, J.; Gao, Y.; Jin, S. Medical entity recognition and knowledge map relationship analysis of Chinese EMRs based on improved BiLSTM-CRF. Comput. Electr. Eng. 2023, 108, 108709. [Google Scholar] [CrossRef]
- Zhao, X.; Shi, Z.; Xiang, Y.; Ren, Y. Chinese Named Entity Recognition Based on Grid Tagging and Semantic Segmentation. In Proceedings of the 2023 IEEE 9th International Conference on Cloud Computing and Intelligent Systems (CCIS), Dali, China, 12–13 August 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 289–294. [Google Scholar]
- Li, J.; Meng, K. MFE-NER: Multi-feature fusion embedding for Chinese named entity recognition. arXiv 2021, arXiv:2109.07877. [Google Scholar]
- Zheng, D.; Zhang, H.; Yu, F. Named entity recognition of Chinese electronic medical records based on adversarial training and feature fusion. In Proceedings of the 2023 International Joint Conference on Robotics and Artificial Intelligence, Macao, China, 19–25 August 2023; pp. 175–179. [Google Scholar]
- Modi, S.; Kasmiran, K.A.; Sharef, N.M.; Sharum, M.Y. Extracting adverse drug events from clinical Notes: A systematic review of approaches used. J. Biomed. Inform. 2024, 151, 104603. [Google Scholar] [CrossRef]
- Zhang, N.; Chen, M.; Bi, Z.; Liang, X.; Li, L.; Shang, X.; Yin, K.; Tan, C.; Xu, J.; Huang, F.; et al. Cblue: A chinese biomedical language understanding evaluation benchmark. arXiv 2021, arXiv:2106.08087. [Google Scholar]
- Zheng, H.; Guan, M.; Mei, Y.; Li, Y.; Wu, Y. Check for updates ECNU-LLM@ CHIP-PromptCBLUE: Prompt Optimization and In-Context Learning for Chinese Medical Tasks. In Proceedings of the Health Information Processing: Evaluation Track Papers: 9th China Conference, CHIP 2023, Hangzhou, China, 27–29 October 2023; Proceedings. Springer Nature: Berlin/Heidelberg, Germany, 2024; Volume 2080, p. 60. [Google Scholar]
- Yi, F.; Liu, H.; Wang, Y.; Wu, S.; Sun, C.; Feng, P.; Zhang, J. Medical Named Entity Recognition Fusing Part-of-Speech and Stroke Features. Appl. Sci. 2023, 13, 8913. [Google Scholar] [CrossRef]
- An, Y.; Xia, X.; Chen, X.; Wu, F.X.; Wang, J. Chinese clinical named entity recognition via multi-head self-attention based BiLSTM-CRF. Artif. Intell. Med. 2022, 127, 102282. [Google Scholar] [CrossRef]
- Li, D.; Tu, Y.; Zhou, X.; Zhang, Y.; Ma, Z. End-to-end chinese entity recognition based on bert-bilstm-att-crf. ZTE Commun. 2022, 20, 27. [Google Scholar]
- Wang, W.; Li, X.; Ren, H.; Gao, D.; Fang, A. Chinese clinical named entity recognition from electronic medical records based on multisemantic features by using robustly optimized bidirectional encoder representation from transformers pretraining approach whole word masking and convolutional neural networks: Model development and validation. JMIR Med. Inform. 2023, 11, e44597. [Google Scholar] [PubMed]
- Yang, H.; Wang, L.; Yang, Y. Named Entity Recognition in Electronic Medical Records Incorporating Pre-trained and Multi-Head Attention. IAENG Int. J. Comput. Sci. 2024, 51, 401–408. [Google Scholar]
- Zhong, J.; Xuan, Z.; Wang, K.; Cheng, Z. A BERT-Span model for Chinese named entity recognition in rehabilitation medicine. PeerJ Comput. Sci. 2023, 9, e1535. [Google Scholar] [CrossRef]
- Tu, H.; Han, L.; Nenadic, G. Extraction of Medication and Temporal Relation from Clinical Text by Harnessing Different Deep Learning Models. arXiv 2023, arXiv:2310.02229. [Google Scholar]
- Kong, W.; Xia, Y.; Yao, W.; Lu, T. A Joint Entity and Relation Extraction Approach Using Dilated Convolution and Context Fusion. In CCF International Conference on Natural Language Processing and Chinese Computing; Springer Nature: Cham, Switzerland, 2023; pp. 135–146. [Google Scholar]
- Jiang, W.; Ye, Z.; Ou, Z.; Zhao, R.; Zheng, J.; Liu, Y.; Liu, B.; Li, S.; Yang, Y.; Zheng, Y. Mcscset: A specialist-annotated dataset for medical-domain Chinese spelling correction. In Proceedings of the 31st ACM International Conference on Information & Knowledge Management, Atlanta, GA, USA, 17–21 October 2022; pp. 4084–4088. [Google Scholar]
- Koroteev, M.V. BERT: A review of applications in natural language processing and understanding. arXiv 2021, arXiv:2103.11943. [Google Scholar]
- Zhang, Y.; Liao, X.; Chen, L.; Kang, H.; Cai, Y.; Wang, Q. Multi-BERT-wwm model based on probabilistic graph strategy for relation extraction. In Proceedings of the Health Information Science: 10th International Conference, HIS 2021, Melbourne, VIC, Australia, 25–28 October 2021; Proceedings 10. Springer International Publishing: Berlin/Heidelberg, Germany, 2021; pp. 95–103. [Google Scholar]
- Cui, Y.; Che, W.; Liu, T.; Qin, B.; Yang, Z. Pre-training with whole word masking for chinese bert. IEEE/ACM Trans. Audio Speech Lang. Process. 2021, 29, 3504–3514. [Google Scholar] [CrossRef]
- Nemkul, K. Use of Bidirectional Encoder Representations from Transformers (BERT) and Robustly Optimized Bert Pretraining Approach (RoBERTa) for Nepali News Classification. Tribhuvan Univ. J. 2024, 39, 124–137. [Google Scholar] [CrossRef]
- Wang, S.; Fei, C.; Zhang, M. BEDA: BERT-wwm-ext with Data Augmentation for Similarity Measurement and Difficulty Evaluation of Test Questions. In Proceedings of the 2023 5th International Academic Exchange Conference on Science and Technology Innovation (IAECST), Guangzhou, China, 8–10 December 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 6–10. [Google Scholar]
- Gu, Y.; Qu, X.; Wang, Z.; Zheng, Y.; Huai, B.; Yuan, N.J. Delving deep into regularity: A simple but effective method for Chinese named entity recognition. arXiv 2022, arXiv:2204.05544. [Google Scholar]
- Liu, N.; Hu, Q.; Xu, H.; Xu, X.; Chen, M. Med-BERT: A pretraining framework for medical records named entity recognition. IEEE Trans. Ind. Inform. 2021, 18, 5600–5608. [Google Scholar] [CrossRef]
- Han, D.; Wang, Z.; Li, Y.; Zhang, J. Segmentation-aware relational graph convolutional network with multi-layer CRF for nested named entity recognition. Complex Intell. Syst. 2024, 1–13. [Google Scholar] [CrossRef]
- Zhou, L.; Chen, Y.; Li, X.; Li, Y.; Li, N.; Wang, X.; Zhang, R. A New Adapter Tuning of Large Language Model for Chinese Medical Named Entity Recognition. Appl. Artif. Intell. 2024, 38, 2385268. [Google Scholar] [CrossRef]
Sample Size | Training Sets | Validators | Test Sets | Medical Entities |
---|---|---|---|---|
196,496 | 157,194 | 19,652 | 19,650 | 81,020 |
Sample Size | Training Sets | Validators | Test Sets |
---|---|---|---|
23,000 | 15,000 | 5000 | 3000 |
Hyperparameterization | Setpoint |
---|---|
Epochs | 100 |
Batch size | 32 |
Learning rate | 5 × 10−5 |
Learning rate decay | 0.01 |
Text length | 512 |
BiLSTM dimensions | 512 |
Data Set | Model | Precision (%) | Recall (%) | F1 (%) |
---|---|---|---|---|
MCSCSet | BERT-base-crf | 90.22 | 87.41 | 88.79 |
BERT-wwm-crf | 90.82 | 88.20 | 89.49 | |
MacBERT-base-crf | 89.13 | 87.98 | 88.55 | |
RoBERTa-crf | 92.10 | 90.52 | 91.30 | |
BERT-GLMU+RMN-FC-CRF (Ours) | 92.28 | 90.61 | 91.53 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Kang, Y.; Yan, Y.; Huang, W. Chinese Medical Named Entity Recognition Based on Context-Dependent Perception and Novel Memory Units. Appl. Sci. 2024, 14, 8471. https://doi.org/10.3390/app14188471
Kang Y, Yan Y, Huang W. Chinese Medical Named Entity Recognition Based on Context-Dependent Perception and Novel Memory Units. Applied Sciences. 2024; 14(18):8471. https://doi.org/10.3390/app14188471
Chicago/Turabian StyleKang, Yufeng, Yang Yan, and Wenbo Huang. 2024. "Chinese Medical Named Entity Recognition Based on Context-Dependent Perception and Novel Memory Units" Applied Sciences 14, no. 18: 8471. https://doi.org/10.3390/app14188471
APA StyleKang, Y., Yan, Y., & Huang, W. (2024). Chinese Medical Named Entity Recognition Based on Context-Dependent Perception and Novel Memory Units. Applied Sciences, 14(18), 8471. https://doi.org/10.3390/app14188471