Paper
13 October 2022 Modified BERT-based end-to-end Chinese named entity recognition model
Yanchun Tan, Youmin Zhu, Wenqing Shi
Author Affiliations +
Proceedings Volume 12287, International Conference on Cloud Computing, Performance Computing, and Deep Learning (CCPCDL 2022); 1228725 (2022) https://doi.org/10.1117/12.2641066
Event: International Conference on Cloud Computing, Performance Computing, and Deep Learning (CCPCDL 2022), 2022, Wuhan, China
Abstract
In this paper, we present an end-to-end model based on modified Bidirectional Encoder Representations from Transformers (BERT) for Chinese named entity recognition (NER) in natural language processing. The model is composed of the SpanBERT layer and the Conditional Random Field (CRF) layer. By using combination, the model can express the input characters in the better form of "word embeddings", eliminating the steps of feature engineering or data processing in conventional approaches, and can be widely applied to the task of Chinese NER. Our experiments demonstrate that the SpanBERT-CRF model can effectively utilize the contextual data features and give more accurate recognition results. On our data set, the SpanBERT-CRF model had excellent performance with a recognition accuracy of 91.33%, outperforming the benchmark NER model BiLSTM-CRF (Bidirectional Long Short Term Memory, Conditional Random Field) and BERT-CRF model in performance and F1 score.
© (2022) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Yanchun Tan, Youmin Zhu, and Wenqing Shi "Modified BERT-based end-to-end Chinese named entity recognition model", Proc. SPIE 12287, International Conference on Cloud Computing, Performance Computing, and Deep Learning (CCPCDL 2022), 1228725 (13 October 2022); https://doi.org/10.1117/12.2641066
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Data modeling

Performance modeling

Statistical modeling

Computer programming

Transformers

Neural networks

Data processing

Back to Top