Computer Science ›› 2023, Vol. 50 ›› Issue (1): 213-220.doi: 10.11896/jsjkx.211100257

• Artificial Intelligence • Previous Articles     Next Articles

Chinese Nested Named Entity Recognition Algorithm Based on Segmentation Attention andBoundary-aware

ZHANG Rujia, DAI Lu, GUO Peng, WANG Bang   

  1. School of Electronic Information and Communications,Huazhong University of Science and Technology,Wuhan 430074,China
  • Received:2021-11-25 Revised:2022-06-17 Online:2023-01-15 Published:2023-01-09
  • About author:ZHANG Rujia,born in 1997,postgra-duate.Her main research interests include natural language processing and nested name dentity recognition.
    WANG Bang,born in 1975,professor,Ph.D supervisor,is a member of China Computer Federation.His main research interests include recommendation algorithm,knowledge graph and so on.
  • Supported by:
    National Natural Science Foundation of China(62172167).

Abstract: Chinese nested named entity recognition(CNNER) is a challenging task due to the absence of natural delimiters in Chinese and the complexity of the nested structure.In this paper,we propose a novel boundary-aware layered neural model(BLNM) with segmentation attention for the CNNER task.To exploit some semantic relation among adjacent characters,we first design a segmentation attention network to capture the potential word information and enhance character representation.Next,we model the nested structure with dynamically stacked Flat NER networks to detect entities in an inner to outer manner.We also design a boundary generative module to connect adjacent Flat NER layers,which can mark the boundary and position of detected entities and greatly alleviate the error propagation problem.Experiment results on ACE 2005 Chinese nested NE dataset show that the proposed model achieves superior performance than the state-of-the-art methods.

Key words: Chinese nested named entity recognition, Segmentation attention, Boundary generative, Layered neural network

CLC Number: 

  • TP391.1
[1]GUPTA N,SINGH S,ROTH D.Entity linking via joint encoding of types,descriptions,and con-text[C]//Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing.Denmark:ACL,2017:2681-2690.
[2]JI Z,SUN A,CONG G,et al.Joint recognition and linking of fine-grained locations from tweets[C]//Proceedings of the 25th International Conference on World Wide Web.Montréal:WWW,2016:1271-1281.
[3]LIN Y,SHEN S,LIU Z,et al.Neural relation extraction with selective attention over in-stances[C]//Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics.Germany:ACL,2016:2124-2133.
[4]ZHENG S,WANG F,BAO H,et al.Joint extraction of entities and relations based on a novel tagging scheme[C]//Proceedings of the 55th Annual Meeting of the Association for Computa-tional Linguistics.Vancouver:ACL,2017:1227-1236.
[5]CHANG K W,SAMDANI R,ROTH D.A constrained latent variable model for coreference resolution [C]//Proceedings of the 2013 Conference on Empirical Methodsin Natural Language Processing.Seattle:EMNLP,2013:601-612.
[6]SHEN D,ZHANG J,ZHOU G,et al.Effective adaptation of hidden markov model-based named entity recognizer for biome-dical domain[C]//Proceedings of the ACL 2003 Workshop on Natural Language Processing in Biomedicine.Japan:ACL,2003:49-56.
[7]XIA C,ZHANG C,YANG T,et al.Multi-grained named entity recognition[C]//Proceedings of the 57th Annual Meeting of the Association for Compu-tational Linguistics.Italy:ACL,2019:1430-1440.
[8]JU M,MIWA M,ANANIADOU S.A neural layered model for nested named entity recognition[C]// Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics.New Orleans:ACL,2018:1446-1459.
[9]LI H,XU H,QIAN L,et al.Multi-layer Joint Learning of Chinese Nested Named Entity Recognition Based on Self-attention Mechanism[C]//Proceedings of the 9th CCF International Conference on Natural Language Processing and Chinese Computing.Cham:Springer International Publishing,2020:144-155.
[10]KURU O,CAN O A,YURET D.CharNER:Character-levelnamed entity recognition[C]//Proceedings of COLING 2016,the 26th International Conference on Computational Linguistics:Technical Papers.Osaka:COLING,2016:911-921.
[11]TRAN Q,MACKINLAY A,YEPES A J.Named entity recognition with stack residual lstm and trainable bias decoding[J].arXiv:1706.07598,2017.
[12]GRIDACH M.Character-level neural network for biomedicalnamed entity recognition[J].Journal of biomedical informatics,2017,70:85-91.
[13]EBERTS M,ULGES A.Span-based joint entity and relation extraction with transformer pre-training[J].arXiv:1909.07755,2019.
[14]HUANG Z,XU W,YU K.Bidirectional LSTM-CRF models for sequence tagging[J].arXiv:1508.01991,2015.
[15]LIU L,SHANG J,REN X,et al.Empower sequence labeling with task-aware neural language model[C]//Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence.New Orleans:AAAI Press,2018:5253-5260.
[16]LUAN Y,WADDEN D,HE L,et al.A general framework for information extraction using dynamic span graphs[C]//Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics.Minnesota:ACL,2019:3036-3046.
[17]DONG C,ZHANG J,ZONG C,et al.Character-based LSTM-CRF with radical-level features for Chinese named entity recognition[C]//International Conference on Computer Processing of Oriental Languages.Cham:Springer International Publishing,2016:239-250.
[18]ZHANG Y,YANG J.Chinese NER using lattice LSTM[C]//Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics.Melbourne:ACL,2018:1554-1564.
[19]LIU W,XU T,XU Q,et al.An encoding strategy based word-character LSTM for Chinese NER[C]//Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics.Minneapolis:ACL,2019:2379-2389.
[20]WU Y,JIANG M,LEI J,et al.Named entity recognition in Chinese clinical text using deep neural network[J].Studies in health technology and informatics,2015,216:624-628.
[21]GUI T,ZOU Y,ZHANG Q,et al.A Lexicon-Based Graph Neural Network forChinese NER[C]//Proceedings of the 2019 Conference on Empirical Methods in Natural Language Proces-sing and the 9th International Joint Conference on Natural Language Processing.Hong Kong:ACL,2019:1040-1050.
[22]LI X,YAN H,QIU X,et al.FLAT:Chinese NER Using Flat-Lattice Transformer[C]//Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.Online:ACL,2020:6836-6842.
[23]MENGGE X,YU B,LIU T,et al.Porous Lattice Transformer Encoder for Chinese NER[C]//Proceedings of the 28th International Conference on Computational Linguistics.Barcelona:International Committee on Computational Linguistics,2020:3831-3841.
[24]GUI T,MA R,ZHANG Q,et al.CNN-Based Chinese NER with Lexicon Rethinking[C]//Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence Main track.Macao:IJCAI,2019:4982-4988.
[25]SUI D,CHEN Y,LIU K,et al.Leverage lexical knowledge for Chinese named entity recognition via collaborative graph network[C]//Proceedings of the 2019 Conference on Empirical Methods in Natural Language.Hong Kong:EMNLP,2019:3830-3840.
[26]LI J H,CHEN M M,WANG H J,et al.Chinese Named Entity Recognition Method Based on ALBERT-BGRU-CRF[J].Computer Engineering,2022,48(6):89-94,106.
[27]ZHONG S S,CHEN X,ZHAO M H,et al.Incorporating word-set attention into Chinese named entity recognition Method[J].Journal of Jilin University(Engineering and Technology Edition),2022,52(5):1098-1105.
[28]GUO X R,LUO P,WANG W L.Chinese named entity recognition based on Transformer encoder[J].Journal of Jilin University(Engineering and Technology Edition),2021,51(3):989-995.
[29]SI Y C,GUAN Y Q.Chinese Named Entity Recognition Model Based on Transformer Encoder[J].Computer Engineering,2022,48(7):66-72.
[30]HU X B,YU X Q,LI S M,et al.Chinese Named Entity Recognition Based on Knowledge Enhancement[J].Computer Engineering,2021,47(11):84-92.
[31]FU C,ZHAO Y,FU G.Exploiting entity-level morphology to Chinese nested named entity recognition[J].International Journal on Asian Language Processing,2012,22(1):33-48.
[32]ZHOU G,ZHANG J,SU J,et al.Recognizing names in biome-dical texts:a machine learning approach[J].Bioinformatics,2004,20(7):1178-1190.
[33]ZHOU G D.Recognizing names in biomedical texts using mutual information independence model and SVM plus sigmoid[C]//Proceedings of the International Joint Workshop on Natural Language Processing in Bio-medicine and its Applications.Geneva,Switzerland:COLING,2004:1-7.
[34]FU C,FU G.Morpheme-based chinese nested named entity recognition[C]//Proceedings of the 2011 Eighth International Conference on Fuzzy Systems and Knowledge Discovery.Shanghai:FSKD,2011:1221-1225.
[35]LAMPLE G,BALLESTEROS M,SUBRAMANIAN S,et al.Neural architectures for named entity recognition[C]//Procee-dings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics.San Diego:NAACL,2016:260-270.
[36]LAFFERTY J,MCCALLUM A,PEREIRA F C N.Conditional random fields:Probabilistic models for segmenting and labeling sequence data[C]//Proceedings of the 8th International Confe-rence on Machine Learning.Evanston:ICML,1991:282-289.
[37]VITERBI A.Error bounds for convolutional codes and an asymptotically optimum decoding algorithm [C]//Proceedings of IEEE Transactions on Information Theory.IEEE,1967:260-269.
[38]VASWANI A,SHAZEER N,PARMAR N,et al.Attention isall you need[J].arXiv:1706.03762,2017.
[39]DODDINGTON G R,MITCHELL A,PRZYBOCKI M A,et al.The automatic content extraction(ace) program-tasks,data,and evaluation[C]//Proceedings of the Fourth International Confe-rence on Language Resources and Evaluation.Portugal:Euro-pean Language Resources Association,2004:837-840.
[40]PASZKE A,GROSS S,MASSA F,et al.Pytorch:An imperative style,high-performance deep learning library[J].arXiv:1912.01703,2019.
[1] YU Juan, ZHANG Chen. Cross-lingual Term Alignment with Kernel-XGBoost [J]. Computer Science, 2022, 49(11A): 211000111-6.
[2] ZHU Ruo-chen, YANG Chang-chun, ZHANG Deng-hui. EGOS-DST:Efficient Schema-guided Approach to One-step Dialogue State Tracking for Diverse Expressions [J]. Computer Science, 2022, 49(11A): 210900246-7.
[3] SHI Wei, FU Yue. Study on Evolution of Sentiment-Topic of Internet Reviews with Time in Emergencies [J]. Computer Science, 2022, 49(11A): 211000193-6.
[4] GUO Jun-cheng, WAN Gang, HU Xin-jie, WANG Shuai, YAN Fa-bao. Study on Solar Radio Burst Event Detection Based on Transfer Learning [J]. Computer Science, 2022, 49(11A): 210900198-7.
[5] ZHANG Bin, LIU Chang-hong, ZENG Sheng, JIE An-quan. Speech-driven Personal Style Gesture Generation Method Based on Spatio-Temporal GraphConvolutional Networks [J]. Computer Science, 2022, 49(11A): 210900094-5.
[6] WU Zi-yi, LI Shao-mei, JIANG Meng-han, ZHANG Jian-peng. Ontology Alignment Method Based on Self-attention [J]. Computer Science, 2022, 49(9): 215-220.
[7] GUO Yu-xin, CHEN Xiu-hong. Automatic Summarization Model Combining BERT Word Embedding Representation and Topic Information Enhancement [J]. Computer Science, 2022, 49(6): 313-318.
[8] HUANG Shao-bin, SUN Xue-wei, LI Rong-sheng. Relation Classification Method Based on Cross-sentence Contextual Information for Neural Network [J]. Computer Science, 2022, 49(6A): 119-124.
[9] MIU Feng, WANG Ping, LI Tai-yong. Implicit Causality Extraction Method Based on Event Action Direction [J]. Computer Science, 2022, 49(3): 276-280.
[10] XIAO Kang, ZHOU Xia-bing, WANG Zhong-qing, DUAN Xiang-yu, ZHOU Guo-dong, ZHANG Min. Review Question Generation Based on Product Profile [J]. Computer Science, 2022, 49(2): 272-278.
[11] MA Jian-hong, ZHANG Tong. Expert Recommendation Algorithm for Enterprise Engineering Problems [J]. Computer Science, 2022, 49(1): 159-165.
[12] YUAN Jing-ling, DING Yuan-yuan, SHENG De-ming, LI Lin. Image-Text Sentiment Analysis Model Based on Visual Aspect Attention [J]. Computer Science, 2022, 49(1): 219-224.
[13] LIU Kai, ZHANG Hong-jun, CHEN Fei-qiong. Name Entity Recognition for Military Based on Domain Adaptive Embedding [J]. Computer Science, 2022, 49(1): 292-297.
[14] ZOU Ao, HAO Wen-ning, JIN Da-wei, CHEN Gang, TIAN Yuan. Study on Text Retrieval Based on Pre-training and Deep Hash [J]. Computer Science, 2021, 48(11): 300-306.
[15] YU Liang, WEI Yong-feng, LUO Guo-liang, WU Chang-xing. Knowledge Distillation Based Implicit Discourse Relation Recognition [J]. Computer Science, 2021, 48(11): 319-326.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!