Computer Science ›› 2019, Vol. 46 ›› Issue (12): 242-249.doi: 10.11896/jsjkx.181102117

• Artificial Intelligence • Previous Articles     Next Articles

Entity and Relationship Joint Extraction Method of Feedback Mechanism

MA Jian-hong, LI Zhen-zhen, ZHU Huai-zhong, WEI Zi-mo   

  1. (School of Artificial Intelligence,Hebei University of Technology,Tianjin 300401,China)
  • Received:2018-11-16 Online:2019-12-15 Published:2019-12-17

Abstract: Entity and relationship extraction are two core tasks in information extraction,and are the important cornerstone of knowledge mapping.At present,entity recognition and relationship extraction usually adopt the method of extracting features and rules manually and realizing them independently in two steps.This method is easy to cause duplicate data preprocessing and error propagation.The two modules are interrelated.Entity recognition is the basis of relationship extraction.The results of entity relationship extraction can also feedback and verify entity information.Therefore,a hybrid neural network model (Mufeedback-Join Model) without adding manual features and with mutual feedback mechanism was proposed to extract entities and their relationships jointly and realize the feedback checking mechanism of entity relationship to entity recognition.The model shares Bi-LSTM feature extraction layer to extract text context features,and introduces attention mechanism to capture key parts based on shared layer features.After decoding the feature,CRF is used to complete the entity sequence labeling task.The shared layer feature and entity feature are input into CNN model to realize entity relationship extraction.Finally,the relationship extraction result loss value is calculated,and the feature extraction layer of loss value feedback correction and the parameters of entity recognition model are combined.In the same hardware environment,this method can shorten the training time of model,improve the accuracy,recall and F1 value of entity and relationship extraction,The F1 value of the joint extraction is improved by 2.91%,the entity identification sub-module F1 is increased by 1.34% on average,and the relationship extraction F1 value is increased by 5.79%.The experimental data show that the joint extraction model can merge two sub-modules to reduce data processing time and error data transmission,and the mechanism of mutual feedback can improve the overall recognition effect.

Key words: Deep learning, Entity recognition, Feedback mechanism, Joint extraction, Relation extraction

CLC Number: 

  • TP391
[1]SUNDHIM B M.Named Entity task definition,version 2.1 [C]//Proc.of the Sixth Message Understanding Conf.America:Morgan Kaufmann Publishers,1995:319 -332.
[2]ZHANG L,ZHAO H.Named entity recognition for Chinese microblog with convolutional neural network[C]//International Conference on Natural Computation,Fuzzy Systems and Knowledge Discovery.China:IEEE Press,2017:87-92.
[3]CHEN Y,ZHENG D Q,ZHAO T J.Chinese relation extraction based on Deep Belief Nets[J].Journal of Software,2012,23(10):2572-2585.(in Chinese)
陈宇,郑德权,赵铁军.基于Deep Belief Nets的中文名实体关系抽取[J].软件学报,2012,23(10):2572-2585.
[4]MIWA M,SASAKI Y.Modeling Joint Entity and Relation Extraction with Table Representation[C]//Conference on Empirical Methods in Natural Language Processing.Qatar:Association for Computational Linguistics ,2014:944-948.
[5]MIWA M,BANSAL M.End-to-End Relation Extraction using LSTMs on Sequences and Tree Structures[C]//Meeting of the Association for Computational Linguistics.Germany:dblp:computer science bibliography,2016:1105-1116.
[6]ZHENG S,HAO Y,LU D,et al.Joint Entity and Relation Extraction Based on A Hybrid Neural Network[J].Neurocomputing,2017,257:1-8.
[7]BAHDANAU D,CHO K,BENGIO Y.Neural Machine Translation by Jointly Learning to Align and Translate[J/OL].Computer Science,2014.https://arxiv.org/pdf/1409.0473v6.pdf.
[8]KAMBHATLA N.Combining lexical,syntactic,and semantic features with maximum entropy models for extracting relations[C]//ACL Interactive Poster Demonstration Sessions,Spain:Association for Computational Linguistics.2004:22-25.
[9]OUDAH M,SHAALAN K.NERA 2.0:Improving coverage and performance of rule-based named entity recognition for Arabic[J].Natural Language Engineering,2017,23:441-472.
[10]ZHOU G D,SU J.Named entity recognition using an HMM- based chunk tagger[C]//Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics.America,2002:473-480.
[11]MCCALLUM A,LI W.Early results for named entity recognition with conditional random fields,feature induction and web-enhanced lexicons[C]//Conference on Natural Language Learning at Hlt-Naacl.Canada:Association for Computational Linguistics,2003:188-191.
[12]ZHANG H N,WU D Y,LIU Y,et al.Chinese Named Entity Recognition Based on Deep Neural Network[J].Journal of Chinese Information Processing,2017,31(4):28-35.(in Chinese)
张海楠,伍大勇,刘悦,等.基于深度神经网络的中文命名实体识别[J].中文信息学报,2017,31(4):28-35.
[13]DONG C H,ZHANG J J,ZONG C Q.Character-Based LSTM-CRF with Radical-Level Features for Chinese Named Entity Recognition[C]//International Conference on Computer Processing of Oriental Languages.Lecture Notes in Computer Science,Cham:Springer,2016:239-250.
[14]LUO L,YANG Z,YANG P,et al.An attention-based BiLSTM-CRF approach to document-level chemical named entity recognition[J].Bioinformatics,2017,34(8):1381-1388.
[15]CHE W X,LIU T,LI S.Automatic Entity Relation Extraction[J].Journal of Chinese Information Processing,2005,19(2):1-6.(in Chinese)
车万翔,刘挺,李生.实体关系自动抽取[J].中文信息学报,2005,19(2):1-6.
[16]MA X J,GUO J Y,XIAN Y T,et al.Entity Hyponymy Acquisition and Organization Combining Word Embedding and Bootstrapping in Special Domain[J].Computer Science,2018,45(1):67-72.(in Chinese)
马晓军,郭剑毅,线岩团,等.结合词向量和Bootstrapping的领域实体上下位关系获取与组织[J].计算机科学,2018,45(1):67-72.
[17]SOCHER R,PENNINGTON J,HUANG E H,et al.Semi-supervised recursive autoencoders for predicting sentiment distributions[C]//Conference on Empirical Methods in Natural Language Processing(EMNLP 2011).UK:DBLP,2011:151-161.
[18]LAI S,XU L,LIU K,et al.Recurrent convolutional neural networks for text classification[C]//Conference of the Association for the Advancement of Artificial Intelligence (AAAI).America:DBLP,2015:2267-2273.
[19]YAN X,MOU L,LI G,et al.Classifying Relations via Long Short Term Memory Networks along Shortest Dependency Path[J].Computer Science,2015,42(1):56-61.
[20]LI J W,LUONG M T,JURAFSKY D,et al.When Are Tree Structures Necessary for Deep Learning of Representations?[J/OL].https://arxiv.org/abs/1503.00185.
[21]DAN R,YIH W T.1 Global Inference for Entity and Relation Identification via a Linear Programming Formulation[M]//Introduction to Statistical Relational Learning.MIT Press,2007:608-636.
[22]YANG B,CARDIE C.Joint Inference for Fine-grained Opinion Extraction[C]//Meeting of the Association for Computational Linguistics.Bulgaria:ACL,2013:1640-1649.
[23]SINGH S,SINGH M,CHANANA S,et al.Operation and control of a hybrid wind-diesel-battery energy system connected to micro-grid[C]//International Conference on Control,Automation,Robotics and Embedded Systems.India:IEEE Press,2013:1-6.
[24]LI Q,JI H.Incremental Joint Extraction of Entity Mentions and Relations[C]//Meeting of the Association for Computational Linguistics.America:BibSonomy,2014:402-412.
[25]DZMITRY B,KYUNGHYUN C,YOSHUA B,et al.Neural machine translation by jointly learning to align and translate [J/OL].https://arxiv.org/abs/1409.0473v2.
[1] XU Yong-xin, ZHAO Jun-feng, WANG Ya-sha, XIE Bing, YANG Kai. Temporal Knowledge Graph Representation Learning [J]. Computer Science, 2022, 49(9): 162-171.
[2] RAO Zhi-shuang, JIA Zhen, ZHANG Fan, LI Tian-rui. Key-Value Relational Memory Networks for Question Answering over Knowledge Graph [J]. Computer Science, 2022, 49(9): 202-207.
[3] TANG Ling-tao, WANG Di, ZHANG Lu-fei, LIU Sheng-yun. Federated Learning Scheme Based on Secure Multi-party Computation and Differential Privacy [J]. Computer Science, 2022, 49(9): 297-305.
[4] WANG Jian, PENG Yu-qi, ZHAO Yu-fei, YANG Jian. Survey of Social Network Public Opinion Information Extraction Based on Deep Learning [J]. Computer Science, 2022, 49(8): 279-293.
[5] HAO Zhi-rong, CHEN Long, HUANG Jia-cheng. Class Discriminative Universal Adversarial Attack for Text Classification [J]. Computer Science, 2022, 49(8): 323-329.
[6] JIANG Meng-han, LI Shao-mei, ZHENG Hong-hao, ZHANG Jian-peng. Rumor Detection Model Based on Improved Position Embedding [J]. Computer Science, 2022, 49(8): 330-335.
[7] SUN Qi, JI Gen-lin, ZHANG Jie. Non-local Attention Based Generative Adversarial Network for Video Abnormal Event Detection [J]. Computer Science, 2022, 49(8): 172-177.
[8] HU Yan-yu, ZHAO Long, DONG Xiang-jun. Two-stage Deep Feature Selection Extraction Algorithm for Cancer Classification [J]. Computer Science, 2022, 49(7): 73-78.
[9] CHENG Cheng, JIANG Ai-lian. Real-time Semantic Segmentation Method Based on Multi-path Feature Extraction [J]. Computer Science, 2022, 49(7): 120-126.
[10] HOU Yu-tao, ABULIZI Abudukelimu, ABUDUKELIMU Halidanmu. Advances in Chinese Pre-training Models [J]. Computer Science, 2022, 49(7): 148-163.
[11] ZHOU Hui, SHI Hao-chen, TU Yao-feng, HUANG Sheng-jun. Robust Deep Neural Network Learning Based on Active Sampling [J]. Computer Science, 2022, 49(7): 164-169.
[12] JIN Fang-yan, WANG Xiu-li. Implicit Causality Extraction of Financial Events Integrating RACNN and BiLSTM [J]. Computer Science, 2022, 49(7): 179-186.
[13] SU Dan-ning, CAO Gui-tao, WANG Yan-nan, WANG Hong, REN He. Survey of Deep Learning for Radar Emitter Identification Based on Small Sample [J]. Computer Science, 2022, 49(7): 226-235.
[14] WANG Jun-feng, LIU Fan, YANG Sai, LYU Tan-yue, CHEN Zhi-yu, XU Feng. Dam Crack Detection Based on Multi-source Transfer Learning [J]. Computer Science, 2022, 49(6A): 319-324.
[15] CHU Yu-chun, GONG Hang, Wang Xue-fang, LIU Pei-shun. Study on Knowledge Distillation of Target Detection Algorithm Based on YOLOv4 [J]. Computer Science, 2022, 49(6A): 337-344.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!