Computer Science ›› 2020, Vol. 47 ›› Issue (6A): 40-44.doi: 10.11896/JsJkx.190700042

• Artificial Intelligence • Previous Articles     Next Articles

Relation Extraction Method Combining Encyclopedia Knowledge and Sentence Semantic Features

YU Yi-lin, TIAN Hong-tao, GAO Jian-wei and WAN Huai-yu   

  1. School of Computer and Information Technology,BeiJing Jiaotong University,BeiJing 100044,China
  • Published:2020-07-07
  • About author:LYU Yi-lin, born in 1997, undergra-duate student.His main research interest is information extraction.
    WAN Huai-yu, born in 1981, Ph.D, associate professor, Ph.D supervisor, is a member of China Computer Federation.His main research interests include social network mining, text mining, user behavior analysis, and traffic data mining.
  • Supported by:
    This work was supported by the National Key R&D Program of China (2018YFC0830200).

Abstract: Relation extraction is one of the important research topics in the field of information extraction.Its typical application scenarios include knowledge graphs,question answering systems,machine translation,etc.Recently,deep learning has been applied in a large amount of relation extraction researches,and deep neural networks based relationship extraction method performs much better than the traditional methods in many situations.However,most of the current deep neural network-based relation extraction methods Just rely on the corpus itself and lack the introduction of external knowledge.To address this issue,this paper proposed a neural network model,which combined encyclopedia knowledge and semantic features of sentences for relation extraction.The model introduced the description information of entities in encyclopedia as external knowledge,and dynamically extracted entity features through attention mechanism.Meanwhile,it employed bidirectional LSTM networks to extract the semantic features contained in the sentence.Finally,the model combined the entity features and the sentence semantic features for relation extraction.A series of experiments were carried out based on a manually labeled dataset.Experimental results demonstrate that the proposed model is superior to other existing relationship extraction methods.

Key words: Entity relation extraction, Chinese encyclopedia, Long short-term memory, Attention mechanism

CLC Number: 

  • TP391
[1] ZENG D,LIU K,LAI S,et al.Relation Classification via Convolutional Deep Neural Network //Proceedings of the 25th International Conference on Computational Linguistics.Stroudsburg,PA:ACL,2014:2335-2344.
[2] SANTOS C N,XIANG B,ZHOU B.Classifying Relations by Ranking with Convolutional Neural Networks .arXiv:1504.06580,2015.
[3] NGUYEN T H,GRISHMAN R.Relation extraction:Perspective from convolutional neural networks //Proceedings of the 1st Workshop on Vector Space Modeling for Natural Language Processing.Stroudsburg,PA:ACL,2015:39-48.
[4] ZHANG D,WANG D.Relation Classification via Recurrent Neural Network .arXiv:1508.01006,2015.
[5] ZHANG S,ZHENG D,HU X,et al.Bidirectional long short-term memory networks for relation classification //Procee-dings of the 29th Pacific Asia Conference on Language,Information and Computation.Stroudsburg,PA:ACL,2015:73-78.
[6] ZHOU P,SHI W,TIAN J,et al.Attention-based bidirectional long short-term memory networks for relation classification //Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2:Short Papers).Stroudsburg,PA:ACL,2016:207-212.
[7] BAHDANAU D,CHO K,BENGIO Y.Neural Machine Translation by Jointly Learning to Align and Translate .arXiv:1409.0473,2014.
[8] KAMBHATLA N.Combining lexical,syntactic,and semantic features with maximum entropy models for extracting relations //Proceedings of the ACL Interactive Poster and Demonstration Sessions.Stroudsburg,PA:ACL,2004:22.
[9] SUCHANEK F M,IFRIM G,WEIKUM G.Combining linguistic and statistical analysis to extract relations from web documents //Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining.New York:ACM,2006:712-717.
[10] QIAN L,ZHOU G,KONG F,et al.Exploiting constituent dependencies for tree kernel-based semantic relation extraction //Proceedings of the 22nd International Conference on Computational Linguistics.Stroudsburg,PA:ACL,2008:697-704.
[11] BUNESCU R C,MOONEY R J.A shortest path dependency kernel for relation extraction //Proceedings of the Conference on Human Language Technology and Empirical Methods in Natural Language Processing.Stroudsburg,PA:ACL,2005:724-731.
[12] MOONEY R J,BUNESCU R C.Subsequence kernels for relation extraction //Proceedings of Advances in Neural Information Processing Systems.Stroudsburg,PA:ACL,2006:171-178.
[13] ZENG D,LIU K,CHEN Y,et al.Distant supervision for relation extraction via piecewise convolutional neural networks //Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing.Stroudsburg,PA:ACL,2015:1753-1762.
[14] HOCHREITER S,SCHMIDHUBER J.Long short-term memory .Neural computation,1997,9(8):1735-1780.
[15] CHO K,VAN M B,GULCEHRE C,et al.Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation .arXiv:1406.1078,2014.
[16] XU Y,MOU L,LI G,et al.Classifying relations via long short term memory networks along shortest dependency path .arXiv:1508.03720,2015.
[17] MIKOLOV T,CHEN K,CORRADO G,et al.Efficient Estimation of Word Representations in Vector Space .arXiv:1301.3781,2013.
[18] CHE W,LI Z,LIU T.LTP:A Chinese Language Technology Platform //Proceedings of the 23rd International Conference on Computational Linguistics Demonstrations Volume.Stroudsburg,PA:ACL,2010:13-16.
[19] KINGMA D P,BA J.Adam:A method for stochastic optimization .arXiv:1412.6980,2014.
[1] BAO Zhen-shan, GUO Jun-nan, XIE Yuan and ZHANG Wen-bo. Model for Stock Price Trend Prediction Based on LSTM and GA [J]. Computer Science, 2020, 47(6A): 467-473.
[2] DIAO Li and WANG Ning. Research on Premium Income Forecast Based on X12-LSTM Model [J]. Computer Science, 2020, 47(6A): 512-516.
[3] NI Hai-qing, LIU Dan, SHI Meng-yu. Chinese Short Text Summarization Generation Model Based on Semantic-aware [J]. Computer Science, 2020, 47(6): 74-78.
[4] HUANG Yong-tao, YAN Hua. Scene Graph Generation Model Combining Attention Mechanism and Feature Fusion [J]. Computer Science, 2020, 47(6): 133-137.
[5] ZHANG Zhi-yang, ZHANG Feng-li, CHEN Xue-qin, WANG Rui-jin. Information Cascade Prediction Model Based on Hierarchical Attention [J]. Computer Science, 2020, 47(6): 201-209.
[6] DENG Yi-jiao, ZHANG Feng-li, CHEN Xue-qin, AI Qing, YU Su-zhe. Collaborative Attention Network Model for Cross-modal Retrieval [J]. Computer Science, 2020, 47(4): 54-59.
[7] ZHANG Peng-fei, LI Guan-yu, JIA Cai-yan. Truncated Gaussian Distance-based Self-attention Mechanism for Natural Language Inference [J]. Computer Science, 2020, 47(4): 178-183.
[8] HUANG Hong-wei,LIU Yu-jiao,SHEN Zhuo-kai,ZHANG Shao-wei,CHEN Zhi-min,GAO Yang. End-to-end Track Association Based on Deep Learning Network Model [J]. Computer Science, 2020, 47(3): 200-205.
[9] ZHANG Yi-fei,WANG Zhong-qing,WANG Hong-ling. Product Review Summarization Using Discourse Hierarchical Structure [J]. Computer Science, 2020, 47(2): 195-200.
[10] LIU Yun,YIN Chuan-huan,HU Di,ZHAO Tian,LIANG Yu. Communication Satellite Fault Detection Based on Recurrent Neural Network [J]. Computer Science, 2020, 47(2): 227-232.
[11] LI Yuan,LI Zhi-xing,TENG Lei,WANG Hua-ming,WANG Guo-yin. Comment Sentiment Analysis and Sentiment Words Detection Based on Attention Mechanism [J]. Computer Science, 2020, 47(1): 186-192.
[12] YANG Dan-hao,WU Yue-xin,FAN Chun-xiao. Chinese Short Text Keyphrase Extraction Model Based on Attention [J]. Computer Science, 2020, 47(1): 193-198.
[13] LIU Hai-bo,WU Tian-bo,SHEN Jing,SHI Chang-ting. Advanced Persistent Threat Detection Based on Generative Adversarial Networks and Long Short-term Memory [J]. Computer Science, 2020, 47(1): 281-286.
[14] SUN Zhong-feng, WANG Jing. RCNN-BGRU-HN Network Model for Aspect-based Sentiment Analysis [J]. Computer Science, 2019, 46(9): 223-228.
[15] SHI Chun-dan, QIN Lin. Chinese Named Entity Recognition Method Based on BGRU-CRF [J]. Computer Science, 2019, 46(9): 237-242.
Full text



No Suggested Reading articles found!