Computer Science ›› 2023, Vol. 50 ›› Issue (2): 275-284.doi: 10.11896/jsjkx.220400271

• Artificial Intelligence • Previous Articles     Next Articles

Event Extraction Method Based on Conversational Machine Reading Comprehension Model

LIU Luping1, ZHOU Xin1,2, CHEN Junjun2, He Xiaohai1, QING Linbo1, WANG Meiling1   

  1. 1 College of Electronics and Information Engineering,Sichuan University,Chengdu 610065,China
    2 China Information Technology Security Evaluation Center,Beijing,100085,China
  • Received:2022-04-27 Revised:2022-10-26 Online:2023-02-15 Published:2023-02-22
  • Supported by:
    National Natural Science Foundation of China(60903098) and Chengdu Major Technology Application Demonstration Project(2019-YF09-00120-SN)

Abstract: Event extraction aims to extract structured information automatically from massive unstructured texts to help people quickly understand the latest developments of events.Traditional methods are mainly implemented by classification or sequence labeling methods,which rely on a large amount of labeled data to train the model.In recent years,researchers have proposed to use machinereading comprehension models for event extraction,and through task conversion and combined use of machine rea-ding comprehension datasets for training to effectively alleviate the issue of insufficient annotation data.However,existing methods are limited to a single round of question answering and lack dependencies between different question and answer rounds.In addition,existing methods do not fully utilize entity knowledge in sentences.To this end,a new machine reading comprehension model for event extraction is proposed,and we extend existing methods in two ways.Firstly,by explicitly adding entity tag information in the sentence,making the model effectively learn the prior knowledge of the entities in the input sentence.Secondly,a historical conversational information encoding module is designed,and the attention mechanism is utilized to select important information from historical conversations to assist in inference.Finally,experiment results on a public dataset show that the new model achieves better performance than the existing methods based on the machine reading comprehension model.

Key words: Event extraction, Conversational machine reading comprehension, Entity information marking, Historical conversational information encoding, Attention mechanism

CLC Number: 

  • TP391
[1]HE R F,DUAN S Y.Joint Chinese event extraction basedmulti-task learning[J].Journal of Software,2019,30(4):1015-1030.
[2]LI W,WU M,LU Q,et al.Extractive summarization using inter-and intra-event relevance[C]//Proceeding of the 21st International Conference on Computational Linguistics and 44th Annual Meeting of the Association for Computational Linguistics.Stroudsburg:ACL,2006:369-376.
[3]HAO Y Z,ZHENG Q H,CHEN Y P,et al.Recognition of ab-normal behavior based on data of public opinion on the web[J].Journal of Computer Research and Development,2016,53(3):611-620.
[4]YANG H,CHEN Y,LIU K,et al.DCFEE:A document-levelChinese financial event extraction system based on automatically labeled training data[C]// Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics.Stroudsburg:ACL,2018:50-55.
[5]XUE C,GAO N,ZHA D R,et al.Event database construction techniques[J].Journal of Cyber Security,2019,4(2):83-106.
[6]BAI L,ZHOU Z Y,LI B Y,et al.The construction of the even-tic graph for the political field[J].Journal of Chinese Information Processing,2021,35(4):66-74,82.
[7]GAROFOLO J.Automatic content extraction(ACE) [EB/OL].Philadelphia,PA:University of Pennsylvanial.(2005-07-01) [2021-10-03].http://itl.gov/iad/mig/-tests/ace/2005.
[8]LIU K,CHEN Y,LIU J,et al.Extracting event and their relations from texts:a survey on recent research progress and challenges[J].AI Open,2020,1:22-39.
[9]LLORENS H,SAQUETE E,NAVARRO B.TimeML eventsrecognition and classification:learning CRF models with semantic roles[C]//Proceedings of the 23rd International Conference on Computational Linguistics.Stroudsburg,PA:ACL,2010:725-733.
[10]AHN D.Thestages of event extraction[C]//Proceedings of the Workshop on Annotating and Reasoning about Time and Events.Stroudsburg,PA:ACL,2006:1-8.
[11]HONG Y,ZHANG J,MA B,et al.Using cross-entity inference to improve event extraction[C]//Proceedings of the 49th An-nual Meeting of the Association for Computational Linguistics.Stroudsburg:ACL,2011:1127-1136.
[12]LIAO S,GRISHMAN R.Using document level cross-event inference to improve event extraction[C]//Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics.Stroudsburg:ACL,2010:789-797.
[13]NGUYEN T H,GRISHMAN R.Event detection and domainadaptation with convolutional neural networks[C]//Procee-dings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th Int Joint Conf on Natural Language Processing.Stroudsburg:ACL,2015:365-371.
[14]CHEN Y,XU L,LIU K,et al.Event extraction via dynamic multi-pooling convolutional neural networks[C]//Proceedings of the 53rd Annual Meeting of the Association for Computa-tional Linguistics and the 7th Int Joint Conf on Natural Language Processing.Stroudsburg:ACL,2015:167-176.
[15]FENG X,HUANG L,TANG D,et al.A language-independent neural network for event detection[C]//Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics.Stroudsburg:ACL,2016:66-71.
[16]SHA L,QIAN F,CHANG B,et al.Jointly extracting event triggers and arguments by dependency-bridge rnn and tensor-based argument interaction[C]//Proceedings of the AAAI Conference on Artificial Intelligence.Menlo Park,CA:AAAI,2018:5916-5923.
[17]YANG S,FENG D,QIAO L,et al.Exploring pre-trained lan-guage models for event extraction and generation[C]//Procee-dings of the 57th Annual Meeting of the Association for Computational Linguistics.Stroudsburg:ACL,2019:5284-5294.
[18]DENG S,ZHANG N,LI L,et al.OntoED:Low-resource Event Detection with Ontology Embedding[C]//Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing.Stroudsburg:ACL,2021:2828-2839.
[19]LYU Q,ZHANG H,SULEM E,et al.Zero-shot event extraction via transfer learning:Challenges and insights[C]//Procee-dings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing(Volume 2:Short Papers).2021:322-332.
[20]CONG X,CUI S,YU B,et al.Few-Shot Event Detection with Prototypical Amortized Conditional Random Field[C]//Proceedings of the Findings of the Association for Computational Linguistics.Stroudsburg:ACL,2021:28-40.
[21]DU X,CARDIE C.Event extraction by answering(almost) natural questions[C]//Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing.Stroudsburg:ACL,2020:671-683.
[22]LIU J,CHEN Y,LIU K,et al.Event extraction as machinereading comprehension[C]//Proceedings of the 2020 Confe-rence on Empirical Methods in Natural Language Processing.Stroudsburg:ACL,2020:1641-1651.
[23]LI F,PENG W,CHEN Y,et al.Event extraction as multi-turn question answering[C]//Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing:Findings.Stroudsburg:ACL,2020:829-838.
[24]LI Q,PENG H,LI J,et al.Reinforcement learning-based dialogue guided event extraction to exploit argument relations[J].IEEE/ACM Transactions on Audio,Speech,and Language Processing,2021,30:520-533.
[25]DEVLIN J,CHANG M W,LEE K,et al.BERT:pre-training of deep bidirectional transformers for language understanding[C]//Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics.Stroudsburg,PA:ACL,2019:4171-4186.
[26]GU Y J,GUI X L,LI D F,et al.Survey of machine reading comprehension based on neural network[J].Journal of Software,2020,31(7):2095-2126.
[27]SOARES L B,FITZGERALD N,LING J,et al.Matching the blanks:distributional similarity for relation learning[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics.Stroudsburg:ACL,2019:2895-2905.
[28]PASZKE A,GROSS S,MASSA F,et al.PyTorch:an imperative style,high-performance deep learning library[C]//Proceedings of the 33rd International Conference on Neural Information Processing Systems.Cambridge,MA:MIT Press,2019:8026-8037.
[29]KINGMA D P,BA J.Adam:A method for stochastic optimization [EB/OL].(2014-12-22) [2021-08-15].https://arxiv.org/abs/1412.6980.
[30]RAJPURKAR P,JIA R,LIANG P.Know what you don'tknow:unanswerable questions for squad[C]//Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics.Stroudsburg:ACL,2018:784-789.
[31]LI Q,JI H,HUANG L.Joint event extraction via structured prediction with global features[C]//Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics.Stroudsburg:ACL,2013:73-82.
[32]NGUYEN T H,CHO K,GRISHMAN R.Joint event extraction via recurrent neural networks[C]//Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics.Stroudsburg,PA:ACL,2016:300-309.
[33]LIU X,LUO Z,HUANG H Y.Jointly multiple events extractions via attention-based graphinformation aggregation[C]//Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing.Stroudsburg,ACL,2018:1247-1256.
[34]NGUYEN T M,NGUYEN T H.One for all:neural joint mode-ling of entities and events[C]//Proceedings of the AAAI Confe-rence on Artificial Intelligence.Menlo Park,CA:AAAI,2019:6851-6858.
[35]ZHANG T,JI H,SIL A.Joint entity and event extraction with generative adversarial imitation learning[J].Data Intelligence,2019,1(2):99-120.
[1] LI Shuai, XU Bin, HAN Yike, LIAO Tongxin. SS-GCN:Aspect-based Sentiment Analysis Model with Affective Enhancement and Syntactic Enhancement [J]. Computer Science, 2023, 50(3): 3-11.
[2] CHEN Fuqiang, KOU Jiamin, SU Limin, LI Ke. Multi-information Optimized Entity Alignment Model Based on Graph Neural Network [J]. Computer Science, 2023, 50(3): 34-41.
[3] ZHOU Mingqiang, DAI Kailang, WU Quanwang, ZHU Qingsheng. Attention-aware Multi-channel Graph Convolutional Rating Prediction Model for Heterogeneous Information Networks [J]. Computer Science, 2023, 50(3): 129-138.
[4] ZOU Yunzhu, DU Shengdong, TENG Fei, LI Tianrui. Visual Question Answering Model Based on Multi-modal Deep Feature Fusion [J]. Computer Science, 2023, 50(2): 123-129.
[5] QU Zhong, WANG Caiyun. Crack Detection of Concrete Pavement Based on Attention Mechanism and Lightweight DilatedConvolution [J]. Computer Science, 2023, 50(2): 231-236.
[6] CAI Xiao, CEHN Zhihua, SHENG Bin. SPT:Swin Pyramid Transformer for Object Detection of Remote Sensing [J]. Computer Science, 2023, 50(1): 105-113.
[7] ZHANG Jingyuan, WANG Hongxia, HE Peisong. Multitask Transformer-based Network for Image Splicing Manipulation Detection [J]. Computer Science, 2023, 50(1): 114-122.
[8] LI Xuehui, ZHANG Yongjun, SHI Dianxi, XU Huachi, SHI Yanyan. AFTM:Anchor-free Object Tracking Method with Attention Features [J]. Computer Science, 2023, 50(1): 138-146.
[9] ZHAO Qian, ZHOU Dongming, YANG Hao, WANG Changchen. Image Deblurring Based on Residual Attention and Multi-feature Fusion [J]. Computer Science, 2023, 50(1): 147-155.
[10] ZHENG Cheng, MEI Liang, ZHAO Yiyan, ZHANG Suhang. Text Classification Method Based on Bidirectional Attention and Gated Graph Convolutional Networks [J]. Computer Science, 2023, 50(1): 221-228.
[11] ZHOU Fang-quan, CHENG Wei-qing. Sequence Recommendation Based on Global Enhanced Graph Neural Network [J]. Computer Science, 2022, 49(9): 55-63.
[12] DAI Yu, XU Lin-feng. Cross-image Text Reading Method Based on Text Line Matching [J]. Computer Science, 2022, 49(9): 139-145.
[13] ZHOU Le-yuan, ZHANG Jian-hua, YUAN Tian-tian, CHEN Sheng-yong. Sequence-to-Sequence Chinese Continuous Sign Language Recognition and Translation with Multi- layer Attention Mechanism Fusion [J]. Computer Science, 2022, 49(9): 155-161.
[14] XIONG Li-qin, CAO Lei, LAI Jun, CHEN Xi-liang. Overview of Multi-agent Deep Reinforcement Learning Based on Value Factorization [J]. Computer Science, 2022, 49(9): 172-182.
[15] RAO Zhi-shuang, JIA Zhen, ZHANG Fan, LI Tian-rui. Key-Value Relational Memory Networks for Question Answering over Knowledge Graph [J]. Computer Science, 2022, 49(9): 202-207.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!