Computer Science ›› 2019, Vol. 46 ›› Issue (8): 244-248.doi: 10.11896/j.issn.1002-137X.2019.08.040

• Artificial Intelligence • Previous Articles     Next Articles

Event Temporal Relation Classification Method Based on Self-attention Mechanism

ZHANG Yi-jie, LI Pei-feng, ZHU Qiao-ming   

  1. (School of Computer Science and Technology,Soochow University,Suzhou,Jiangsu 215006,China)
    (Province Key Lab of Computer Information Processing Technology of Jiangsu,Suzhou,Jiangsu 215006,China)
  • Received:2018-07-09 Online:2019-08-15 Published:2019-08-15

Abstract: Classifying temporal relation between events is a significant subsequent study of event extraction.With the development of deep learning,neural network plays a vital role in the task of event temporal relation classification.However,it remains a major challenge for conventional RNNs or CNNs to handle structural information and capture long distance dependence relations.To address this issue,this paper proposed a neural architecture for event temporal relation classification based on self-attention mechanism,which can directly capture relationships between two arbitrary tokens.The classification performance is improved significantly through combing this mechanism with nonlinear layers.The contrast experiments on TimeBank-Dense and Richer Event Description datasets prove that the proposed method outperforms most of the existing neural methods.

Key words: Temporal relation, Deep learning, Self-attention mechanism

CLC Number: 

  • TP391.1
[1] LIN J,YUAN C F.Extraction and Computation of Chinese Temporal Relation[J].Journal of Chinese Information Processing,2009,23(5):62-67.(in Chinese) 林静,苑春法.汉语时间关系抽取与计算[J].中文信息学报,2009,23(5):62-67.
[2] ZHONG Z M,LIU Z T,ZHOU W,et al.The Model of Event Relation Representation[J].Journal of Chinese Information Processing,2009,23(6):56-60.(in Chinese) 仲兆满,刘宗田,周文,等.事件关系表示模型[J].中文信息学报,2009,23(6):56-60.
[3] WANG F E,TAN H Y,QIAN Y L.Recognition of Temporal Relation in One Sentence Based on Maximum Entropy[J].Computer Engineering,2012,38(4):37-39.(in Chinese) 王风娥,谭红叶,钱揖丽.基于最大熵的句内时间关系识别[J].计算机工程,2012,38(4):37-39.
[4] MARCU D,ECHIHABI A.Anunsupervised approach to recognizing discourse relations[C]∥Proceedings of the Association for Computational Linguistics.Association for Computational Linguistics,2002:368-375.
[5] MANI I,VERHAGEN M,WELLNER B,et al.Machine lear- ning of temporal relations[C]∥Proceedings of the Association for Computational Linguistics.Association for Computational Linguistics,2006:753-760.
[6] CHAMBERS N,WANG S,JURAFSKY D.Classifying temporal relations between events[C]∥Proceeding of the ACL on Inte-ractive Poster and Demonstration Sessions.Association for Computational Linguistics,2007:173-176.
[7] LI P F,ZHU Q M,ZHOU G D,et al.Global Inference to Chinese Temporal Relation Extraction[C]∥Proceedings of the International Conference on Computational Linguistics.2016:1451-1460.
[8] CHENG F,MIYAO Y.Classifying Temporal Relations by Bidirectional LSTM over Dependency Paths[C]∥Proceedings of the Association for Computational Linguistics(Short Papers).Association for Computational Linguistics.2017:1-6.
[9] MENG Y,RUMSHISKY A,ROMANOV A.Temporal Infor- mation Extraction for Question Answering Using Syntactic Dependencies in an LSTM-based Architecture[C]∥Proceedings of the Conference on Empirical Methods in Natural Language Processing.Association for Computational Linguistics,2017:887-896.
[10] CHOUBEY P K,HUANG R H.A Sequential Model for Classifying Temporal Relations between Intra-Sentence Events[C]∥Proceedings of the Conference on Empirical Methods in Natural Language Processing.Association for Computational Linguistics.2017:1796-1802.
[11] TOURILLE J,FERRET O,TANNIER X,et al.Neural Architecture for Temporal Relation Extraction:A Bi-LSTM Approach for Detecting Narrative Containers[C]∥Proceedings of the Association for Computational Linguistics.Association for Computational Linguistics,2017:224-230.
[12] VASWANI A,SHAZEER N,PARMAR N,et al.Attentionis all you need[J].arXiv:1706.03762.
[13] CHENG J P,DONG L,LAPATA M.Long Short-Term Memory-Networks for Machine Reading[J].arXiv:1601.06733.
[14] LIN Z H,FENG M W,SANTOS C N,et al.A Structured Self-attentive Sentence Embedding[J].arXiv:1703.03130.
[15] PAULUS R,XIONG C M,SOCHER R.A Deep Reinforced Model for Abstractive Summarization[J].arXiv:1705.04304.
[16] SHEN T,ZHOU T Y,LONG G D,et al.DiSAN:Directional Self-Attention Network for RNN/CNN-free Language Understanding[J].arXiv :1709.04696.
[17] DEY R,SALEMT F M.Gate-variants of Gated Recurrent Unit (GRU)neural networks[J].arXiv:1701.05923.
[18] DAUPHIN Y N,FAN A,AULI M,et al.Language Modeling with Gated Convolutional Networks[J].arXiv:1612.08083.
[19] MIRZA P,TONELLI S.On the contribution of word embeddings to temporal relationclassification[C]∥Proceedings of the International Conference on Computational Linguistics.2016:2818-2828.
[1] WANG Rui-ping, JIA Zhen, LIU Chang, CHEN Ze-wei, LI Tian-rui. Deep Interest Factorization Machine Network Based on DeepFM [J]. Computer Science, 2021, 48(1): 226-232.
[2] YU Wen-jia, DING Shi-fei. Conditional Generative Adversarial Network Based on Self-attention Mechanism [J]. Computer Science, 2021, 48(1): 241-246.
[3] TONG Xin, WANG Bin-jun, WANG Run-zheng, PAN Xiao-qin. Survey on Adversarial Sample of Deep Learning Towards Natural Language Processing [J]. Computer Science, 2021, 48(1): 258-267.
[4] DING Yu, WEI Hao, PAN Zhi-song, LIU Xin. Survey of Network Representation Learning [J]. Computer Science, 2020, 47(9): 52-59.
[5] HE Xin, XU Juan, JIN Ying-ying. Action-related Network:Towards Modeling Complete Changeable Action [J]. Computer Science, 2020, 47(9): 123-128.
[6] YE Ya-nan, CHI Jing, YU Zhi-ping, ZHAN Yu-liand ZHANG Cai-ming. Expression Animation Synthesis Based on Improved CycleGan Model and Region Segmentation [J]. Computer Science, 2020, 47(9): 142-149.
[7] DENG Liang, XU Geng-lin, LI Meng-jie, CHEN Zhang-jin. Fast Face Recognition Based on Deep Learning and Multiple Hash Similarity Weighting [J]. Computer Science, 2020, 47(9): 163-168.
[8] BAO Yu-xuan, LU Tian-liang, DU Yan-hui. Overview of Deepfake Video Detection Technology [J]. Computer Science, 2020, 47(9): 283-292.
[9] YUAN Ye, HE Xiao-ge, ZHU Ding-kun, WANG Fu-lee, XIE Hao-ran, WANG Jun, WEI Ming-qiang, GUO Yan-wen. Survey of Visual Image Saliency Detection [J]. Computer Science, 2020, 47(7): 84-91.
[10] WANG Wen-dao, WANG Run-ze, WEI Xin-lei, QI Yun-liang, MA Yi-de. Automatic Recognition of ECG Based on Stacked Bidirectional LSTM [J]. Computer Science, 2020, 47(7): 118-124.
[11] LIU Yan, WEN Jing. Complex Scene Text Detection Based on Attention Mechanism [J]. Computer Science, 2020, 47(7): 135-140.
[12] ZHANG Zhi-yang, ZHANG Feng-li, TAN Qi, WANG Rui-jin. Review of Information Cascade Prediction Methods Based on Deep Learning [J]. Computer Science, 2020, 47(7): 141-153.
[13] JIANG Wen-bin, FU Zhi, PENG Jing, ZHU Jian. 4Bit-based Gradient Compression Method for Distributed Deep Learning System [J]. Computer Science, 2020, 47(7): 220-226.
[14] CHEN Jin-yin, ZHANG Dun-Jie, LIN Xiang, XU Xiao-dong and ZHU Zi-ling. False Message Propagation Suppression Based on Influence Maximization [J]. Computer Science, 2020, 47(6A): 17-23.
[15] CHENG Zhe, BAI Qian, ZHANG Hao, WANG Shi-pu and LIANG Yu. Improving Hi-C Data Resolution with Deep Convolutional Neural Networks [J]. Computer Science, 2020, 47(6A): 70-74.
Full text



[1] LEI Li-hui and WANG Jing. Parallelization of LTL Model Checking Based on Possibility Measure[J]. Computer Science, 2018, 45(4): 71 -75 .
[2] SUN Qi, JIN Yan, HE Kun and XU Ling-xuan. Hybrid Evolutionary Algorithm for Solving Mixed Capacitated General Routing Problem[J]. Computer Science, 2018, 45(4): 76 -82 .
[3] ZHANG Jia-nan and XIAO Ming-yu. Approximation Algorithm for Weighted Mixed Domination Problem[J]. Computer Science, 2018, 45(4): 83 -88 .
[4] WU Jian-hui, HUANG Zhong-xiang, LI Wu, WU Jian-hui, PENG Xin and ZHANG Sheng. Robustness Optimization of Sequence Decision in Urban Road Construction[J]. Computer Science, 2018, 45(4): 89 -93 .
[5] SHI Wen-jun, WU Ji-gang and LUO Yu-chun. Fast and Efficient Scheduling Algorithms for Mobile Cloud Offloading[J]. Computer Science, 2018, 45(4): 94 -99 .
[6] ZHOU Yan-ping and YE Qiao-lin. L1-norm Distance Based Least Squares Twin Support Vector Machine[J]. Computer Science, 2018, 45(4): 100 -105 .
[7] LIU Bo-yi, TANG Xiang-yan and CHENG Jie-ren. Recognition Method for Corn Borer Based on Templates Matching in Muliple Growth Periods[J]. Computer Science, 2018, 45(4): 106 -111 .
[8] GENG Hai-jun, SHI Xin-gang, WANG Zhi-liang, YIN Xia and YIN Shao-ping. Energy-efficient Intra-domain Routing Algorithm Based on Directed Acyclic Graph[J]. Computer Science, 2018, 45(4): 112 -116 .
[9] CUI Qiong, LI Jian-hua, WANG Hong and NAN Ming-li. Resilience Analysis Model of Networked Command Information System Based on Node Repairability[J]. Computer Science, 2018, 45(4): 117 -121 .
[10] WANG Zhen-chao, HOU Huan-huan and LIAN Rui. Path Optimization Scheme for Restraining Degree of Disorder in CMT[J]. Computer Science, 2018, 45(4): 122 -125 .