计算机科学 ›› 2023, Vol. 50 ›› Issue (6A): 220800189-6.doi: 10.11896/jsjkx.220800189
段建勇1,2, 杨潇1,2, 王昊1,2,3, 何丽1,2, 李欣1,2
DUAN Jianyong1,2, YANG Xiao1,2, WANG Hao1,2,3, HE Li1,2, LI Xin1,2
摘要: 为解决现有模型对文档的结构信息挖掘不足的问题,提出一种基于句间信息的图注意力卷积网络模型。该模型改进了一种文档级编码器,该编码器使用了一种新的注意力机制——句间注意力机制,使得句子的最终表示更加关注前一个句子和之前文档中的重要信息,更有利于挖掘文档的结构信息。实验结果表明,所提模型在DocRED数据集上的F1评价指标达到56.3%,性能优于基线模型。在融入句间注意力机制时,由于模型需要对每一句话分别进行句间注意力操作,因此训练模型时需要消耗更多的内存和时间。基于句间信息的图注意力卷积网络模型可以有效地对文档中的相关信息进行聚合,并且增强对文档的结构信息的挖掘能力,从而使得模型在文档级关系抽取任务中效果得到提升。
中图分类号:
[1]CHE W X,LIU T,LI S.Automatic extraction of entity relationship[J].Chinese Journal of Information,2005,19(2):2-7. [2]HUANG Q Z,ZHU S Q,FENG Y S,et al.Three sentences are all you need:Local path enhanced document relation extraction[C]//Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing(Volume 2:Short Papers).2021. [3]QUIRK C,POON H.Distant supervision for relation extraction beyond the sentence boundary[C]//Proceedings of EACL.2017. [4]KIPF T N,WELLING M.Semisupervised classification withgraph convolutional networks[C]//Proceedings of ICLR.2017. [5]HUANG K,QI P,WANG G T,et al.Entity and evidence guided document-level relation extraction[C]//Proceedings of the 6th Workshop on Representation Learning for NLP.2021:140-149. [6]GUPTA P,RAJARAM S,SCHÜTZE H,et al.Neural relation extraction within and across sentence boundaries[C]//Procee-dings of the AAAI Conference on Artificial Intelligence.2019:6513-6520. [7]NAN G,GUO Z,SEKULIC I,et al.Reasoning with LatentStructure Refinement for Document-Level Relation Extraction[C]//ACL.2020:1073-1087. [8]ZHOU W X,HUANG K,MA T Y,et al.Document-level relation extraction with adaptive thresholding and localized context pooling[C]//Proceedings of the AAAI Conference on Artificial Intelligence.2021:1631-1640. [9]YAO Y,YE D M,LI P,et al.DocRED:A large-scale document-level relation extraction dataset[C]//Proceedings of ACL.2019:189-198. [10]TODERICI G,O’MALLEY S M,HWANG S J,et al.Recurrent Neural Network Regularization[J].arXiv:1409.2329,2015. [11]ZHOU P,SHI W,TIAN J,et al.Attention-based bidirectional long short-term memory networks for relation classification[C]//Proceedings of ACL.2016:1412-1421. [12]DEVLIN J,WEI M,CHANG K,et al.BERT:Pre-training of Deep Bidirectional Transformers for Language Understanding[J].arXiv:1810.04805v2,2019. [13]VASWANI A,SHAZEER N,PARMAR N.Attention Is All You Need[J].arXiv:1706.03762v5,2017. [14]GU J H,SUN F Q,QIAN L H,et al.Database:The Journal of Biological Databases and Curation[J].The Journal of Biological Databases and Curation,2017:2049-2058. [15]CHUNG J Y,GULCEHRE C,CHO K H.Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling[J].arXiv:1412.3555,2014. [16]GUO Z J,ZHANG Y,LU W.Attention Guided Graph Convolutional Networks for Relation Extraction[C]//Proceedings of ACL.2019:241-251. [17]CHRISTOPOULOU F,MIWA M,ANANIADOU S.Connec-ting the dots:Document-level neural relation extraction with edge-oriented graphs[C]//Proceedings of EMNLP.2019:1532-1543. [18]WU T,KONG F.Document level relation extraction based on graph attention convolution neural network[J].Chinese Journal of information,2021,35(10):73-80. |
|