计算机科学 ›› 2024, Vol. 51 ›› Issue (1): 252-265.doi: 10.11896/jsjkx.230200100

• 人工智能 • 上一篇    下一篇

限定域关系抽取技术研究综述

侯景1,2, 邓晓梅1, 汉鹏武1   

  1. 1 中国科学院空间应用工程与技术中心 北京100094
    2 中国科学院大学 北京100094
  • 收稿日期:2023-02-15 修回日期:2023-06-29 出版日期:2024-01-15 发布日期:2024-01-12
  • 通讯作者: 汉鹏武(hanpengwu@csu.ac.cn)
  • 作者简介:(houjing21@mails.ucas.ac.cn)

Survey on Domain Limited Relation Extraction

HOU Jing1,2, DENG Xiaomei1, HAN Pengwu1   

  1. 1 Technology and Engineering Center for Space Utilization,Chinese Academy of Sciences,Beijing 100094,China
    2 University of Chinese Academy of Sciences,Beijing 100094,China
  • Received:2023-02-15 Revised:2023-06-29 Online:2024-01-15 Published:2024-01-12
  • About author:HOU Jing,born in 1999,postgraduate.Her main research interests include information extraction and standard digitization.
    HAN Pengwu,born in 1985,Ph.D,se-nior engineer.His main research in-terests include knowledge map,space Internet of things and computer control.

摘要: 限定域关系抽取技术是在预定义实体类型和关系类型的前提下,从文本中捕获关键信息的技术,多采用由头尾实体和关系构成的三元组作为信息表示形式。作为信息抽取领域的重要研究方向之一,其在知识问答、信息检索等任务中被广泛应用。文中在介绍相关概念和任务范式的基础上,分析了深度学习背景下限定域关系抽取任务的研究进展,根据句中实体是否可见,分为关系分类任务和三元组抽取任务,依据任务表现特征,前者可细分为有监督条件下的关系分类任务、小样本关系分类任务和远程监督条件下的关系分类任务。文中探讨和分析了以上任务中常用的技术方法及其优缺点,最后归纳总结了关系抽取技术在低资源、多模态等更为接近真实情景下的发展潜力和现存的挑战。

关键词: 限定域关系抽取, 深度学习, 关系分类, 三元组, 远程监督

Abstract: Domain-limited relation extraction aims to capture essential text information from the text under the premise of predefined entity types and relation types,and mostly uses triples composed of head and tail entities and relations as structured information representation.As one of the important tasks of information extraction,it plays an important role in question answering and information retrieval.Based on its concepts and task paradigms,this paper systematically sorts out the technical methods in domain-limited relation extraction under the background of deep learning.Whether the entity is visible or not,it is divided into relation classification and triplet extraction.According to the performance characteristics of the task,the former can be divided into relation classification under supervised conditions,few-shot relation classification,and relation classification under distant supervision.This paper discusses and analyzes the commonly used technical methods and their advantages and disadvantages in the above tasks.Finally,we summarize the development potential and existing challenges of relation extraction technology in low-resource,multimodal and other situations that are closer to the real world.

Key words: Domain-limited, Deep learning, Relation classification, Triples, Distant supervision

中图分类号: 

  • TP391
[1]NGUYEN T H,GRISHMAN R.Relation Extraction:Perspec-tive from Convolutional Neural Networks[C]//Proceedings of NAACL-HLT 2015.2015:39-48.
[2]FU J,HUANG X,LIU P.SpanNER:Named Entity Re-/Recognition as Span Prediction[C]//Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing(Volume 1:Long Papers).Association for Computational Linguistics,2021:7183-7195.
[3]YI L,HE L,MARI O,et al.Multi-Task Identification of Entities,Relations,and Coreference for Scientific Knowledge Graph Construction[C]//Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing.Brussels,Belgium,Association for Computational Linguistics,2018:3219-3232.
[4]WALKER C,STRASSEL S,MEDERO J,et al.ACE 2005 multilingual training corpus[OL].(2006) [2019-09-13].https://www.cdc.gov/violenceprevention/aces/about.html.
[5]ZHOU B,ZHAO H,PUIG X,et al.Scene parsing throughADE20K dataset[OL].(2017) [2020-05-23].https://groups.csail.mit.edu/vision/datasets/ADE20K.
[6]LI D M,ZHANG Y,LI D Y,et al.A survey on entity relation extraction methods [J].Journal of Computer Research and Development,2020,57(7):1424-1448.
[7]ZHUANG C Z,JIN X L,ZHU W J,et al.A survey on relation extraction based on deep learning [J].Journal of Chinese Information Processing,2019,33(12):1-18.
[8]NAYAK T,MAJUMDER N,GOYAL P,et al.Deep Neural Ap-proaches to Relation Triplets Extraction:a Comprehensive Survey[J].Cognitive Computation,2021,13(5):1215-1232.
[9]ZHANG S W,WANG X,CHEN Z R,et al.A survey on supervised joint entity and relation extraction methods [J].Journal of Frontiers of Computer Science and Technology,2022,16(4):1-24.
[10]BAI L,JIN X L,XI P B,et al.A survey on relation extraction based on distant supervision [J].Journal of Chinese Information Processing,2019,33(10):10-17.
[11]SHANG Y,HUANG H,MAO X.OneRel:Joint Entity and Relation Extraction with One Module in One Step[OL].(2022-03) [2022-5-27].https://arxiv.org/abs/2203.05412v1.
[12]ZHAO K,XU H,YANG J,et al.Consistent RepresentationLearning for Continual Relation Extraction[C]//Findings of the Association for Computational Linguistics.Association for Computational Linguistics,2022:3402-3411.
[13]QIN P,XU W,GUO J.Designing an adaptive attention mechanism for relation classification[C]//2017 International Joint Conference on Neural Networks(IJCNN) IEEE.2017:4356-4362.
[14]SARKAR S,ARZOO K,REBECCA P,et al.CONTAINERFew-Shot Named Entity Recognition via Contrastive learning[C]//Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics(Volume 1:Long Papers).Association for Computational Linguistics,2022:6338-6353.
[15]YU H,ZHANG N,DENG S,et al.Bridging Text and Know-ledge with Multi-Prototype Embedding for Few-Shot Relational Triple Extraction[C]//Proceedings of the 28th International Conference on Computational Linguistics.International Committee on Computational Linguistics,2020:6399-6410.
[16]FENG J,HUANG M,ZHAO L,et al.Reinforcement Learning for Relation Classification from Noisy Data[C]//Association for Advancement Artificial Intelligence.New Orleans,LA,2018:5779-5786.
[17]JIANG X,WANG Q,LI P,et al.Relation Extraction withMulti-instance Multi-label Convolutional Neural Networks[C]//The 26th International Conference on Computational Linguistics(COLING 2016).The COLING 2016 Organizing Committee,2016:1471-1480.
[18]WU F,WELD D.Autonomously semantifying wikipedia[C]//The 16th ACM International Conference on Information and Knowledge Management(CIKM 2007).Association for Computing Machinery,2007:41-50.
[19]MIKE M,STEVEN B,RION S,et al.Distant supervision for relation extraction without labeled data[C]//The Joint Confe-rence of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP(ACL-IJCNLP 2009).Association for Computational Linguistics,2009:1003-1011.
[20]ZHONG Z,CHEN D.A Frustratingly Easy Approach for Entity and Relation Extraction[C]//Proceedings of the 2021 Confe-rence of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies.Association for Computational Linguistics,2021:50-61.
[21]TIAN Y,CHEN G,SONG Y,et al.Dependency-driven Relation Extraction with Attentive Graph Convolutional Networks[C]//The 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing(ACL-IJCNLP 2021).Association for Computational Linguistics,2021:4458-4471.
[22]LAI T,JI H,ZHAI C,et al.Joint Biomedical Entity and Relation Extraction with Knowledge-Enhanced Collective Inference[C]//The 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing(ACL-IJCNLP 2021).Association for Computational Linguistics,2021:6248-6260.
[23]ZHANG Y,CHEN W,LI Y,et al.The TAC Relation Extraction Dataset[OL].(2018-12-15) [2020-02-20].https://catalog.ldc.upenn.edu/LDC2018T24.
[24]GAO T,HAN X,ZHU H,et al.FewRel 2.0:Towards More Challenging Few-Shot Relation Classification[C]//The 2019 Conference on Empirical Methods in Natural Language Proces-sing and the 9th International Joint Conference on Natural Language Processing.Association for Computational Linguistics,2019:6250-6255.
[25]SEBASTIAN R,YAO L,ANDREW M.The New York Times Company[OL].(2010) [2021-01-21].http://iesl.cs.umass.edu/riedel/ecml.
[26]GARDENT C,SHIMORINA A,NARAYAN S,et al.Creating Training Corpora for NLG Micro-Planning[C]//Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics(Volume 1:Long Papers).Association for Computational Linguistics,2017:179-188.
[27]YUAN X.Schema-based-Knowledge-Extraction[OL].(2020-04) [2023-3-13].https://github.com/yuanxiaosc/Schema-based-Knowledge-Extraction.
[28]WANG Y,YU B,ZHANG Y,et al.TPLinker:Single-stage Joint Extraction of Entities and Relations Through Token Pair Lin-king[C]//Proceedings of the 28th International Conference on Computational Linguistics Barcelona.International Committee on Computational Linguistics,2020:1572-1582.
[29]ALT C,GABRYSZAK A,HENNIG L.TACRED Revisited:A Thorough Evaluation of the TACRED Relation Extraction Task[C]//Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.Association for Computational Linguistics,2020:1558-1569.
[30]ZHOU G Z,SU J,ZHANG J,et al.Exploring Various Know-ledge in Relation Extraction[C]//Proceedings of the 43rdAn-nual Meeting of the Association for Computational Linguistics(ACL’05).Association for Computational Linguistics,2005:427-434.
[31]ZHOU G,ZHANG M,JI D H,et al.Tree Kernel-Based Relation Extraction with Context-Sensitive Structured Parse Tree Information[C]//Proceedings of the 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computa-tional Natural Language Learning.Association for Computa-tional Linguistics,2007:728-736.
[32]ZHANG D,WANG D.Relation Classification via RecurrentNeural Network[OL].(2015-08) [2022-02-21].https://go.exlibris.link/swcDB6f5.
[33]ZHANG C,CUI C,GAO S,et al.Multi-Gram CNN-Based Self-Attention Model for Relation Classification[J].IEEE Access,2019,7:5343-5357.
[34]ZHANG S,ZHENG D,HU X,et al.Bidirectional Long Short-Term Memory Networks for Relation Classification[C]//Proceedings of the 29th Pacific Asia Conference on Language,Information and Computation.2015:73-78.
[35]ZHOU P,SHI W,TIAN J,et al.,Attention-Based Bidirectional Long Short-Term Memory Networks for Relation Classification[C]//Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics(Volume 2:Short Papers).Association for Computational Linguistics,2016:207-212.
[36]LEE J,SEO S,CHOI Y S.Semantic Relation Classification via Bidirectional LSTM Networks with Entity-Aware Attention Using Latent Entity Typing[J].Symmetry,2019,11(6):785.
[37]SANTOS C N D,XIANG B,ZHOU B.Classifying Relations by Ranking with Convolutional Neural Networks[C]//53rd An-nual Meeting of the Association for Computational Linguistics(ACS)/7th International Joint Conference on Natural Language Processing of the Asian Federation of Natural Language Processing(IJCNLP).Assoc Computational Linguistics-Acl,2015:626-634.
[38]ZHANG C,CUI C,GAO S,et al.Multi-Gram CNN-Based Self-Attention Model for Relation Classification[J].IEEE Access,2019,7:5343-5357.
[39]ZHANG Y,QI P,MANNING C D.Graph Convolution overPruned Dependency Trees Improves Relation Extraction[C]//Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing.Association for Computational Linguistics,2018:2205-2215.
[40]GUO Z,ZHANG Y,LU W.Attention Guided Graph Convolutional Networks for Relation Extraction[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics.Association for Computational Linguistics,2019:241-251.
[41]ZHOU L,WANG T,LIU Y.AWeighted GCN with Logical Adjacency Matrix for Relation Extraction[C]//The 24th European Conference on Artificial Intelligence.2020:2314-2321.
[42]ZENG D,LIU K,LAI S,et al.Relation classification via convolutional deep neural network[C]//Proceedings of COLING 2014,the 25th International Conference on Computational Linguistics:Technical Papers.Association for Computational Linguistics,2014:2335-2344.
[43]ZHOU L,WANG T,QU H,et al.A Weighted GCN with Logical Adjacency Matrix for Relation Extraction[C]//24th Euro-pean Conference on Artificial Intelligence(ECAI 2020).2020:2314-2321.
[44]MATTHEW E,MARK N,MOHIT I,et al.Deep contextualized word representations[C]//Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies.Association for Computational Linguistics,2018:2227-2237.
[45]RADFORD A,NARASIMHAN K,SALIMANS T.ImprovingLanguage Understanding by Generative Pre-Training[OL].(2018) [2023-3-13].https://s3-us-west-2.amazonaws.com/openai-assets/research-covers/language-unsupervised/language_understanding_paper.pdf.
[46]MIKE L,LIU Y,NAMAN G,et al.BART:Denoising Sequence-to-Sequence Pre-training for Natural Language Generation,Translation,and Comprehension[C]//Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.Association for Computational Linguistics,2020:7871-7880.
[47]RAFFEL C,SHAZEER N,ROBERTS A.Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer[OL].(2019-10) [2021-5-17].https://arxiv.org/abs/1910.10683.
[48]HU S,DING N,WANG H,et al.Knowledgeable Prompt-tu-ning:Incorporating Knowledge into Prompt Verbalizer for Text Classification[C]//Findings of the Association for Computational Linguistics:ACL 2022.Association for Computational Linguistics,2022:2225-2240.
[49]CHIA Y K,BING L,PORIA S,et al.RelationPrompt:Leveraging Prompts to Generate Synthetic Data for Zero-Shot Relation Triplet Extraction[C]//Findings of the Association for Computational Linguistics:ACL 2022.Association for Computational Linguistics,2022:45-57.
[50]LU Y,LIU Q,DAI D,et al.Unified Structure Generation forUniversal Information Extraction[C]//Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics(Volume 1:Long Papers).Association for Computational Linguistics,2022:5755-5772.
[51]KOJIMA T,GU S S,REID M,et al.Large Language Models are Zero-Shot Reasoners[J].arXiv:2205.11916,2021.
[52]TAYLOR S,YASAMAN R,ROBERT L,et al.AUTO-PROMPT:Eliciting Knowledge from Language Models with Automatically Generated Prompts[C]//Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing(EMNLP).Association for Computational Linguistics,2020:4222-4235.
[53]QIN G,JASON E.Learning How to Ask:Querying LMs with Mixtures of Soft Prompts[C]//Proceedings of the 2021 Confe-rence of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies.Association for Computational Linguistics,2021:5203-5212.
[54]TIMO S,HELMUT S,HINRICH S.Automatically Identifying Words That Can Serve as Labels for Few-Shot Text Classification[C]//Proceedings of the 28th International Conference on Computational Linguistics,Barcelona,Spain.International Committee on Computational Linguistics,2020:5569-5578.
[55]OSCAR S,OIER L,GORKA L,et al.Label Verbalization and Entailment for Effective Zero- and Few-Shot Relation Extraction[C]//Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing Online and Punta Cana,Dominican Republic.Association for Computational Linguistics,2021:1199-1212.
[56]CUI G,HU S,DING D,et al.Prototypical Verbalizer for Prompt-based Few-shot Tuning[C]//Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics(Volume 1:Long Papers),Dublin,Ireland.Association for Computational Linguistics,2022:7014-7024.
[57]FABIO P,TIM R,PATRICK L,et al.Language Models asKnowledge Bases?[J].arXiv:1909.01066,2019.
[58]ZHANG P,LU W.Better Few-Shot Relation Extraction withLabel Prompt Dropout[C]//Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing,Abu Dhabi.Association for Computational Linguistics,2022:6996-7006.
[59]ZHAO T Z,WALLACE E,FENG S,et al.Calibrate BeforeUse:Improving Few-Shot Performance of Language Models[J].arXiv:2102.09690,2021.
[60]JIANG ZB,XU F,JUN A.How Can We Know What Language Models Know[C]//Transactions of the Association for Computational Linguistics,Volume 8.MIT Press,2020:423-438.
[61]CUI L,WU Y,LIU J,et al.Template-Based Named Entity Re-cognition Using BART[C]//Findings of the Association for Computational Linguistics:ACL-IJCNLP 2021.Association for Computational Linguistics,2021:1835-1845.
[62]LEE D,KADAKIA A,TAN K,et al.Good Examples Make A Faster Learner:Simple Demonstration-based Learning for Low-resource NER[C]//Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics(Volume 1:Long Papers).Association for Computational Linguistics,2021:2687-2700.
[63]CHEN J W,LIU Q,LIN H Y,et al.Few-shot Named EntityRecognition with Self-describing Networks[C]//Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics(Volume 1:Long Papers).Association for Computational Linguistics,2022:5711-5722.
[64]JACOB D,CHANG M W,KENTON L.BERT:Pre-training of Deep Bidirectional Transformers for Language Understanding[C]//Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies,Volume 1(Long and Short Papers).Association for Computational Linguistics,2018:4171-4186.
[65]HAN X,ZHU H,YU P,et al.FewRel:A Large-Scale Supervised Few-Shot Relation Classification Dataset with State-of-the-Art Evaluation[C]//Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing.Association for Computational Linguistics,2018:4803-4809.
[66]DONG M,PAN C,LUO Z.MapRE:An Effective SemanticMapping Approach for Low-resource Relation Extraction[C]//Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing Online and Punta Cana,Dominican Republic.Association for Computational Linguistics,2021:2694-2704.
[67]LIU Y,HU J,WAN X,et al.A Simple yet Effective Relation Information Guided Approach for Few-Shot Relation Extraction[C]//Findings of the Association for Computational Linguistics:ACL 2022.Association for Computational Linguistics,2022:757-763.
[68]ZHANG Z,XU H,LIU Z Y,et al.ERNIE:Enhanced Language Representation with Informative Entities[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics.Association for Computational Linguistics,2019:1441-1451.
[69]YANG S,ZHANG Y F,NIU G L,et al.Enhanced few shot relation extration through concept[C]//Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing(Volume 2:Short Papers).Association for Computational Linguistics,2021:987-991.
[70]CHEN X,ZHANG N,XIE X,et al.KnowPrompt:Knowledge-aware Prompt-tuning with Synergistic Optimization for Relation Extraction[C]//The Web Conference 2021.Association for Computing Machinery,2022:2778-2788.
[71]HAN X,ZHAO W,DING N,et al.PTR:Prompt Tuning with Rules for Text Classification[OL].(2021-05) [2022-03-24].https://arxiv.org/abs/2105.11259.
[72]ZENG D J,LIU K,CHEN Y B,et al.Distant Supervision for Relation Extraction via Piecewise Convolutional Neural Networks[C]//Distant Supervision for Relation Extraction via Piecewise Convolutional Neural Network.Association for Computational Linguistics,2015:1753-1762.
[73]LIN Y K,SHEN S Q,LIU Z Y,et al.Neural Relation Extraction with Selective Attention over Instances[C]//Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics(Volume 1:Long Papers).Association for Computational Linguistics,2016:2124-2133.
[74]LIN X Y,LIU T Y,JIA W J,et al.Distantly Supervised Relation Extraction using Multi-Layer Revision Network and Confidence-based Multi-Instance Learning[C]//Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing Online and Punta Cana,Dominican Republic.Asso-ciation for Computational Linguistics,2021:165-174.
[75]HOGAN W,LI J C,SHANG J B.Fine-grained ContrastiveLearning for Relation Extraction[C]//Proceedings of the 2022 Conference on Empirical Methods in Natural Language Proces-sing Abu Dhabi.Association for Computational Linguistics,2022:1083-1095.
[76]DONG Y L,ZHANG T L,NAN H,et al.HiCLRE:A Hierarchical Contrastive Learning Framework for Distantly Supervised Relation Extraction[C]//Findings of the Association for Computational Linguistics:ACL 2022.Association for Computational Linguistics,2022:2567-2578.
[77]HU L,ZHANG L,SHI C,et al.Improving Distantly-SupervisedRelation Extraction with Joint Label Embedding[C]//Procee-dings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing(EMNLP-IJCNLP).Association for Computational Linguistics,2019:3821-3829.
[78]ZHANG N,DENG S,SUN Z,et al.Long-tail Relation Extraction via Knowledge Graph Embeddings and Graph Convolution Networks[C]//Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies,Volume 1(Long and Short Papers).Association for Computational Linguistics,2019:3016-3025.
[79]WEI J,DAI D,XIN X Y.ARNOR:Attention Regularizationbased Noise Reduction for Distant Supervision Relation Classification[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics.Association for Computational Linguistics,2019:1399-1408.
[80]MA R,GUI T,LI L,et al.SENT:Sentence-level Distant Relation Extraction via Negative Training[C]//Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing(Volume 1:Long Papers).Association for Computational Linguistics,2021:6201-6213.
[81]XIE C,LIANG J,LIU J,et al.Revisiting the Negative Data of Distantly Supervised Relation Extraction[C]//Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing(Volume 1:Long Papers).Association for Computational Linguistics,2021:3572-3581.
[82]ZHENG H,WEN R,CHEN X,et al.PRGC:Potential Relation and Global Correspondence Based Joint Relational Triple Extraction[C]//Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing(Vo-lume 1:Long Papers).Association for Computational Linguistics,2021:6225-6235.
[83]ZHENG S,HAO Y,LU D,et al.Joint entity and relation extraction based on a hybrid neural network[J].Neurocomputing,2017,257:59-66.
[84]ZHENG S,WANG F,BAO H,et al.Joint Extraction of Entities and Relations Based on a Novel Tagging Scheme[C]//Procee-dings of the 55th Annual Meeting of the Association for Computational Linguistics(Volume 1:Long Papers).Association for Computational Linguistics,2017:1227-1236.
[85]REN F L,ZHANG L H,YIN S J,et al.A Novel Global Feature-Oriented Relational Triple Extraction Model based on Table Filling[C]//Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing.Association for Computational Linguistics,2021:2646-2656.
[86]WANG Y,SUN C,WU Y,et al.UniRE:A Unified Label Space for Entity Relation Extraction[C]//Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing.Association for Computational Linguistics,2021:220-231.
[87]ZENG X,ZENG D,HE S,et al.Extracting Relational Facts by an End-to-End Neural Model with Copy Mechanism[C]//Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics(Volume 1:Long Papers).Association for Computational Linguistics,2018:506-514.
[88]LI X,YIN F,SUN Z,et al.Entity-Relation Extraction as Multi-Turn Question Answering[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics.Association for Computational Linguistics,2019:1340-1350.
[89]SUI D B,CHEN Y B,LIU K.Joint Entity and Relation Extraction with Set Prediction Networks[OL].(2020-11) [2021-03-24].https://arxiv.org/abs/2011.01675.
[90]BEKOULIS G,DELEU J,DEMEESTER T,et al.Joint entity recognition and relation extraction as a multi-head selection problem[J].Expert Systems with Applications,2018,114:34-45.
[91]WEI Z,SU J,WANG Y,et al.A Novel Cascade Binary Tagging Framework for Relational Triple Extraction[C]//Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.Association for Computational Linguistics,2020:1476-1488.
[92]LI X M,LUO X T,DONG C H,et al.TDEER:An EfficientTranslating Decoding Schema for Joint Extraction of Entities and Relations[C]//Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing,Dominican Republic.Association for Computational Linguistics,2021:8055-8064.
[93]YE D M,LIN Y K,LI P,et al.Packed Levitated Marker for Entity and Relation Extraction[C]//Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics(Volume 1:Long Papers).Association for Computational Linguistics,2022:4904-4917.
[94]FU T J,LI P H,MA W Y.GraphRel:Modeling Text as Relational Graphs for Joint Entity and Relation Extraction[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics.Association for Computational Linguistics,2019:1409-1418.
[95]YAN Z,ZHANG C,FU J,et al.A Partition Filter Network for Joint Entity and Relation Extraction[C]//Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing,Dominican Republic.Association for Computational Linguistics,2021:185-197.
[96]TONG M H,WANG S,XU B,et al.Learning from Miscellaneous Other-Class Words for few shot named entity recognition[C]//Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing(Volume 1:Long Papers).Association for Computational Linguistics,2021:6236-6247.
[97]WANG Y R,SHINDO H,MATSUMOTO Y J,et al.Nested Named Entity Recognition via Explicitly Excluding the Influence of the Best Path[C]//Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing(Volume 1:Long Papers).Association for Computational Linguistics,2021:3547-3557.
[98]LI F,LIN Z C,ZHANG M S.A Span-Based Model for JointOverlapped and Discontinuous Named Entity Recognition[C]//Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing(Volume 1:Long Papers).Association for Computational Linguistics,2021:4814-4828.
[99]ZHU E W,LI J P.Boundary Smoothing for Named Entity Re-cognition[C]//Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics(Volume 1:Long Papers).Association for Computational Linguistics,2022:7096-7108.
[100]YE D,LIN Y,LI P,et al.A Simple but Effective Pluggable Entity Lookup Table for Pre-trained Language Models[C]//Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics(Volume 2:Short Papers).Association for Computational Linguistics,2022:523-529.
[101]YANG A,WANG Q,LIU J,et al.Enhancing Pre-Trained Language Representations with Rich Knowledge[C]//Proceedings of the 57th Annual Meeting of the Association for Computa-tional Linguistics.Association for Computational Linguistics,2019:2346-2357.
[102]LU L,KONG F.Conversation-oriented entity relationship extraction with knowledge [C]//Proceedings of the 20th Chinese National Conference on Computational Linguistics.Chinese Information Processing Society of China,2021.
[103]ZHANG R,HRISTOVSKI D,SCHUTTE D,et al.Drug repurposing for COVID-19 via knowledge graph completion[J].Journal of Biomedical Informatics,2021,115:103696.
[104]YEW K C,BING L D,PORIA S.RelationPrompt:Leveraging Prompts to Generate Synthetic Data for Zero-Shot Relation Triplet Extraction[C]//Findings of the Association for Computational Linguistics:ACL 2022.Association for Computational Linguistics,2022:45-57.
[105]LU Y,BARTOLO M,MOORE A,et al.Fantastically Ordered Prompts and Where to Find Them:Overcoming Few-Shot Prompt Order Sensitivity[C]//Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics(Vo-lume 1:Long Papers).Association for Computational Linguistics,2021:8086-8098.
[106]SI C L,ZHANG Z Z,CHEN Y F,et al.READIN A Chinese Multi-Task Benchmark with realistic and diverse noises[OL].(2023-02) [2023-02-20].https://arxiv.org/abs/2302.07324.
[107]WANG X,GUI M,JIANG Y,et al.ITA:Image-Text Alignments for Multi-Modal Named Entity Recognition[C]//Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies Seattle.Association for Computational Linguistics,2022:3176-3189.
[108]LI Y Q,LI W J,NIE L Q.MMCoQA:Conversational Question Answering over Text,Tables,and Images[C]//Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics(Volume 1:Long Papers).Association for Computational Linguistics,2022:4220-4231.
[109]HAN J L,CHENG B, LU W.Exploring Task Difficulty forFew-Shot Relation Extraction[C]//Proceedings of the 2021 Conference on Empirical Methods in Natural Language Proces-sing.Association for Computational Linguistics,2021:2605-2616.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!