计算机科学 ›› 2023, Vol. 50 ›› Issue (4): 181-187.doi: 10.11896/jsjkx.220700164

• 人工智能 • 上一篇    下一篇

融合多粒度抽取式特征的关键词生成

甄田歌, 宋明阳, 景丽萍   

  1. 北京交通大学计算机与信息技术学院 北京 100044
    交通数据分析与挖掘北京市重点实验室(北京交通大学) 北京 100044
  • 收稿日期:2022-07-18 修回日期:2022-11-21 出版日期:2023-04-15 发布日期:2023-04-06
  • 通讯作者: 景丽萍(lpjing@bjtu.edu.cn)
  • 作者简介:(20120457@bjtu.edu.cn)
  • 基金资助:
    国家自然科学基金(61822601,61773050,61632004);北京市自然科学基金(Z180006);北京市科委项目(Z181100008918012)

Incorporating Multi-granularity Extractive Features for Keyphrase Generation

ZHEN Tiange, SONG Mingyang, JING Liping   

  1. School of Computer and Information Technology,Beijing Jiaotong University,Beijing 100044,China
    Beijing Key Lab of Traffic Data Analysis and Mining,Beijing Jiaotong University,Beijing 100044,China
  • Received:2022-07-18 Revised:2022-11-21 Online:2023-04-15 Published:2023-04-06
  • About author:ZHEN Tiange,born in 1997,bachelor.Her main research interests include machine learning and natural language processing.
    JING Liping,born in 1978,Ph.D,professor,Ph.D supervisor,is a senior member of China Computer Federation.Her main research interests include machine learning and its applications.
  • Supported by:
    National Natural Science Foundation of China(61822601,61773050,61632004),Natural Science Foundation of Beijing,China(Z180006) and Program of Beijing Municipal Science & Technology Commission(Z181100008918012).

摘要: 关键词是概括给定文本核心主题及关键内容的一组短语。在信息过载日益严重的今天,从给定的大量文本信息中预测出具有其中心思想的关键词至关重要。因此,关键词预测作为自然语言处理的基本任务之一,受到越来越多研究者的关注。其对应方法主要包括两类:关键词抽取和关键词生成。关键词抽取是从给定文本中快速、准确地抽取文中出现过的显著性短语作为关键词。与关键词抽取不同,关键词生成既能预测出现在给定文本中的关键词,也能预测未出现在给定文本中的关键词。总而言之,这两类方法各有优劣。然而,现有的关键词生成工作大多忽视了抽取式特征可能为关键词生成模型带来的潜在收益。抽取式特征能指明原文本的重要片段,对于模型学习原文本的深层语义表示起到重要作用。因此,结合抽取式和生成式方法的优势,提出了一种新的融合多粒度抽取式特征的关键词生成模型(incorporating Multi-Granularity Extractive features for keyphrase generation,MGE-Net)。在一系列公开数据集上的实验结果表明,和近年来的关键词生成模型相比,所提模型在大多数评价指标上取得了显著的性能提升。

关键词: 自然语言处理, 序列到序列, 关键词生成, 抽取式特征, 多任务学习

Abstract: Keyphrase is a set of phrases that summarizes the core theme and key content of a given text.At present,information overload is becoming more and more serious,it is crucial to predict phrases with their central ideas for a given large amount of textual information.Therefore,keyphrase prediction,as one of the basic tasks of natural language processing,has received more and more attention from research scholars.Its corresponding methods mainly contain two categories,namely keyphrase extraction and keyphrase generation.Keyphrase extraction is the fast and accurate extraction of salient phrases that appear in the given text.Unlike keyphrase extraction,keyphrase generation predicts both phrases that appear in the given text and those do not appear in the given text.In summary,both have their advantages and disadvantages.However,most of the existing work on keyphrase ge-neration has ignored the potential benefits that extractive features may bring to keyphrase generation models.Extractive features can indicate important fragments of the original text and play an important role for the model to learn the deep semantic representation of the original text.Therefore,combining the advantages of extractive and generative approaches,this paper proposes a new keyphrase generation model incorporating multi-granularity extractive features(MGE-Net).Compared with recent keyphrase ge-neration models on a series of publicly available datasets,the proposed model achieves significant performance improvements in most evaluation metrics.

Key words: Natural language processing, Sequence-to-Sequence, Keyphrase generation, Extractive features, Multi-task learning

中图分类号: 

  • TP391
[1]JONES S,STAVELEY M S.Phrasier:a system for interactivedocument retrieval using keyphrases[C]//Proceedings of the 22nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval.New York:ACM,1999:160-167.
[2]HULTH A,MEGYESI B.A study on automatically extracted keywords in text categorization[C]//Proceedings of the 21st International Conference on Computational Linguistics and 44th Annual Meeting of the Association for Computational Linguistics.Stroudsburg,PA:ACL,2006:537-544.
[3]QAZVINIAN V,RADEV D,ÖZGÜR A.Citation summari-zation through keyphrase extraction[C]//Proceedings of the 23rd International Conference on Computational Linguistics(COLING 2010).New York:ACM,2010:895-903.
[4]TOMOKIYO T,HURST M.A language model approach tokeyphrase extraction[C]//Proceedings of the ACL 2003 Workshop on Multiword Expressions:Analysis,Acquisition and Treatment.Stroudsburg,PA:ACL,2003:33-40.
[5]LIU Z Y,LI P,ZHENG Y B,et al.Clustering to find exemplar terms for keyphrase extraction[C]//Proceedings of the 2009 Conference on Empirical Methods in Natural Language Proces-sing.Stroudsburg,PA:ACL,2009:257-266.
[6]CHO K,VAN M B,GULCEHRE C,et al.Learning phrase representations using RNN encoder-decoder for statistical machine translation[J].arXiv:1406.1078,2014.
[7]SUTSKEVER I,VINYALS O,LE Q V.Sequence to sequence learning with neural networks[C]//Proceedings of the 27th International Conference on Neural Information Processing Systems-Volume 2.Cambridge,MA:MIT Press,2014:3104-3112.
[8]ALZAIDY R,CARAGEA C,GILES C L.Bi-LSTM-CRF se-quence labeling for keyphrase extraction from scholarly documents[C]//The World Wide Web Conference.New York:ACM,2019:2551-2557.
[9]BASALDELLA M,CHIARADIA G,TASSO C.Evaluatinganaphora and coreference resolution to improve automatic keyphrase extraction[C]//Proceedings of COLING 2016,the 26th International Conference on Computational Linguistics:Technical Papers.New York:ACM,2016:804-814.
[10]LOPEZ P,ROMARY L.HUMB:Automatickey term extraction from scientific articles in GROBID[C]//Proceedings of the 5th International Workshop on Semantic Evaluation.Stroudsburg,PA:ACL,2010:248-251.
[11]MIHALCEA R,TARAU P.Textrank:Bringing order into text[C]//Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing.Stroudsburg,PA:ACL,2004:404-411.
[12]MENG R,ZHAO S Q,HAN S G,et al.Deep keyphrase generation[C]//Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics(Volume 1:Long Papers).Stroudsburg,PA:ACL,2017:582-592.
[13]YUAN X D,WANG T,MENG R,et al.Onesize does not fit all:Generating and evaluating variable number of keyphrases[C]//Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.Stroudsburg,PA:ACL,2020:7961-7975.
[14]CHEN J,ZHANG X M,WU Y,et al.Keyphrase generationwith correlation constraints[C]//Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing.Stroudsburg,PA:ACL,2018:4057-4066.
[15]ZHAO J,ZHANG Y X.Incorporating linguistic constraints into keyphrase generation[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics.Stroudsburg,PA:ACL,2019:5224-5233.
[16]BAHULEYAN H,EL A L.Diverse keyphrase generation with neural unlikelihood training[C]//Proceedings of the 28th International Conference on Computational Linguistics.New York:ACM,2020:5271-5287.
[17]YE H,WANG L.Semi-supervised learning for neural keyphrase generation[C]//Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing.Stroudsburg,PA:ACL,2018:4142-4153.
[18]CHEN W,GAO Y F,ZHANG J N,et al.Title-guided encoding for keyphrase generation[C]//Proceedings of the AAAI Confe-rence on Artificial Intelligence.Menlo Park,CA:AAAI,2019:6268-6275.
[19]ZHANG Y,FANG Y,XIAO W D.Deep keyphrase generation with a convolutional sequence to sequence model[C]//2017 4th International Conference on Systems and Informatics(ICSAI).Piscataway,NJ:IEEE,2017:1477-1485.
[20]CHAN H P,CHEN W,WANG L,et al.Neural keyphrase ge-neration via reinforcement learning with adaptive rewards[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics.Stroudsburg,PA:ACL,2019:2163-2174.
[21]WANG Y,LI J,CHAN H P,et al.Topic-aware neural key-phrase generation for social media language[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics.Stroudsburg,PA:ACL,2019:2516-2526.
[22]CHEN W,CHAN H P,LI P J,et al.Exclusive hierarchical decoding for deep keyphrase generation[C]//Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.Stroudsburg,PA:ACL,2020:1095-1105.
[23]AHMAD W,BAI X,LEE S,et al.Select,extract and generate:Neural keyphrase generation with layer-wise coverage attention[C]//Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confernece on Natural Language Processing(Volume 1:Long Papers).Stroudsburg,PA:ACL,2021:1389-1404.
[24]BA J L,KIROS J R,HINTON G E.Layer normalization[J].arXiv:1607.06450,2016.
[25]HULTH A.Improved automatic keyword extraction given morelinguistic knowledge[C]//Proceedings of the 2003 Conference on Empirical Methods in Natural Language Processing.Stroudsburg,PA:ACL,2003:216-223.
[26]KRAPIVIN M,AUTAEU A,MARCHESE M.Large datasetfor keyphrases extraction[R].Trento,Italy:Information Engineering and Computer Science Department of Trento Univer-sity,2009.
[27]NGUYEN T D,KAN M Y.Keyphrase extraction in scientific publications[C]//Proceedings of the 10th International Confe-rence on Asian Digital Libraries.Berlin:Springer,2007:317-326.
[28]KINGMA D P,BA J.Adam:A method for stochastic optimization[J].arXiv:1412.6980,2014.
[1] 孙凯丽, 罗旭东, 罗有容.
预训练语言模型的应用综述
Survey of Applications of Pretrained Language Models
计算机科学, 2023, 50(1): 176-184. https://doi.org/10.11896/jsjkx.220800223
[2] 郑诚, 梅亮, 赵伊研, 张苏航.
基于双向注意力机制和门控图卷积网络的文本分类方法
Text Classification Method Based on Bidirectional Attention and Gated Graph Convolutional Networks
计算机科学, 2023, 50(1): 221-228. https://doi.org/10.11896/jsjkx.211100095
[3] 闫佳丹, 贾彩燕.
基于双图神经网络信息融合的文本分类方法
Text Classification Method Based on Information Fusion of Dual-graph Neural Network
计算机科学, 2022, 49(8): 230-236. https://doi.org/10.11896/jsjkx.210600042
[4] 侯钰涛, 阿布都克力木·阿布力孜, 哈里旦木·阿布都克里木.
中文预训练模型研究进展
Advances in Chinese Pre-training Models
计算机科学, 2022, 49(7): 148-163. https://doi.org/10.11896/jsjkx.211200018
[5] 杜丽君, 唐玺璐, 周娇, 陈玉兰, 程建.
基于注意力机制和多任务学习的阿尔茨海默症分类
Alzheimer's Disease Classification Method Based on Attention Mechanism and Multi-task Learning
计算机科学, 2022, 49(6A): 60-65. https://doi.org/10.11896/jsjkx.201200072
[6] 李小伟, 舒辉, 光焱, 翟懿, 杨资集.
自然语言处理在简历分析中的应用研究综述
Survey of the Application of Natural Language Processing for Resume Analysis
计算机科学, 2022, 49(6A): 66-73. https://doi.org/10.11896/jsjkx.210600134
[7] 赵凯, 安卫超, 张晓宇, 王彬, 张杉, 相洁.
共享浅层参数多任务学习的脑出血图像分割与分类
Intracerebral Hemorrhage Image Segmentation and Classification Based on Multi-taskLearning of Shared Shallow Parameters
计算机科学, 2022, 49(4): 203-208. https://doi.org/10.11896/jsjkx.201000153
[8] 杨晓宇, 殷康宁, 候少麒, 杜文仪, 殷光强.
基于特征定位与融合的行人重识别算法
Person Re-identification Based on Feature Location and Fusion
计算机科学, 2022, 49(3): 170-178. https://doi.org/10.11896/jsjkx.210100132
[9] 张虎, 柏萍.
融入句子中远距离词语依赖的图卷积短文本分类方法
Graph Convolutional Networks with Long-distance Words Dependency in Sentences for Short Text Classification
计算机科学, 2022, 49(2): 279-284. https://doi.org/10.11896/jsjkx.201200062
[10] 朱艺娜, 曹阳, 钟靖越, 郑泳智.
事件抽取技术研究综述
Survey on Event Extraction Technology
计算机科学, 2022, 49(12): 264-273. https://doi.org/10.11896/jsjkx.211100226
[11] 刘晓影, 王淮, 乌吉斯古愣.
基于GAN和中文词汇网的文本摘要技术
GAN and Chinese WordNet Based Text Summarization Technology
计算机科学, 2022, 49(12): 301-304. https://doi.org/10.11896/jsjkx.210600166
[12] 阿布都克力木·阿布力孜, 张雨宁, 阿力木江·亚森, 郭文强, 哈里旦木·阿布都克里木.
预训练语言模型的扩展模型研究综述
Survey of Research on Extended Models of Pre-trained Language Models
计算机科学, 2022, 49(11A): 210800125-12. https://doi.org/10.11896/jsjkx.210800125
[13] 徐晖, 王中卿, 李寿山, 张民.
结合情感信息的个性化对话生成
Personalized Dialogue Generation Integrating Sentimental Information
计算机科学, 2022, 49(11A): 211100019-6. https://doi.org/10.11896/jsjkx.211100019
[14] 王浩宇.
软件需求工程技术综述
Review on Technologies of Requirement Engineering of Software
计算机科学, 2022, 49(11A): 210900132-14. https://doi.org/10.11896/jsjkx.210900132
[15] 郑顺源, 胡良校, 吕晓倩, 孙鑫, 张盛平.
基于边缘引导的自校正皮肤检测
Edge Guided Self-correction Skin Detection
计算机科学, 2022, 49(11): 141-147. https://doi.org/10.11896/jsjkx.220600012
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!