计算机科学 ›› 2021, Vol. 48 ›› Issue (8): 60-65.doi: 10.11896/jsjkx.200700008

• 数据库&大数据&数据科学* • 上一篇    下一篇

基于细粒度差异特征的文本匹配方法

王胜, 张仰森, 陈若愚, 向尕   

  1. 北京信息科技大学智能信息处理研究所 北京100101
  • 收稿日期:2020-07-01 修回日期:2020-08-20 发布日期:2021-08-10
  • 通讯作者: 张仰森(zhangyangsen@163.com)
  • 基金资助:
    国家自然科学基金(61772081);国家重点研发计划(2018YFB1403104);北京信息科技大学科研基金(2035008)

Text Matching Method Based on Fine-grained Difference Features

WANG Sheng, ZHANG Yang-sen, CHEN Ruo-yu, XIANG Ga   

  1. Institute of Intelligent Information Processing,Beijing Information Science and Technology University,Beijing 100101,China
  • Received:2020-07-01 Revised:2020-08-20 Published:2021-08-10
  • About author:WANG Sheng,born in 1996,postgra-duate.His main research interests include natural language processing and machine learning.(1028742881@qq.com)ZHANG Yang-sen,born in 1962,postdoctoral,professor,is a distinguished member of China Computer Federation.His main research interests include na-tural language processing and artificial intelligence.
  • Supported by:
    National Natural Science Foundation of China(61772081),National Key Research and Development Plan(2018YFB1403104) and Research Fund of Beijing Information Science and Technology University(2035008).

摘要: 文本匹配是检索系统中的关键技术之一。针对现有文本匹配模型对文本语义差异捕获不准确的问题,文中提出了一种基于细粒度差异特征的文本匹配方法。首先,使用预训练模型作为基础模型对匹配文本进行语义的抽取与初步匹配;然后,引入对抗学习的思想,在模型的编码阶段人为构造虚拟对抗样本进行训练,以提升模型的学习能力与泛化能力;最后,通过引入文本的细粒度差异特征,纠正文本匹配的初步预测结果,有效提升了模型对细粒度差异特征的捕获能力,进而提升了文本匹配模型的性能。在两个数据集上进行了实验验证,其中在LCQMC数据集上的实验结果显示,所提方法在ACC性能指标上达到了88.96%,优于已知的最好模型。

关键词: 差异特征, 对抗学习, 文本匹配, 语义相似度, 预训练模型

Abstract: Text matching is one of the key technologies in the retrieval system.Aiming at the problem that the existing text ma-tching models can't capture the semantic differences of texts accurately,this paper proposes a text matching method based on fine-grained difference features.Firstly,the pre-trained model is used as the basic model to extract the matching text semantics and preliminarily match them.Then,the idea of adversarial learning is introduced in the embedding layer,and by constructing the virtual confrontation samples artificially for training,the learning ability and generalization ability of the model are improved.Finally,by introducing the fine-grained difference feature of the text to correct the preliminary prediction results of the text ma-tching,the capture ability of the model for fine-grained difference features is effectively improved,and then the performance of the text matching model is improved.In this paper,two datasets are tested,and the experiment on LCQMC dataset shows that the performance index of ACC is 88.96%,which is better than the best known model.

Key words: Adversarial learning, Difference feature, Pre-trained model, Semantic similarity, Text match

中图分类号: 

  • TP391.1
[1]LI X.Research on paragraph retrieval technology for questionanswering system[D].Chengdu:University of Science and Technology of China,2010.
[2]SALTON G,BUCKLEY C.Term-weighting approaches in automatic text retrieval[J].Information Processing & Management,1988,24(5):513-523.
[3]SONG F,CROFT W B.A general language model for information retrieval[C]//Proceedings of the Eighth International Conference on Information and Knowledge Management.1999:316-321.
[4]LE Q,MIKOLOV T.Distributed representations of sentencesand documents[C]//International Conference on Machine Learning.2014:1188-1196.
[5]LOGESWARAN L,LEE H.An efficient framework for learning sentence representations[J].arXiv:1803.02893,2018.
[6]CER D,YANG Y,KONG S,et al.Universal sentence encoder[J].arXiv:1803.11175,2018.
[7]YIN W,SCHÜTZE H,XIANG B,et al.Abcnn:Attention-based convolutional neural network for modeling sentence pairs[J].Transactions of the Association for Computational Linguistics,2016,4:259-272.
[8]CHEN Q,ZHU X,LING Z,et al.Enhanced lstm for natural language inference[J].arXiv:1609.06038,2016.
[9]WANG Z,HAMZA W,FLORIAN R.Bilateral multi-perspective matching for natural language sentences[J]. arXiv:1702.03814,2017.
[10]RADFORD A,NARASIMHAN K,SALIMANS T,et al.Improving language understanding with unsupervised learning[R/OL].Technical Report,OpenAI,2018.https://openai.com/blog/language-unsupervised/.
[11]DEVLIN J,CHANG M W,LEE K,et al.Bert:Pre-training ofdeep bidirectional transformers for language understanding[J].arXiv:1810.04805,2018.
[12]RADFORD A,WU J,CHILD R,et al.Language models are unsupervised multitask learners[J].OpenAI Blog,2019,1(8):9.
[13]SUN Y,WANG S,LI Y,et al.Ernie:Enhanced representation through knowledge integration[J].arXiv:1904.09223,2019.
[14]RAFFEL C,SHAZEER N,ROBERTS A,et al.Exploring the limits of transfer learning with a unified text-to-text transformer[J].arXiv:1910.10683,2019.
[15]HU W,DANG A,TAN Y.A Survey of State-of-the-Art Short Text Matching Algorithms[C]//International Conference on Data Mining and Big Data.Singapore:Springer,2019:211-219.
[16]SAKATA W,SHIBATA T,TANAKA R,et al.FAQ retrieval using query-question similarity and BERT-based query-answer relevance[C]//Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval.2019:1113-1116.
[17]WU Y,WANG R J.Application of BERT-Based Semantic Ma-tching Algorithm in Question Answering System[J].Instrumentation Technology,2020(6):19-22,30.
[18]WANG N Z.Research on improved text representation model based on BERT[D].Southwest University.2019.
[19]GOODFELLOW I J,SHLENS J,SZEGEDY C.Explaining and harnessing adversarial examples[J].arXiv:1412.6572,2014.
[20]ZHU C,CHENG Y,GAN Z,et al.Freelb:Enhanced adversarial training for natural language understanding[C]//International Conference on Learning Representations.2019.
[21]YAN J,BRACEWELL D B,REN F,et al.A semantic analyzer for aiding emotion recognition in Chinese[C]//International Conference on Intelligent Computing.Berlin,Heidelberg:Springer,2006:893-901.
[22]SONG Y,SHI S,LI J,et al.Directional skip-gram:Explicitlydistinguishing left and right context for word embeddings[C]//Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies,Volume 2 (Short Papers).2018:175-180.
[23]LIU X,CHEN Q,DENG C,et al.Lcqmc:A large-scale chinese question matching corpus[C]//Proceedings of the 27th International Conference on Computational Linguistics.2018:1952-1962.
[24]ZHANG X,LU W,ZHANG G,et al.Chinese Sentence Semantic Matching Based on Multi-Granularity Fusion Model[C]//Paci-fic-Asia Conference on Knowledge Discovery and Data Mining.Cham:Springer,2020:246-257.
[25]LIU W,ZHOU P,ZHAO Z,et al.K-BERT:Enabling Language Representation with Knowledge Graph[C]//AAAI.2020:2901-2908.
[26]CHEN J,CAO C,JIANG X.SiBert:Enhanced Chinese Pre-trained Language Model with Sentence Insertion[C]//Procee-dings of The 12th Language Resources and Evaluation Confe-rence.2020:2405-2412.
[27]MENG Y,WU W,WANG F,et al.Glyce:Glyph-vectors forChinese character representations[C]//Advances in Neural Information Processing Systems.2019:2746-2757.
[28]CUI Y,CHE W,LIU T,et al.Pre-training with whole word masking for chinese bert[J].arXiv:1906.08101,2019.
[29]KOEHN P.Statistical significance tests for machine translation evaluation[C]//Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing.2004:388-395.
[1] 吕晓锋, 赵书良, 高恒达, 武永亮, 张宝奇.
基于异质信息网的短文本特征扩充方法
Short Texts Feautre Enrichment Method Based on Heterogeneous Information Network
计算机科学, 2022, 49(9): 92-100. https://doi.org/10.11896/jsjkx.210700241
[2] 曹晓雯, 梁美玉, 鲁康康.
基于细粒度语义推理的跨媒体双路对抗哈希学习模型
Fine-grained Semantic Reasoning Based Cross-media Dual-way Adversarial Hashing Learning Model
计算机科学, 2022, 49(9): 123-131. https://doi.org/10.11896/jsjkx.220600011
[3] 曾志贤, 曹建军, 翁年凤, 蒋国权, 徐滨.
基于注意力机制的细粒度语义关联视频-文本跨模态实体分辨
Fine-grained Semantic Association Video-Text Cross-modal Entity Resolution Based on Attention Mechanism
计算机科学, 2022, 49(7): 106-112. https://doi.org/10.11896/jsjkx.210500224
[4] 侯钰涛, 阿布都克力木·阿布力孜, 哈里旦木·阿布都克里木.
中文预训练模型研究进展
Advances in Chinese Pre-training Models
计算机科学, 2022, 49(7): 148-163. https://doi.org/10.11896/jsjkx.211200018
[5] 赵丹丹, 黄德根, 孟佳娜, 董宇, 张攀.
基于BERT-GRU-ATT模型的中文实体关系分类
Chinese Entity Relations Classification Based on BERT-GRU-ATT
计算机科学, 2022, 49(6): 319-325. https://doi.org/10.11896/jsjkx.210600123
[6] 刘硕, 王庚润, 彭建华, 李柯.
基于混合字词特征的中文短文本分类算法
Chinese Short Text Classification Algorithm Based on Hybrid Features of Characters and Words
计算机科学, 2022, 49(4): 282-287. https://doi.org/10.11896/jsjkx.210200027
[7] 侯宏旭, 孙硕, 乌尼尔.
蒙汉神经机器翻译研究综述
Survey of Mongolian-Chinese Neural Machine Translation
计算机科学, 2022, 49(1): 31-40. https://doi.org/10.11896/jsjkx.210900006
[8] 刘立波, 苟婷婷.
融合深度典型相关分析和对抗学习的跨模态检索
Cross-modal Retrieval Combining Deep Canonical Correlation Analysis and Adversarial Learning
计算机科学, 2021, 48(9): 200-207. https://doi.org/10.11896/jsjkx.200600119
[9] 汤世征, 张岩峰.
DragDL:一种易用的深度学习模型可视化构建系统
DragDL:An Easy-to-Use Graphical DL Model Construction System
计算机科学, 2021, 48(8): 220-225. https://doi.org/10.11896/jsjkx.200900045
[10] 吕乐宾, 刘群, 彭露, 邓维斌, 王崇宇.
结合多粒度信息的文本匹配融合模型
Text Matching Fusion Model Combining Multi-granularity Information
计算机科学, 2021, 48(6): 196-201. https://doi.org/10.11896/jsjkx.200700100
[11] 詹瑞, 雷印杰, 陈训敏, 叶书函.
基于多重差异特征网络的街景变化检测
Street Scene Change Detection Based on Multiple Difference Features Network
计算机科学, 2021, 48(2): 142-147. https://doi.org/10.11896/jsjkx.200500158
[12] 吴俣, 李舟军.
检索式聊天机器人技术综述
Survey on Retrieval-based Chatbots
计算机科学, 2021, 48(12): 278-285. https://doi.org/10.11896/jsjkx.210900250
[13] 黄欣, 雷刚, 曹远龙, 陆明名.
基于深度学习的交互式问答研究综述
Review on Interactive Question Answering Techniques Based on Deep Learning
计算机科学, 2021, 48(12): 286-296. https://doi.org/10.11896/jsjkx.210100209
[14] 张玉帅, 赵欢, 李博.
基于BERT和BiLSTM的语义槽填充
Semantic Slot Filling Based on BERT and BiLSTM
计算机科学, 2021, 48(1): 247-252. https://doi.org/10.11896/jsjkx.191200088
[15] 倪海清, 刘丹, 史梦雨.
基于语义感知的中文短文本摘要生成模型
Chinese Short Text Summarization Generation Model Based on Semantic-aware
计算机科学, 2020, 47(6): 74-78. https://doi.org/10.11896/jsjkx.190600006
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!