计算机科学 ›› 2021, Vol. 48 ›› Issue (2): 212-216.doi: 10.11896/jsjkx.200700137

• 人工智能 • 上一篇    下一篇

基于回复生成的对话意图预测

王博宇, 王中卿, 周国栋   

  1. 苏州大学计算机科学与技术学院 江苏 苏州215006
  • 收稿日期:2020-07-21 修回日期:2020-09-01 出版日期:2021-02-15 发布日期:2021-02-04
  • 通讯作者: 王中卿(wangzq.antony@gmail.com)
  • 作者简介:wby990108@163.com
  • 基金资助:
    国家自然科学基金(61806137,61702149,61836007,61702518);江苏省高等学校自然科学研究面上项目(18KJB520043)

Dialogue Act Prediction Based on Response Generation

WANG Bo-yu, WANG Zhong-qing, ZHOU Guo-dong   

  1. School of Computer Science and Technology,Soochow University,Suzhou,Jiangsu 215006,China
  • Received:2020-07-21 Revised:2020-09-01 Online:2021-02-15 Published:2021-02-04
  • About author:WANG Bo-yu,born in 1999,postgra-duate.His main research interests include natrual language processing and machine learning.
    WANG Zhong-qing,born in 1987,Ph.D,lecturer.His main research interests include natrual language processing,sentiment analysis and event extraction.
  • Supported by:
    This work was supported by the National Natural Science Foundation of China(61806137,61702149,61836007,61702518) and Jiangsu High School Research Grant(18KJB520043).

摘要: 随着人机对话系统的不断发展,让计算机能够准确理解对话者的对话意图,并根据对话的历史信息对回复进行意图预测,对于人机对话系统有着十分重要的意义。已有研究重点关注根据对话文本和已有标签对回复进行意图预测,但是,在很多场景下回复可能并没有生成。因此,文中提出了一种结合回复生成的对话意图预测模型。在生成部分,使用Seq2Seq结构,根据对话历史信息生成文本,作为对话中未来回复的文本信息;在分类部分,利用LSTM模型,将生成的回复文本与已有的对话信息转变为子句级别的表示,并结合注意力机制突出同一轮次对话句与生成回复的联系。实验结果表明,所提出的模型相比简单基线模型取得了2.54%的F1-score提升,并且联合训练的方式有助于提升模型性能。

关键词: 对话意图, 文本生成, 预测模型, 注意力机制

Abstract: With the continuous development of the human-machine dialogue system,it is of great significance for the computer to accurately understand the speaker's dialogue act and predict the act of response according to the history information of the dialogue.Previous research work focus on act prediction of responses based on dialogue text and existing labels.But in many scena-rios,the reply has not been generated.Therefore,this paper proposes a dialogue act prediction model based on reply generation.In the generation part,the Seq2Seq structure is used to generate text based on the conversation history information as text information for future replies in the conversation;in the classification part,the LSTM model is used to express the generated reply text and the existing conversation information as clause level representations.Combined with the attention mechanism,it highlights the connection between the dialogue sentence of the same round and the generated response.The experimental results show that the proposed model a chieves a 2.54% F1-score improvement compared to the simple baseline model,and the joint training method contributes to the improvement of model performance.

Key words: Attention mechanism, Dialogue act, Prediction model, Text generation

中图分类号: 

  • TP391
[1] TUR G,CELIKYILMAZ A,HAKKANI-TÜR D.Latent se-mantic modeling for slot filling in conversational understanding[C]//2013 IEEE International Conference on Acoustics,Speech and Signal Processing.IEEE,2013:8307-8311.
[2] EYBEN F,WÖLLMER M,GRAVES A,et al.On-line emotion recognition in a 3-D activation-valence-time continuum using acoustic and linguistic cues[J].Journal on Multimodal User Interfaces,2010,3(1/2):7-19.
[3] LIU J,LI Y L,LIN M.Review of Intent Detection Methods in Human-Machine Dialogue System[J].Computer Engineering and Applications,2019,55(12):1-7,43.
[4] LI W B,ZHANG L,SHU X.Application and Research on Generative Automatic Question Answering System Based on Seq2Seq [J].Modern Computer,2017(36):59-62.
[5] HOCHREITER S,SCHMIDHUBER J.Long short-term memory[J].Neural Computation,1997,9(8):1735-1780.
[6] BAHDANAU D,CHO K,BENGIO Y.Neural machine translation by jointly learning to align and translate[J/OL].https://arxiv.org/abs/1409.0473.
[7] KALCHBRENNER N,BLUNSOM P.Recurrent Convolutional Neural Networks for Discourse Compositionality[C]//Procee-dings of the Workshop on Continuous Vector Space Models and their Compositionality.2013:119-126.
[8] MIKOLOV T,KARAFIT M,BURGET L,et al.Recurrent neural network based language model [C]//Proceedings of 11th Annual Conference of the International Speech Communication Association.2010:1045-1048.
[9] KHANPOUR H,GUNTAKANDLA N,NIELSEN R.Dialogue act classification in domain-independent conversations using a deep recurrent neural network[C]//Proceedings of COLING 2016,the 26th International Conference on Computational Linguistics.2016:2012-2021.
[10] KUMAR H,AGARWAL A,DASGUPTA R,et al.Dialogue act sequence labeling using hierarchical encoder with crf[C]//Thirty-Second AAAI Conference on Artificial Intelligence.2018:3440-3447.
[11] LAFFERTY J.Conditional random fields:Probabilistic modelsfor segmenting and labeling sequence data[C]//Proceedings of the 18th International Conference on Machine Learning.2001.
[12] RAHEJA V,TETREAULT J.Dialogue Act Classification with Context-Aware Self-Attention[C]//Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies.2019:3727-3733.
[13] LI R,LIN C,COLLINSON M,et al.A Dual-Attention Hierarchical Recurrent Neural Networkfor Dialogue Act Classification[C]//Proceedings of the 23rd Conference on Computational Natural Language Learning.ACL,2019:383-392.
[14] TANAKA K,TAKAYAMA J,ARASE Y.Dialogue-Act Prediction of Future Responses Based on Conversation History[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics:Student Research Workshop.2019:197-202.
[15] CHO K,VAN MERRI$\widetilde{E}$NBOER B,GULCEHRE C,et al.Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation[C]//Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP).2014:1724-1734.
[16] FREITAG M,AL-ONAIZAN Y.Beam Search Strategies forNeural Machine Translation[J].Association for Computational Linguistics,2017,2017:56.
[17] VASWANI A,SHAZEER N,PARMAR N,et al.Attention isall you need[C]//Advances in Neural Information Processing Systems.2017:5998-6008.
[18] LI Y,SU H,SHEN X,et al.DailyDialog:A Manually Labelled Multi-turn Dialogue Dataset[C]//Procee-dings of the Eighth International Joint Conference on Natural Language Processing.2017:986-995.
[19] PAPINENI K,ROUKOS S,WARD T,et al.BLEU:a method for automatic evaluation of machine translation[C]//Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics.2002:311-318.
[1] 周芳泉, 成卫青.
基于全局增强图神经网络的序列推荐
Sequence Recommendation Based on Global Enhanced Graph Neural Network
计算机科学, 2022, 49(9): 55-63. https://doi.org/10.11896/jsjkx.210700085
[2] 戴禹, 许林峰.
基于文本行匹配的跨图文本阅读方法
Cross-image Text Reading Method Based on Text Line Matching
计算机科学, 2022, 49(9): 139-145. https://doi.org/10.11896/jsjkx.220600032
[3] 周乐员, 张剑华, 袁甜甜, 陈胜勇.
多层注意力机制融合的序列到序列中国连续手语识别和翻译
Sequence-to-Sequence Chinese Continuous Sign Language Recognition and Translation with Multi- layer Attention Mechanism Fusion
计算机科学, 2022, 49(9): 155-161. https://doi.org/10.11896/jsjkx.210800026
[4] 熊丽琴, 曹雷, 赖俊, 陈希亮.
基于值分解的多智能体深度强化学习综述
Overview of Multi-agent Deep Reinforcement Learning Based on Value Factorization
计算机科学, 2022, 49(9): 172-182. https://doi.org/10.11896/jsjkx.210800112
[5] 饶志双, 贾真, 张凡, 李天瑞.
基于Key-Value关联记忆网络的知识图谱问答方法
Key-Value Relational Memory Networks for Question Answering over Knowledge Graph
计算机科学, 2022, 49(9): 202-207. https://doi.org/10.11896/jsjkx.220300277
[6] 汪鸣, 彭舰, 黄飞虎.
基于多时间尺度时空图网络的交通流量预测模型
Multi-time Scale Spatial-Temporal Graph Neural Network for Traffic Flow Prediction
计算机科学, 2022, 49(8): 40-48. https://doi.org/10.11896/jsjkx.220100188
[7] 姜梦函, 李邵梅, 郑洪浩, 张建朋.
基于改进位置编码的谣言检测模型
Rumor Detection Model Based on Improved Position Embedding
计算机科学, 2022, 49(8): 330-335. https://doi.org/10.11896/jsjkx.210600046
[8] 朱承璋, 黄嘉儿, 肖亚龙, 王晗, 邹北骥.
基于注意力机制的医学影像深度哈希检索算法
Deep Hash Retrieval Algorithm for Medical Images Based on Attention Mechanism
计算机科学, 2022, 49(8): 113-119. https://doi.org/10.11896/jsjkx.210700153
[9] 孙奇, 吉根林, 张杰.
基于非局部注意力生成对抗网络的视频异常事件检测方法
Non-local Attention Based Generative Adversarial Network for Video Abnormal Event Detection
计算机科学, 2022, 49(8): 172-177. https://doi.org/10.11896/jsjkx.210600061
[10] 闫佳丹, 贾彩燕.
基于双图神经网络信息融合的文本分类方法
Text Classification Method Based on Information Fusion of Dual-graph Neural Network
计算机科学, 2022, 49(8): 230-236. https://doi.org/10.11896/jsjkx.210600042
[11] 张颖涛, 张杰, 张睿, 张文强.
全局信息引导的真实图像风格迁移
Photorealistic Style Transfer Guided by Global Information
计算机科学, 2022, 49(7): 100-105. https://doi.org/10.11896/jsjkx.210600036
[12] 曾志贤, 曹建军, 翁年凤, 蒋国权, 徐滨.
基于注意力机制的细粒度语义关联视频-文本跨模态实体分辨
Fine-grained Semantic Association Video-Text Cross-modal Entity Resolution Based on Attention Mechanism
计算机科学, 2022, 49(7): 106-112. https://doi.org/10.11896/jsjkx.210500224
[13] 徐鸣珂, 张帆.
Head Fusion:一种提高语音情绪识别的准确性和鲁棒性的方法
Head Fusion:A Method to Improve Accuracy and Robustness of Speech Emotion Recognition
计算机科学, 2022, 49(7): 132-141. https://doi.org/10.11896/jsjkx.210100085
[14] 孟月波, 穆思蓉, 刘光辉, 徐胜军, 韩九强.
基于向量注意力机制GoogLeNet-GMP的行人重识别方法
Person Re-identification Method Based on GoogLeNet-GMP Based on Vector Attention Mechanism
计算机科学, 2022, 49(7): 142-147. https://doi.org/10.11896/jsjkx.210600198
[15] 金方焱, 王秀利.
融合RACNN和BiLSTM的金融领域事件隐式因果关系抽取
Implicit Causality Extraction of Financial Events Integrating RACNN and BiLSTM
计算机科学, 2022, 49(7): 179-186. https://doi.org/10.11896/jsjkx.210500190
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!