Computer Science ›› 2022, Vol. 49 ›› Issue (3): 232-238.doi: 10.11896/jsjkx.210200153

• Artificial Intelligence • Previous Articles     Next Articles

Conversational Comprehension Model for Question Generation

SHI Yu-tao, SUN Xiao   

  1. School of Computer and Information,Hefei University of Technology,Hefei 230601,ChinaKey Laboratory of Affective Computing and Advanced Intelligent Machines of Anhui Province,Hefei University of Technology,Hefei 230601, China
  • Received:2021-02-24 Revised:2021-06-13 Online:2022-03-15 Published:2022-03-15
  • About author:SHI Yu-tao,born in 1997,postgraduate.His main research interests include na-tural language processing and machine learning.
    SUN Xiao,born in 1980,Ph.D,professor,is a member of China Computer Federation.His main research interests include affective computing,natural language processing,machine learning and human-machine interactions.
  • Supported by:
    National Natural Science Foundation of China(61976078).

Abstract: Conversational question generation (CQG) is different from the question generation task of generating single-round questions based on paragraphs and answers.CQG additionally considers the conversational information composed of historical question and answer pairs,and the generated questions inherit the historical content of the conversation and maintain high consistency.In response to this feature,the article proposes word-level and sentence-level attention mechanism modules to enhance the ability to extract conversation history information,ensuring that the current round of questions integrates the characteristics of each word and sentence in the conversation history,thereby generating a coherent,high-quality question.The accuracy of the question word is more important.The generated question needs to match the answer type corresponding to the original question in the data set.An additional loss function is constructed in the question word prediction module as a limitation of the question word type.The conversational comprehension network (CCNet) model is obtained by synthesizing each module.Experiments show that this model is higher than the baseline model in most evaluation indicators.On the CoQA dataset,Bleu1 and Bleu2 reach 39.70 and 23.76,respectively,and the quality of the generated questions is higher.The model is proved to be effective in ablation experiments and cross-dataset experiments,indicating that the CCNet model has strong general capabilities.

Key words: Attention mechanism, Conversational question generation, Gated network, Question generation, Recurrent neural network

CLC Number: 

  • TP391
[1]SERBAN I V,GARCIA-DURAN A,GULCEHRE C,et al.Ge-nerating Factoid Questions With Recurrent Neural Networks:The 30M Factoid Question-Answer Corpus[C]//Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics.2016:588-598.
[2]GUO D,SUN Y,TANG D,et al.Question Generation fromSQL Queries Improves Neural Semantic Parsing[C]//Procee-dings of the 2018 Conference on Empirical Methods in Natural Language Processing.2018:1597-1607.
[3]DU X,SHAO J,CARDIE C.Learning to Ask:Neural Question Generation for Reading Comprehension[C]//Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics.2017:1342-1352.
[4]DU X,CARDIE C.Harvesting Paragraph-level Question-An-swer Pairs from Wikipedia[C]//Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics.2018:1907-1917.
[5]WANG J,LIU J,BI W,et al.Dual Dynamic Memory Network for End-to-End Multi-turn Task-oriented Dialog Systems[C]//Proceedings of the 28th International Conference on Computational Linguistics.2020:4100-4110.
[6]GAO Y,LI P,KING I,et al.Interconnected Question Generation with Coreference Alignment and Conversation Flow Mode-ling [C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics.2019:4853-4862.
[7]REDDY S,CHEN D,MANNINGC D.CoQA:A Conversational Question Answering Challenge[J].Transactions of the Association for Computational Linguistics,2019,7:249-266.
[8]CHOI E,HE H,IYYER M,et al.QuAC:Question Answering in Context[C]//Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing.2018:2174-2184.
[9]PAN B,LI H,YAO Z,et al.Reinforced Dynamic Reasoning for Conversational Question Generation[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics.2019:2114-2124.
[10]VANDERWENDE L.Answering and Questioning for Machine Reading[C]//AAAI Spring Symposium:Machine Reading.2007:91.
[11]HEILMAN M,SMITH N A.Good question! statistical ranking for question generation[C]//Human Language Technologies:The 2010 Annual Conference of the North American Chapter of the Association for Computational Linguistics.2010:609-617.
[12]SUTSKEVER I,VINYALS O,LEQ V.Sequence to sequencelearning with neural networks[C]//Proceedings of the 27th International Conference on Neural Information Processing Systems.2014:3104-3112.
[13]LUONG M T,PHAM H,MANNINGC D.Effective Approaches to Attention-based Neural Machine Translation[C]//Procee-dings of the 2015 Conference on Empirical Methods in Natural Language Processing.2015:1412-1421.
[14]ZHOU Q,YANG N,WEI F,et al.Neural question generation from text:A preliminary study[C]//National CCF Conference on Natural Language Processing and Chinese Computing.Cham:Springer,2017:662-671.
[15]SUBRAMANIAN S,WANG T,YUAN X,et al.Neural Models for Key Phrase Extraction and Question Generation[C]//Proceedings of the Workshop on Machine Reading for Question Answering.2018:78-88.
[16]DUAN N,TANG D,CHEN P,et al.Question generation for question answering[C]//Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing.2017:866-874.
[17]SERBAN I V,SORDONI A,BENGIO Y,et al.Building end-to-end dialogue systems using generative hierarchical neural network models[C]//Proceedings of the Thirtieth AAAI Confe-rence on Artificial Intelligence.2016:3776-3783.
[18]XING C,WU Y,WU W,et al.Hierarchical recurrent attentionnetwork for response generation[C]//Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence and Thirtieth Innovative Applications of Artificial Intelligence Conference and Eighth AAAI Symposium on Educational Advances in Artificial Intelligence.2018:5610-5617.
[19]HOCHREITER S,SCHMIDHUBER J.Long short-term memory[J].Neural Computation,1997,9(8):1735-1780.
[20]XIONG C,ZHONG V,SOCHER R.Dynamic coattention networks for question answering [J].arXiv:1611.01604,2016.
[21]SEO M,KEMBHAVI A,FARHADI A,et al.Bidirectional attention flow for machine comprehension[J].arXiv:1611.01603,2016.
[22]SEE A,LIU P J,MANNINGC D.Get To The Point:Summarization with Pointer-Generator Networks[C]//Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics.2017:1073-1083.
[23]PENNINGTON J,SOCHER R,MANNING C D.Glove:Global vectors for word representation[C]//Proceedings of the 2014 Conference on Empirical Methods in Natural Language Proces-sing (EMNLP).2014:1532-1543.
[24]PAPINENI K,ROUKOS S,WARD T,et al.Bleu:a method for automatic evaluation of machine translation[C]//Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics.2002:311-318.
[25]BANERJEE S,LAVIE A.METEOR:An automatic metric for MT evaluation with improved correlation with human judgments[C]//Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization.2005:65-72.
[26]LIN C Y.Rouge:A package for automatic evaluation of summaries[C]//Text Summarization Branches Out.2004:74-81.
[1] ZHOU Fang-quan, CHENG Wei-qing. Sequence Recommendation Based on Global Enhanced Graph Neural Network [J]. Computer Science, 2022, 49(9): 55-63.
[2] DAI Yu, XU Lin-feng. Cross-image Text Reading Method Based on Text Line Matching [J]. Computer Science, 2022, 49(9): 139-145.
[3] ZHOU Le-yuan, ZHANG Jian-hua, YUAN Tian-tian, CHEN Sheng-yong. Sequence-to-Sequence Chinese Continuous Sign Language Recognition and Translation with Multi- layer Attention Mechanism Fusion [J]. Computer Science, 2022, 49(9): 155-161.
[4] XIONG Li-qin, CAO Lei, LAI Jun, CHEN Xi-liang. Overview of Multi-agent Deep Reinforcement Learning Based on Value Factorization [J]. Computer Science, 2022, 49(9): 172-182.
[5] RAO Zhi-shuang, JIA Zhen, ZHANG Fan, LI Tian-rui. Key-Value Relational Memory Networks for Question Answering over Knowledge Graph [J]. Computer Science, 2022, 49(9): 202-207.
[6] WANG Ming, PENG Jian, HUANG Fei-hu. Multi-time Scale Spatial-Temporal Graph Neural Network for Traffic Flow Prediction [J]. Computer Science, 2022, 49(8): 40-48.
[7] JIANG Meng-han, LI Shao-mei, ZHENG Hong-hao, ZHANG Jian-peng. Rumor Detection Model Based on Improved Position Embedding [J]. Computer Science, 2022, 49(8): 330-335.
[8] ZHU Cheng-zhang, HUANG Jia-er, XIAO Ya-long, WANG Han, ZOU Bei-ji. Deep Hash Retrieval Algorithm for Medical Images Based on Attention Mechanism [J]. Computer Science, 2022, 49(8): 113-119.
[9] SUN Qi, JI Gen-lin, ZHANG Jie. Non-local Attention Based Generative Adversarial Network for Video Abnormal Event Detection [J]. Computer Science, 2022, 49(8): 172-177.
[10] YAN Jia-dan, JIA Cai-yan. Text Classification Method Based on Information Fusion of Dual-graph Neural Network [J]. Computer Science, 2022, 49(8): 230-236.
[11] ZHANG Ying-tao, ZHANG Jie, ZHANG Rui, ZHANG Wen-qiang. Photorealistic Style Transfer Guided by Global Information [J]. Computer Science, 2022, 49(7): 100-105.
[12] ZENG Zhi-xian, CAO Jian-jun, WENG Nian-feng, JIANG Guo-quan, XU Bin. Fine-grained Semantic Association Video-Text Cross-modal Entity Resolution Based on Attention Mechanism [J]. Computer Science, 2022, 49(7): 106-112.
[13] XU Ming-ke, ZHANG Fan. Head Fusion:A Method to Improve Accuracy and Robustness of Speech Emotion Recognition [J]. Computer Science, 2022, 49(7): 132-141.
[14] MENG Yue-bo, MU Si-rong, LIU Guang-hui, XU Sheng-jun, HAN Jiu-qiang. Person Re-identification Method Based on GoogLeNet-GMP Based on Vector Attention Mechanism [J]. Computer Science, 2022, 49(7): 142-147.
[15] JIN Fang-yan, WANG Xiu-li. Implicit Causality Extraction of Financial Events Integrating RACNN and BiLSTM [J]. Computer Science, 2022, 49(7): 179-186.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!