Computer Science ›› 2021, Vol. 48 ›› Issue (6A): 206-212.doi: 10.11896/jsjkx.200900196

• Big Data & Data Science • Previous Articles     Next Articles

Query Suggestion Method Based on Autoencoder and Reinforcement Learning

HU Xiao-wei, CHEN Yu-zhong   

  1. College of Mathematics and Computer Science,Fuzhou University,Fuzhou 350116,China
    Fujian Provincial Key Laboratory of Network Computing and Intelligent Information Processing,Fuzhou 350116,China
  • Online:2021-06-10 Published:2021-06-17
  • About author:HU Xiao-wei,born in 1996,postgra-duate.His main research interests include query suggestion and machine reading comprehension.
    CHEN Yu-zhong,Ph.D,professor.His main research interests include natural language processing and data mining.
  • Supported by:
    National Natural Science Foundation of China(61672158,61672159,61502104,61502105),Industry-Academy Cooperation Project(2018H6010) and Natural Science Foundation of Fujian Province,China(201801795).

Abstract: The purpose of query suggestion is to explore the query intent of search engine users and provide relevant query suggestion.Traditional query suggestion methods mainly rely on manually extracting relevant features of queries,such as query frequency,query time,user clicks and dwell time,etc.,and use statistical learning algorithms or ranking algorithms to give query suggestion.In recent years,deep learning methods have been widely used in query suggestion problems.The existing deep learning methods for query recommendation are mostly based on recurrent neural networks,which predict the next query of the user by modeling the semantic features of all queries in the query log.However,the existing deep learning methods have poor context awareness of query suggestion,it is difficult to accurately capture user query intentions,and the influence of time factors on query suggestion is not fully considered,and it lacks timeliness and diversity.In response to the above problems,this paper proposes a query suggestion model combining autoencoder and reinforcement learning(Latent Variable Hierarchical Recurrent Encoder-Decoder with Time Information of Query and Reinforcement Learning,VHREDT-RL).VHREDT-RL introduces a reinforcement learning joint training generator and discriminator,thereby enhancing the context awareness of generating query suggestion,using latent variable hierarchical recursive autoencoders that integrate query time information as a generator,and making query suggestion better time-sensitive and diversity.The experimental results on the AOL data set show that the VHREDT-RL model proposed in this paper achieves better accuracy,robustness and stability than the benchmark method.

Key words: Query intention, Query suggestion, Reinforcement learning, Time information, Variable hierarchical recursive autoencoder

CLC Number: 

  • TP391
[1] HUANG C K,CHIEN L F,OYANG Y J.Relevant term suggestion in interactive web search based on contextual information in query session logs[J].Journal of the American Society for Information Science and Technology,2003,54(7):638-649.
[2] CHEN W,CAI F,CHEN H,et al.Personalized query suggestion diversification[C]//Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval.2017:817-820.
[3] CHO K,VAN MERRINBOER B,GULCEHRE C,et al.Learning phrase representations using RNN encoder-decoder for statistical machine translation[J].arXiv:1406.1078,2014.
[4] OZERTEM U,CHAPELLE O,DONMEZ P,et al.Learning to suggest:a machine learning framework for ranking query suggestions[C]//Proceedings of the 35th International ACM SIGIR Conference on Research and Development in Information Retrieval.2012:25-34.
[5] SANTOS R L T,MACDONALD C,OUNIS I.Learning to rank query suggestions for adhoc and diversity search[J].Information Retrieval,2013,16(4):429-451.
[6] KOMBRINK S,MIKOLOV T,KARAFIÁT M,et al.Recurrent neural network based language modeling in meeting recognition[C]//Twelfth Annual Conference of the International Speech Communication Association.2011.
[7] MIKOLOV T,KOMBRINK S,BURGET L,et al.Extensions of recurrent neural network language model[C]//2011 IEEE International Conference on Acoustics,Speech and Signal Proces-sing (ICASSP).IEEE,2011:5528-5531.
[8] SUTSKEVER I,MARTENS J,HINTON G E.Generating text with recurrent neural networks[C]//ICML.2011.
[9] SHUANG C,YI D.Structure of recurrent neural networks[J].Computer Applications,2004,24(8):18-20.
[10] YUAN L,MENG Y.Combination prediction model of network traffic based on recurrent neural networks[J].Computer Engineering and Design,2008(3):700-702.
[11] JIAN Z,DAN Q,ZHEN L.Recurrent neural network language model based on work vector features[J].Pattern Recognition and Artificial Intelligence,2015,28(4):299-305.
[12] ONAL K D,ZHANG Y,ALTINGOVDE I S,et al.Neural information retrieval:At the end of the early years[J].Information Retrieval Journal,2018,21(2/3):111-182.
[13] SERBAN I V,SORDONI A,BENGIO Y,et al.Building end-to-end dialogue systems using generative hierarchical neural network models[J].arXiv:1507.04808,2015.
[14] CAI F,DE RIJKE M.A Survey of Query Auto Completion in Information Retrieval[J].Foundations and Trends in Information Retrieval,2016,10(4):273-363.
[15] SERBAN I V,SORDONI A,LOWE R,et al.A hierarchical latent variable encoder-decoder model for generating dialogues[J].arXiv:1605.06069,2016.
[16] BAEZA-YATES R,HURTADO C,MENDOZA M.Query recommendation using query logs in search engines[C]//International Conference on Extending Database Technology.Berlin:Springer,2004:588-596.
[17] CAO H,JIANG D,PEI J,et al.Context-aware query suggestion by mining click-through and session data[C]//Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining.2008:875-883.
[18] BARAGLIA R,NARDINI F M,CASTILLO C,et al.The effects of time on query flow graph-based models for query suggestion[M]//Adaptivity,Personalization and Fusion of Heterogeneous Information.2010:182-189.
[19] WANG J G,HUANG J Z,GUO J,et al.Recommending high-utility search engine queries via a query-recommending model[J].Neurocomputing,2015,167:195-208.
[20] CHEN W,HAO Z,SHAO T,et al.Personalized query suggestion based on user behavior[J].International Journal of Modern Physics C,2018,29(4):1850036.
[21] SORDONI A,BENGIO Y,VAHABI H,et al.A hierarchical recurrent encoder-decoder for generative context-aware query suggestion[C]//Proceedings of the 24th ACM International on Conference on Information and Knowledge Management.2015:553-562.
[22] SONG J,XIAO J,WU F,et al.Hierarchical contextual attention recurrent neural network for map query suggestion[J].IEEE Transactions on Knowledge and Data Engineering,2017,29(9):1888-1901.
[23] CHEN W,CAI F,CHEN H,et al.Attention-based hierarchical neural query suggestion[C]//The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval.2018:1093-1096.
[24] WU B,XIONG C,SUN M,et al.Query suggestion with feedback memory network[C]//Proceedings of the 2018 World Wide Web Conference.2018:1563-1571.
[25] JIANG J Y,WANG W.RIN:reformulation inference network for context-aware query suggestion[C]//Proceedings of the 27th ACM International Conference on Information and Know-ledge Management.2018:197-206.
[26] BENGIO Y.Learning deep architectures for AI[M].Now Publishers Inc,2009.
[27] LILLICRAP T P,HUNT J J,PRITZEL A,et al.Continuouscontrol with deep reinforcement learning[J].arXiv:1509.02971,2015.
[28] SILVER D,HUANG A,MADDISON C J,et al.Mastering the game of Go with deep neural networks and tree search[J].Nature,2016,529(7587):484-489.
[29] MNIH V,KAVUKCUOGLU K,SILVER D,et al.Playing atariwith deep reinforcement learning[J].arXiv:1312.5602,2013.
[30] YU L,ZHANG W,WANG J,et al.Seqgan:Sequence generative adversarial nets with policy gradient[C]//Thirty-first AAAI Conference on Artificial Intelligence.2017.
[31] LI J,MONROE W,SHI T,et al.Adversarial learning for neural dialogue generation[J].arXiv:1701.06547,2017.
[32] SUTTON R S,BARTO A G.Reinforcement learning:An introduction[M].MIT press,2018.
[1] LIU Xing-guang, ZHOU Li, LIU Yan, ZHANG Xiao-ying, TAN Xiang, WEI Ji-bo. Construction and Distribution Method of REM Based on Edge Intelligence [J]. Computer Science, 2022, 49(9): 236-241.
[2] YUAN Wei-lin, LUO Jun-ren, LU Li-na, CHEN Jia-xing, ZHANG Wan-peng, CHEN Jing. Methods in Adversarial Intelligent Game:A Holistic Comparative Analysis from Perspective of Game Theory and Reinforcement Learning [J]. Computer Science, 2022, 49(8): 191-204.
[3] SHI Dian-xi, ZHAO Chen-ran, ZHANG Yao-wen, YANG Shao-wu, ZHANG Yong-jun. Adaptive Reward Method for End-to-End Cooperation Based on Multi-agent Reinforcement Learning [J]. Computer Science, 2022, 49(8): 247-256.
[4] YU Bin, LI Xue-hua, PAN Chun-yu, LI Na. Edge-Cloud Collaborative Resource Allocation Algorithm Based on Deep Reinforcement Learning [J]. Computer Science, 2022, 49(7): 248-253.
[5] LI Meng-fei, MAO Ying-chi, TU Zi-jian, WANG Xuan, XU Shu-fang. Server-reliability Task Offloading Strategy Based on Deep Deterministic Policy Gradient [J]. Computer Science, 2022, 49(7): 271-279.
[6] XIE Wan-cheng, LI Bin, DAI Yue-yue. PPO Based Task Offloading Scheme in Aerial Reconfigurable Intelligent Surface-assisted Edge Computing [J]. Computer Science, 2022, 49(6): 3-11.
[7] HONG Zhi-li, LAI Jun, CAO Lei, CHEN Xi-liang, XU Zhi-xiong. Study on Intelligent Recommendation Method of Dueling Network Reinforcement Learning Based on Regret Exploration [J]. Computer Science, 2022, 49(6): 149-157.
[8] GUO Yu-xin, CHEN Xiu-hong. Automatic Summarization Model Combining BERT Word Embedding Representation and Topic Information Enhancement [J]. Computer Science, 2022, 49(6): 313-318.
[9] FAN Jing-yu, LIU Quan. Off-policy Maximum Entropy Deep Reinforcement Learning Algorithm Based on RandomlyWeighted Triple Q -Learning [J]. Computer Science, 2022, 49(6): 335-341.
[10] ZHANG Jia-neng, LI Hui, WU Hao-lin, WANG Zhuang. Exploration and Exploitation Balanced Experience Replay [J]. Computer Science, 2022, 49(5): 179-185.
[11] LI Peng, YI Xiu-wen, QI De-kang, DUAN Zhe-wen, LI Tian-rui. Heating Strategy Optimization Method Based on Deep Learning [J]. Computer Science, 2022, 49(4): 263-268.
[12] OUYANG Zhuo, ZHOU Si-yuan, LYU Yong, TAN Guo-ping, ZHANG Yue, XIANG Liang-liang. DRL-based Vehicle Control Strategy for Signal-free Intersections [J]. Computer Science, 2022, 49(3): 46-51.
[13] ZHOU Qin, LUO Fei, DING Wei-chao, GU Chun-hua, ZHENG Shuai. Double Speedy Q-Learning Based on Successive Over Relaxation [J]. Computer Science, 2022, 49(3): 239-245.
[14] LI Su, SONG Bao-yan, LI Dong, WANG Jun-lu. Composite Blockchain Associated Event Tracing Method for Financial Activities [J]. Computer Science, 2022, 49(3): 346-353.
[15] HUANG Xin-quan, LIU Ai-jun, LIANG Xiao-hu, WANG Heng. Load-balanced Geographic Routing Protocol in Aerial Sensor Network [J]. Computer Science, 2022, 49(2): 342-352.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!