Computer Science ›› 2022, Vol. 49 ›› Issue (12): 163-169.doi: 10.11896/jsjkx.211200080

• Database & Big Data & Data Science • Previous Articles     Next Articles

Disentangled Sequential Variational Autoencoder for Collaborative Filtering

WU Mei-lin1, HUANG Jia-jin2, QIN Jin1   

  1. 1 School of Computer Science and Technology,Guizhou University,Guiyang 550025,China
    2 International WIC Institute,Beijing University of Technology,Beijing 100000,China
  • Received:2021-12-06 Revised:2022-03-25 Published:2022-12-14
  • About author:WU Mei-lin,born in 1997,postgra-duate.His main research interests include recommendation system and so on.HUANG Jia-jin,born in 1977,Ph.D.His main research interests include recommendation system and so on.
  • Supported by:
    Guizhou Province Science and Technology Foundation([2020]1Y275).

Abstract: Recommendation models typically use user’s historical behaviors to obtain user preference representations for recommendations.Most of the methods of learning user representations always entangle different preference factors,while the disentangled learning method can be used to decompose user behavior characteristics.In this paper,a variational autoencoder based framework DSVAECF is proposed to disentangle the static and dynamic factors from user’s historical behaviors.Firstly,two encoders of the model use multi-layer perceptron and recurrent neural network to model the user behavior history respectively,so as to obtain the static and dynamic preference representation of the user.Then,the concatenate static and dynamic preference representations are treated as disentangled representation input decoders to capture user’s decisions and reconstruct user’s behavior.On the one hand,in the model training phase,DSVAECF learns model parameters by maximizes the mutual information between reconstructed user’s behaviors and actual user’s behaviors.On the other hand,DSVAECF minimizes the difference between disentangled representations and their prior distribution to retain the generation ability of the model.Experimental results on Amazon and MovieLens data sets show that,compared with the baselines,DSVAECF significantly improves the normalized discounted cumulative gain,recall,and precision,and has better recommendation performance.

Key words: Variational autoencoder, Deep learning, Sequence modeling, Disentangled learning, Collaborative filtering

CLC Number: 

  • TP391
[1]CHAE D K,KANG J S,KIM S W,et al.Cfgan:A generic collaborative filtering framework based on generative adversarial networks[C]//Proceedings of the 27th ACM International Conference on Information and Knowledge Management.2018:137-146.
[2]KOREN Y,BELL R,VOLINSKY C.Matrix factorization techniques for recommender systems[J].Computer,2009,42(8):30-37.
[3]ZHANG S,YAO L,SUN A,et al.Deep learning based recommender system:A survey and new perspectives[J].ACM Computing Surveys(CSUR),2019,52(1):1-38.
[4]KINGMA D P,WELLING M.Auto-Encoding Variational Bayes[J].arXiv:1312.6114,2013.
[5]REZENDE D J,MOHAMED S,WIERSTRA D.StochasticBackpropagation and Approximate Inference in Deep Generative Models[C]//Proceedings of the International Conference on Machine Learning.PMLR,2014:1278-1286.
[6]WANG S J,HU L,WANG Y,et al.Sequential Recommender Systems:Challenges,Progress and Prospects[C]//Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence{IJCAI-19}.2019.
[7]ZHANG Y,ZHU Z W,HE Y,et al.Content-Collaborative Disentangled representations Learning for Enhanced Recommendation[C]//Proceedings of the Fourteenth ACM Conference on Recommender Systems.2020:43-52.
[8]HE X N,LIAO L Z,ZHANG H W,et al.Neural Collaborative Filtering[C]//Proceedings of the 26th International Conference on World Wide Web.2017:173-182.
[9]HIDASI B,KARATZOGLOU A,BALTRUNAS L,et al.Session-based Recommendations with Recurrent Neural Networks[J].arXiv:1511.06939,2018.
[10]WU C Y,AHMED A,BEUTEL A,et al.Recurrent recommender networks[C]//Proceedings of the Tenth ACM International Conference on Web Search and Data Mining.2017:495-503.
[11]SEDHAIN S,MENON A K,SANNER S,et al.AutoRec:Autoencoders Meet Collaborative Filtering[C]//Proceedings of the 24th International Conference on World Wide Web.2015:111-112.
[12]SHENBIN I,ALEKSEEV A,TUTUBALINA E,et al.Rec-VAE:A new variational autoencoder for Top-N recommendations with implicit feedback[C]//Proceedings of the 13th International Conference on Web Search and Data Mining.2020:528-536.
[13]SACHDEVA N,MANCO G,RITACCO E,et al.Sequential va-riational autoencoders for collaborative filtering[C]//Procee-dings of the Twelfth ACM International Conference on Web Search and Data Mining.2019:600-608.
[14]LIANG D W,KRISHNAN R G,HOFFMAN M D,et al.Variational autoencoders for collaborative filtering[C]//Proceedings of the 2018 World wide Web Conference.2018:689-698.
[15]LI Y Z,MANDT S.Disentangled sequential autoencoder[C]//Proceedings of the 35th International Conference on Machine Learning.2018:5656-5665.
[16]HAN J,MIN M R,HAN L,et al.Disentangled recurrent wasser-stein autoencoder[C]//Proceedings of the 2021 International Conference on Learning Representations.2021.
[17]BAI J W,WANG W R,CARLA P G.Contrastively disentangled sequential variational autoencoder[C]//Proceedings of the 2021 Annual Conferenceon Neural Information Processing Systems.2021.
[18]HAN J,MIN M R,HAN L G,et al.Disentangled RecurrentWasserstein Autoencoder[C]//Proceedings of the 9th International Conference on Learning Representations.2020.
[19]BOWMAN S R,VILNIS L,VINYALS O,et al.Generating sentences from a continuous space[C]//Proceedings of the 20th SIGNLL Conference on Computational Natural Language Learning,CoNLL.Association for Computational Linguistics(ACL),2016:10-21.
[20]TOLSTIKHIN I,BOUSQUET O,GELLY S,et al.Wasserstein auto-encoders[C]//Proceedings of the 6th International Confe-rence on Learning Representations(ICLR 2018).OpenReview.net,2018.
[21]KABBUR S,NING X,KARYPIS G.FISM:Factored Item Similarity Models for top-N Recommender Systems[C]//Procee-dings of the 19th ACM SIGKDD International.2013:659-667.
[1] CHENG Zhang-tao, ZHONG Ting, ZHANG Sheng-ming, ZHOU Fan. Survey of Recommender Systems Based on Graph Learning [J]. Computer Science, 2022, 49(9): 1-13.
[2] WANG Guan-yu, ZHONG Ting, FENG Yu, ZHOU Fan. Collaborative Filtering Recommendation Method Based on Vector Quantization Coding [J]. Computer Science, 2022, 49(9): 48-54.
[3] XU Yong-xin, ZHAO Jun-feng, WANG Ya-sha, XIE Bing, YANG Kai. Temporal Knowledge Graph Representation Learning [J]. Computer Science, 2022, 49(9): 162-171.
[4] RAO Zhi-shuang, JIA Zhen, ZHANG Fan, LI Tian-rui. Key-Value Relational Memory Networks for Question Answering over Knowledge Graph [J]. Computer Science, 2022, 49(9): 202-207.
[5] TANG Ling-tao, WANG Di, ZHANG Lu-fei, LIU Sheng-yun. Federated Learning Scheme Based on Secure Multi-party Computation and Differential Privacy [J]. Computer Science, 2022, 49(9): 297-305.
[6] SUN Qi, JI Gen-lin, ZHANG Jie. Non-local Attention Based Generative Adversarial Network for Video Abnormal Event Detection [J]. Computer Science, 2022, 49(8): 172-177.
[7] WANG Jian, PENG Yu-qi, ZHAO Yu-fei, YANG Jian. Survey of Social Network Public Opinion Information Extraction Based on Deep Learning [J]. Computer Science, 2022, 49(8): 279-293.
[8] HAO Zhi-rong, CHEN Long, HUANG Jia-cheng. Class Discriminative Universal Adversarial Attack for Text Classification [J]. Computer Science, 2022, 49(8): 323-329.
[9] JIANG Meng-han, LI Shao-mei, ZHENG Hong-hao, ZHANG Jian-peng. Rumor Detection Model Based on Improved Position Embedding [J]. Computer Science, 2022, 49(8): 330-335.
[10] SU Dan-ning, CAO Gui-tao, WANG Yan-nan, WANG Hong, REN He. Survey of Deep Learning for Radar Emitter Identification Based on Small Sample [J]. Computer Science, 2022, 49(7): 226-235.
[11] SUN Xiao-han, ZHANG Li. Collaborative Filtering Recommendation Algorithm Based on Rating Region Subspace [J]. Computer Science, 2022, 49(7): 50-56.
[12] HU Yan-yu, ZHAO Long, DONG Xiang-jun. Two-stage Deep Feature Selection Extraction Algorithm for Cancer Classification [J]. Computer Science, 2022, 49(7): 73-78.
[13] CHENG Cheng, JIANG Ai-lian. Real-time Semantic Segmentation Method Based on Multi-path Feature Extraction [J]. Computer Science, 2022, 49(7): 120-126.
[14] HOU Yu-tao, ABULIZI Abudukelimu, ABUDUKELIMU Halidanmu. Advances in Chinese Pre-training Models [J]. Computer Science, 2022, 49(7): 148-163.
[15] ZHOU Hui, SHI Hao-chen, TU Yao-feng, HUANG Sheng-jun. Robust Deep Neural Network Learning Based on Active Sampling [J]. Computer Science, 2022, 49(7): 164-169.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!