Computer Science ›› 2017, Vol. 44 ›› Issue (Z6): 29-32.doi: 10.11896/j.issn.1002-137X.2017.6A.006

Previous Articles     Next Articles

Survey for Methods of Parameter Estimation in Topic Models

DU Hui, CHEN Yun-fang and ZHANG Wei   

  • Online:2017-12-01 Published:2018-12-01

Abstract: Topic models extract low-dimensional representation of the topic from high-dimensional sparse data set of word by using fast machine learning algorithms,achieving a word document clustering.It is an important work in this field to study the model parameter estimation.The paper detailed the probabilitic latent semantic analysis model,the latent Dirichlet model and basic methods of parameter estimation in topic model.In addition,the paper gave an experimental analysis of perplexity in topic model.

Key words: Topic models,Probabilitic latent semantic analysis,Latent Dirichlet allocation,Pa rameter estimation

[1] DEERWESTER S C,DUMAIS S T,LANDAUER T K,et al.Indexing by latent semantic analysis[J].Journal of the American Society for Information Science,1990,41(6):391-407.
[2] HOFMANN T.Probabilistic latent semantic indexing[C]∥Proceedings of the 22nd Annual International SIGIR Conference.New York:ACM Press,1999:50-57.
[3] BLEI D,NG A,JORDAN M.Latent Dirichlet allocation[J].Journal of Machine Learning Research,2003,3:993-1022.
[4] 徐戈,王厚峰.自然语言处理中主题模型的发展[J].计算机学报,2011(8):1423-1436.
[5] DEMPSTER A P,LAIRD N M,RUBIN D B.Maximum likelihood from incomplete data via the EM algorithm[J].Journal of the Royal Statistical Society,1997,B39(1):1-38.
[6] GRIFFITHS T L,STEYVERS M.Finding scientific topics [J].Proceedings of the National academy of Sciences of the United States of America,2004,101(Suppl 1):5228-5235.
[7] BLEI D M,NG A Y,JORDAN M I.Latent dirichlet allocation [J].The Journal of Machine Learning Research,2003,3:993-1022.
[8] ZEHG J,CHEUNG W K W,Liu J M.Learning topic models by belief propagation [C]∥IEEE Transactions on Pattern Analysis and Machine Intelligence.2012.
[9] 赵鑫,李晓明.主题模型在文本挖掘中的应用: PKU-CS-NCIS-TR2011XX [R].Technical Report,2011.
[10] HEINRICH G.Parameter estimation for text analysis[R].Tech-nical Report,2008.
[11] BLEI D M,LAFFERTY J D.Correlated topic models[C]∥Advances in Neural Information Processing Systems 18.Cambridge,MA:MIT Press,2006.
[12] BLEI D,LAFFERTY J.Dynamic topic models[C]∥Proceedings of the 23rd International Conference on Machine Learning.Pittsburgh,Pennsylvania,USA,2006:113-120.
[13] ROSEN-ZVI M,GRIFFITHS T,STEYVERS M,et al.The Auther-Topic Model for Author-Topic Model for Authors and Do-cuments[C]∥Proc.20th Conf.Uncertainty in Artificial Intelligence.2004:487-494.
[14] BISHOP C M.Pattern Recognition and Machine Learning[M].New York,USA:Springer,2006.
[15] 赖裕平,周亚建,丁洪伟,等.混合逆狄利克雷分布的变分学习及应用[J].电子学报,2014,42(7):1435-1440.
[16] PORTEOUS I,NNWMAN D,IHLER A,et al.Fast collapsed gibbs sampling for latent dirichlet allocation[C]∥Proceeding of The 14th ACM SIGKDD International Conference on Know-ledge Discovery and Data Mining(KDD’08).New York,NY,USA,2008:569-577.
[17] HOFFMAN M,BLEI D M,BACH F.Online learning for latent dirichlet allocation[C]∥NIPS.2010:856-864.
[18] YAO L M,MIMNO D,MCCALLUM A.Efficient methods for topic model inference on streaming document collections[C]∥Proceedings of the 15th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining(KDD’09).New York,NY,USA,ACM,2009:937-946.
[19] BANERJEE A,BASU S.Topic Models over Text Streams:A Study of Batch and Online Unsupervised Learning[C]∥SDM.SIAM,2007.
[20] NEWMAN D,ASUNCION A,SMYTH P,et al.Distributed Inference for Latent Dirichlet Allocation[C]∥Conference on Neural Information Processing Systems.2007:1-6.
[21] LIU Z Y,ZHANG Y Z,CHANG E Y,et al.Plda+:Parallel latent dirichlet allocation with data placement and pipeline processing[J].ACM Trans.Intell.Syst.Tschnol.,2011,2(3):26
[22] YAN F,XU N Y,QI Y.Parallel inference for latent dirichlet allocation on graphics processing units[C]∥NIPS.2009:2134-2142.
[23] ASUNCION A,SMYTH P,WELLING M.Asynchronous distributed learning of topic models[C]∥NIPS.2008:81-88.

No related articles found!
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!