Computer Science ›› 2015, Vol. 42 ›› Issue (9): 195-198.doi: 10.11896/j.issn.1002-137X.2015.09.037

Previous Articles     Next Articles

Efficient Algorithm for Large-scale Support Vector Machine

FENG Chang, LI Zi-da and LIAO Shi-zhong   

  • Online:2018-11-14 Published:2018-11-14

Abstract: The algorithm for solving large-scale support vector machine(SVM) needs large memory requirement and computation time.Therefore,large-scale SVMs are performed on computer clusters or supercomputers.An efficient algorithm for large-scale SVM was presented,which can be operated on a daily-life PC.First,the large-scale training examples were subsampled to reduce the data size.Then,the random Fourier mapping was explicitly applied to the subsample to generate the random feature space,making it possible to apply a linear SVM to uniformly approximate to the Gaussian kernel SVM.Finally,a parallelized linear SVM algorithm was implemented to speed up the training further.Experimental results on benchmark datasets demonstrate the feasibility and efficiency of the proposed algorithm.

Key words: Large-scale support vector machine,Subsampling,Random Fourier features,Parallelized linear SVM

[1] Vapnik V.The Nature of Statistical Learning Theory[M].New York:Springer Verlag,2000
[2] 文益民,王耀南,吕宝粮,等.支持向量机处理大规模问题算法综述[J].计算机科学,2009,36(7):20-25 Wen Yi-min,Wang Yao-nan,Lv Bao-liang,et al.Survey of Applying Support Vector Machines to Handle Large-scale Problems[J].Computer Science,2009,36(7):20-25
[3] Platt J C.Fast training of support vector machines using sequential minimal optimization[M]∥Schlkopf B,Burges C,Smola A.Advances in Kernel Methods:Support Vector Learning.Cambridge:MIT Press,1999:185-208
[4] Tsang L W,Kwok J,Cheung P M.Core vector machines:Fast SVM training on very large data sets[J].Journal of Machine Learning Research,2005,6(2):363-392
[5] Chang E Y,Zhu Kai-hua,Wang Hao,et al.Parallelizing support vector machines on distributed computers[M]∥ Platt J C,Koller D,Singer Y,et al.,eds.Advances in Neural Information Processing Systems 20.Cambridge:MIT Press,2008:257-264
[6] Zhang Tong.Solving large scale linear prediction problems using stochastic gradient descent algorithms[C]∥Proceedings of the 21st International Conference on Machine Learning,2004.New York,USA,2004:919-926
[7] Fan R E,Chang K W,Hsieh C J.LIBLINEAR:A library for large linear classification[J].Journal of Machine Learning Research,2008,9(12):1871-1874
[8] Rahimi A,Recht B.Random features for large-scale kernel machines[M]∥Platt J C,Koller D,Singer Y,et al.,eds.Advances in Neural Information Processing Systems 20.Cambridge:MIT Press,2008:1177-1184
[9] Boyd S,Parikh N,Chu E,et al.Distributed optimization and statistical learning via the alternating direction method of multip-liers[J].Foundations and Trends in Machine Learning,2011,3(1):1-122
[10] Chang Y W,Hsieh C J,Chang K W,et al.Training and testing low-degree polynomial data mappings via linear SVM[J].Journal of Machine Learning Research,2010,11(4):1471-1490
[11] Zhang K,Lan L,Wang Z,et al.Scaling up kernel SVM on limi-ted resources:A low-rank linearization approach[C]∥Procee-dings of the 15th International Conference on Artificial Intelligence and Statistics,2012.Canary,Spain,2012:1425-1434

No related articles found!
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!