Computer Science ›› 2020, Vol. 47 ›› Issue (8): 171-177.doi: 10.11896/jsjkx.190600150

Previous Articles     Next Articles

Incremental Classification Model Based on Q-learning Algorithm

LIU Ling-yun, QIAN Hui, XING Hong-jie, DONG Chun-ru, ZHANG Feng   

  1. Hebei Key Laboratory of Machine Learning and Computational Intelligence, College of Mathematics and Information Science, Hebei University, Baoding, Hebei 071002, China
  • Online:2020-08-15 Published:2020-08-10
  • About author:LIU Ling-yun, born in 1993, postgradua-te.Her main research interests include machine learning, group decision ma-king.
    ZHANG Feng, born in 1976, Ph.D, associate professor, is a member of China Computer Federation.Her main research interests include machine lear-ning, intelligent decision making.
  • Supported by:
    This work was supported by the National Natural Science Foundation of China (61672205), Natural Science Foundation of Hebei Province(F2018201115, F2017201020) and Youth Foundation of Hebei Education Department (QN2017019).

Abstract: The traditional classification models are insufficient to take full advantage of the sequential data with their continuous and explosive growth due to the imprecision of the data.Therefore, the incremental learning is provided to handle this problem.However, the difference sequence of the training samples may have strong impact on performance of a classifier.Especially when the classifier is undertrained, traditional incremental learning method takes the risk of utilizing the noise samples with wrong labels to train the classifier.To overcome this problem, this paper proposes an incremental classification model based on Q-learning algorithm.The model employs the classical Q-learning algorithm in reinforcement learning to select the sequence samples incrementally, which is capable of softening the negative impact of the noise data and labels samples automatically as well.To overcome the problem of computational complexity along with the increasing of state space and action space of Q-learning, an improved batch incremental classification model based on Q-learning algorithm is proposed.Compared with the traditionally trained classifiers, the proposed model combines the ideas of online incremental learning and reinforcement learning, which is able to achieve high accuracy and can be updated online.Finally, the validity of the model is verified on three UCI datasets.The experimental results show that choosing training sets incrementally is helpful to improve the performance of the classifier and the precision of the classifier trained by different incremental training sequences varies greatly as well.The proposed incremental classification model based on Q-learning algorithm can make use of the limited available dataset for supervised initial training, and then construct new-added self-supervised training set based on the Q value of each unlabeled sample to improve the accuracy of the classifier.Therefore, the incremental classification model based on Q-learning algorithm can be used to solve the problem of lack of supervisory information, and has a potential application.

Key words: Classification, Incremental learning, Online learning, Q-learning, Reinforcement learning

CLC Number: 

  • TP391
[1]KRIZHEVSKY A, SUTSKEVER I, HINTON G E.ImageNet Classification with Deep Convolutional Neural Networks[C]∥International Conference on Neural Information Processing Systems.2012:1097-1105.
[2]GOODFELIIOW I J, POUGET-ABADIE J, MIRZA M, et al.Generative Adversarial Networks[J].Advances in Neural Information Processing Systems, 2014, 3:2672-2680.
[3]XIAO R, WANG J C, SUN Z X, et al.An Incremental SVM Learning Algorithm α-ISVM[J].Journal of Software, 2001, 12(12):1818-1824.
[4]KIVINEN J, SMOLA A J, WILLIAMSON R C.Online Learning with Kernels[J].IEEE Transactions on Signal Processing, 2004, 52(8):2165-2176.
[5]GONG X J, LIU S H, SHI Z Z.An Incremental Bayes Classification Model[J].Chinese Journal of Computers, 2002, 25(6):645-650.
[6]RICHARD S, ANDREW B.Reinforcement Learning:An Introduction[M].Cambridge, MA:MIT Press, 1998.
[7]FOERSTER J, NARDELLI N, FARQUHAR G, et al.Stabili-sing Experience Replay for Deep Multi-agent Reinforcement Learning[J].arXiv:1702.08887v1.
[8]LI J.Incremental Learning and Its Applications to Image Recognition[D].Shanghai:Shanghai Jiao Tong University, 2008.
[9]COPPOCK H W, FREUND J E.All-or-none Versus Incremental Learning of Errorless Shock Ecapes by the Rat[J].Science, 1962, 135 (3500):318-319.
[10]SYED N, LIU H, SUNG K.Incremental Learning with Support Vetcor Machines[C]∥Proceedings of the Workshop on Support Vetcor Machines at the International Joint Conference on Artificial Intelligence.Stockholm:Morgan Kaufmann Publishers, 1999:876-892.
[11]ZENG W H, MA J.An Incremental Learning Algorithm forSupport Vector Machine and its Application[J].Computer Integrated Manufacturing System, 2003, 9(S1):144-148.
[12]ZHAO Y H, WANG K N, ZHONG P, et al.Incremental support vector machine based on border samples[J].Computer Engineering and Design, 2010(1):161-163.
[13]PI W J, GONG X J.Data driven parallel incremental support vector machine learning algorithm based on Hadoop framework[J].Journal of Computer Applications, 2016(11):3044-3049.
[14]VO M T.Incremental Learning Using the Time Delay NeuralNetwork[C]∥Proceedings of ICASSP’94.IEEE International Conference on Acoustics, Speech and Signal Processing.IEEE, 1994, 2(2):629-632.
[15]WANG Z.A Modified Neutral Network Increment Study Algorithm[J].Computer Science, 2007, 34(6):177-178.
[16]ZHAO C C.Research of Ensemble Incremental Learning Based on RBF[D].Tianjin:Hebei University of Technology, 2014.

[17]NAKAMURA Y, HASEGAWA O.Nonparametric Density Estimation Based on Self-Organizing Incremental Neural Network for Large Noisy Data[J].IEEE Transactions on Neural Networks & Learning Systems, 2017, 28(1):8-17.
[18]ZHANG Q X, ZHENG J J, NIU Z D, et al.Increment Learning Algorithm Based on Bayesian Classifier Integration[J].Transactions of Beijing Institute of Technology, 2008, 28(5):397-400.
[19]WEI Y, XU M, ZHENG Y.Incremental Learning Method ofBayesian Classification Combined with Feedback Information[J].Journal of Computer Applications, 2011, 1(9):643-648.
[20]SU Z T, LI Y.On Improved Incremental Bayesian Classification Model[J].Computer Applications and Software, 2016, 33(8):254-259.
[21]KOCHUROV M, GARIPOV T, PODOPRIKIN D, et al.Ba-yesian Incremental Learning for Deep Neural Networks[J].ar-Xiv:1802.07329, 2018.
[22]KAELBLING L P.Reinforcement Learning:A Survey[J].Journal of Artificial Intelligence Research, 1996, 4:237-285.
[23]WATKINS C J C H.Learning from Delayed Rewards[J].Robotics & Autonomous Systems, 1989, 15(4):233-235.
[24]WATKINS C J C H, DAYAN P.Technical Note:Q-Learning[J].Machine Learning, 1992, 8(3/4):279-292.
[25]MITCHELL T M.Machine Learning[M].Beijing:China Machine Press, 2014:270-271.
[26]SHANNON C E, WEAVER W.A Mathematical Theory ofCommunication[J].Bell Labs Technical Journal, 1948, 27(4):379-423.
[27]DEKE O, RAN G B, SHAMIR O, et al.Optimal Distributed Online Prediction Using Mini-batches[J].Journal of Machine Learning Research, 2012, 13(1):165-202.
[1] CHEN Zhi-qiang, HAN Meng, LI Mu-hang, WU Hong-xin, ZHANG Xi-long. Survey of Concept Drift Handling Methods in Data Streams [J]. Computer Science, 2022, 49(9): 14-32.
[2] ZHOU Xu, QIAN Sheng-sheng, LI Zhang-ming, FANG Quan, XU Chang-sheng. Dual Variational Multi-modal Attention Network for Incomplete Social Event Classification [J]. Computer Science, 2022, 49(9): 132-138.
[3] LIU Xing-guang, ZHOU Li, LIU Yan, ZHANG Xiao-ying, TAN Xiang, WEI Ji-bo. Construction and Distribution Method of REM Based on Edge Intelligence [J]. Computer Science, 2022, 49(9): 236-241.
[4] SHI Dian-xi, ZHAO Chen-ran, ZHANG Yao-wen, YANG Shao-wu, ZHANG Yong-jun. Adaptive Reward Method for End-to-End Cooperation Based on Multi-agent Reinforcement Learning [J]. Computer Science, 2022, 49(8): 247-256.
[5] HAO Zhi-rong, CHEN Long, HUANG Jia-cheng. Class Discriminative Universal Adversarial Attack for Text Classification [J]. Computer Science, 2022, 49(8): 323-329.
[6] WU Hong-xin, HAN Meng, CHEN Zhi-qiang, ZHANG Xi-long, LI Mu-hang. Survey of Multi-label Classification Based on Supervised and Semi-supervised Learning [J]. Computer Science, 2022, 49(8): 12-25.
[7] LIU Dong-mei, XU Yang, WU Ze-bin, LIU Qian, SONG Bin, WEI Zhi-hui. Incremental Object Detection Method Based on Border Distance Measurement [J]. Computer Science, 2022, 49(8): 136-142.
[8] YUAN Wei-lin, LUO Jun-ren, LU Li-na, CHEN Jia-xing, ZHANG Wan-peng, CHEN Jing. Methods in Adversarial Intelligent Game:A Holistic Comparative Analysis from Perspective of Game Theory and Reinforcement Learning [J]. Computer Science, 2022, 49(8): 191-204.
[9] TAN Ying-ying, WANG Jun-li, ZHANG Chao-bo. Review of Text Classification Methods Based on Graph Convolutional Network [J]. Computer Science, 2022, 49(8): 205-216.
[10] YAN Jia-dan, JIA Cai-yan. Text Classification Method Based on Information Fusion of Dual-graph Neural Network [J]. Computer Science, 2022, 49(8): 230-236.
[11] YU Bin, LI Xue-hua, PAN Chun-yu, LI Na. Edge-Cloud Collaborative Resource Allocation Algorithm Based on Deep Reinforcement Learning [J]. Computer Science, 2022, 49(7): 248-253.
[12] LI Meng-fei, MAO Ying-chi, TU Zi-jian, WANG Xuan, XU Shu-fang. Server-reliability Task Offloading Strategy Based on Deep Deterministic Policy Gradient [J]. Computer Science, 2022, 49(7): 271-279.
[13] GAO Zhen-zhuo, WANG Zhi-hai, LIU Hai-yang. Random Shapelet Forest Algorithm Embedded with Canonical Time Series Features [J]. Computer Science, 2022, 49(7): 40-49.
[14] YANG Bing-xin, GUO Yan-rong, HAO Shi-jie, Hong Ri-chang. Application of Graph Neural Network Based on Data Augmentation and Model Ensemble in Depression Recognition [J]. Computer Science, 2022, 49(7): 57-63.
[15] ZHANG Hong-bo, DONG Li-jia, PAN Yu-biao, HSIAO Tsung-chih, ZHANG Hui-zhen, DU Ji-xiang. Survey on Action Quality Assessment Methods in Video Understanding [J]. Computer Science, 2022, 49(7): 79-88.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!