Computer Science ›› 2019, Vol. 46 ›› Issue (5): 290-297.doi: 10.11896/j.issn.1002-137X.2019.05.045

Previous Articles     Next Articles

Improved Learning Model for Cloud Computing Swarm Optimization Time Efficiency

JIAN Cheng-feng, KUANG Xiang, ZHANG Mei-yu   

  1. (Computer Science and Technology College,Zhejiang University of Technology,Hangzhou 310023,China)
  • Published:2019-05-15

Abstract: Aiming at the time-consuming problem when the traditional task scheduling models of cloud computing deal with the tasks,this paper proposed an ONBA algorithm combining DE (Differential Evolution) to get the scheduling data of task.Then,the obtained scheduling data are used to train the improved IDBN model.By adjusting the learning rate and training times,the time efficiency can be improved,thus achieving fast and accurate prediction of cloud computing scheduling results.The experimental results show that the improved IDBN model trained by this method can effectively shorten the actual scheduling time on the premise of ensuring precise prediction results and make up for the defect of long running time in traditional swarm optimization models.

Key words: Cloud computing, Deep learning, Learning rate, Scheduling forecast

CLC Number: 

  • TP391
[1]MELL P,GRANCE T.The NIST definition of cloud computing [J].Communications of the Acm,2011,53(6):50.
[2]DASHTI S E,RAHMANI A M.Dynamic VMs placement for energy efficiency by PSO in cloud computing [J].Journal of Experimental & Theoretical Artificial Intelligence,2016,28(1-2):97-112.
[3]XIAO Z,JIANG J,ZHU Y,et al.A solution of dynamic VMsplacement problem for energy consumption optimization based on evolutionary game theory [J].Journal of Systems & Software,2015,101(C):260-272.
[4]LU Y,SUN N.An effective task scheduling algorithm based on dynamic energy management and efficient resource utilization in green cloud computing environment [J].Cluster Computing,DOI:10.1007/s10586-017-1272-y.
[5]KATYAL M,MISHRA A.A Comparative Study of Load Balancing Algorithms in Cloud Computing Environment [J].Compu-ter Science,2014,6(1):25-36.
[6]XU G,PANG J,FU X.A load balancing model based on cloudpartitioning for the public cloud [J].Tsinghua Science and Technology,2013,18(1):34-39.
[7]ZHAO J,YANG K,WEI X,et al.A heuristic clustering-based task deployment approach for load balancing using bayes theorem in cloud environment [J].IEEE Transactions on Parallel and Distributed Systems,2016,27(2):305-316.
[8]GUO L,ZHAO S,SHEN S,et al.Task scheduling optimization in cloud computing based on heuristic algorithm[J].JNW,2012,7(3):547-553.
[9]BHOI U,RAMANUJ P N.Enhanced max-min task scheduling algorithm in cloud computing [J].International Journal of Application or Innovation in Engineering and Management (IJAIEM),2013,2(4):259-264.
[10]ZHANG Z Y,MENG F C,CHU D H,et al.Dynamic Scheduling Algorithm of Service Request Based on Aware QoS Revenue [J].Mini-micro Systems,2017(12):2791-2796.(in Chinese)张中一,孟凡超,初佃辉,等.一种QoS收益感知的服务请求动态调度算法[J].小型微型计算机系统,2017(12):2791-2796.
[11]WANG W J,CHANG Y S,LO W T,et al.Adaptive scheduling for parallel tasks with QoS satisfaction for hybrid cloud environments[J].The Journal of Supercomputing,2013,66(2):783-811.
[12]ABDULLAH M,OTHMAN M.Cost-based Multi-QoS JobScheduling Using Divisible Load Theory in Cloud Computing [J].Procedia Computer Science,2013,18(1):928-935.
[13]RAMEZANI F,LU J,HUSSAIN F K.Task-Based System Load Balancing in Cloud Computing Using Particle Swarm Optimization [J].International Journal of Parallel Programming,2014,42(5):739-754.
[14]MIRJALILI S,HASHIM S Z M,SARDROUDI H M.Training feedforward neural networks using hybrid particle swarm optimization and gravitational search algorithm.Applied Mathematics and Computation,2012,218(22):11125-11137.
[15]YU Q,CHEN L,LI B.Ant colony optimization applied to web service compositions in cloud computing [J].Computers & Electrical Engineering,2015,41(C):18-27.
[16]MOON Y J,YU H C,GIL J M,et al.A slave ants based ant co-lony optimization algorithm for task scheduling in cloud computing environments [J].Human-centric Computing and Information Sciences,2017,7(1):28.
[17]DASGUPTA K,MANDAL B,DUTTA P,et al.A Genetic Algorithm (GA) based Load Balancing Strategy for Cloud Computing [J].Procedia Technology,2013,10(2):340-347.
[18]DEAN J,CORRADO G S,MONGA R,et al.Large scale distri-buted deep networks[C]∥International Conference on Neural Information Processing Systems.Curran Associates Inc,2012:1223-1231.
[19]FENG N,RECHT B,RE C,et al.HOGWILD!:A Lock-FreeApproach to Parallelizing Stochastic Gradient Descent[J].Advances in Neural Information Processing Systems,2011,24:693-701.
[20]MARTENS J,SUTSKEVER I.Learning Recurrent Neural Networks with Hessian-Free Optimization[C]∥International Conference on Machine Learning(ICML 2011).Bellevue,Washington,USA.DBLP,2011:1033-1040.
[21]MARTENS J.Deep learning via Hessian-free optimization[C]∥International Conference on Machine Learning.DBLP,2010:735-742.
[22]NGIAM J,COATES A,LAHIRI A,et al.On optimizationmethods for deep learning[C]∥Proceedings of the 28th International Conference on Machine Learning (ICML-11).2011:265-272.
[23]HINTON G E,SALAKHUTDINOV R R.Reducing the dimensionality of data with neural networks[J].Science,2006,313(5786):504-507.
[24]DUCHI J,HAZAN E,SINGER Y.Adaptive Subgradient Me-thods for Online Learning and Stochastic Optimization [J].Journal of Machine Learning Research,2011,12(7):257-269.
[25]SENIOR A,HEIGOLD G,RANZATO M,et al.An empirical study of learning rates in deep neural networks for speech recognition [C]∥2013 IEEE International Conference on Acoustics,Speech and Signal Processing.Vancouver,BC,Canada,2013.
[26]CHO K H,RAIKO T,ILIN A,et al.Enhanced Gradient and Adaptive Learning Rate for Training Restricted Boltzmann Machines[C]∥International Conference on Machine Learning(ICML 2011).Bellevue,Washington,USA,2011:105-112.
[27]TSAI C H,YU W J,WONG W H,et al.A 41.3/26.7 pJ per Neuron Weight RBM Processor Supporting On-Chip Learning/Inference for IoT Applications [J].IEEE Journal of Solid-State Circuits,2017,52(10):2601-2612.
[28]LOPES N,RIBEIRO B.Towards adaptive learning with im-proved convergence of deep belief networks on graphics proces-sing units [J].Pattern Recognition,2014,47(1):114-127.
[29]KIM L W.DeepX:Deep Learning Accelerator for Restricted Boltzmann Machine Artificial Neural Networks [J].IEEE Tran-sactions on Neural Networks & Learning Systems,2017,PP(99):1-13.
[30]CHEN C,ZHANG C,CHEN L,et al.Fuzzy Restricted Boltz-mann Machine and Its Fast Learning Algorithm [J].IEEE Transactions on Fuzzy Systems,2015,23(6):2163-2173.
[31]MENG X B,LIU Y,LIU Y,et al.A novel bat algorithm with habitat selection and Doppler effect in echoes for optimization [J].Expert Systems with Applications,2015,42(17):6350-6364.
[32]JIAN C F,WANG Y K.Batch Task Scheduling-oriented Opti-mization Modelling and Simulation in Cloud Manufacturing[J].International Journal of Simulation Modelling,2014,13(1):93-101.
[33]JIAN C F,WANG B,ZHANG M Y.Modified Breeding Particle Swarm Optimization Algorithm for Global QoS-aware Cloud Service Composition Optimization Problems [J].Journal of Chinese Mini-Micro Computer Systems,2017,38(7):1562-1567.(in Chinese)简琤峰,王斌,张美玉,等.一种求解云服务组合全局QoS最优问题的改进杂交粒子群算法[J].小型微型计算机系统,2017,38(7):1562-1567.
[34]NEAL R M.Connectionist learning of belief networks[J].Artificial Intelligence,1992,56(1):71-113.
[35]HINTON G E.Training products of experts by minimizing contrastive divergence [J].Neural Computation,2002,14(8):1771-1800.
[36]MOHAMMAD A K,MOHAMMAD M H.Deep Belief Network Training Improvement Using Elite Samples Minimizing Free Energy [J].International Journal of Pattern Recognition & Artificial Intelligence,2015,29(5):1-16.
[37]BENGIO Y.Learning Deep Architectures for AI [J].Founda-tions & Trends® in Machine Learning,2009,2(1):1-127.
[38]ZHANG R,XU Z B,HUANG G B,et al.Global convergence of online BP training with dynamic learning rate [J].IEEE Tran-sactions on Neural Networks & Learning Systems,2012,23(2):330-341.
[39]SUN H,LIU L,SUN Y F.Automatic Monitoring System of BP Digital Recognition [J].Mini-micro Systems,2009,30(3):503-506.
[40]ZOU Y Y,SUN J,SHAO K Y,et al.Fuzzy Neural NetworkController Based on Adaptive Learning Rate [J].Control and Instruments In Chemical Industry,2015,42(8):855-859.(in Chinese)邹彦艳,孙晶,邵克勇,等.基于自适应学习速率的模糊神经网络控制器[J].化工自动化及仪表,2015,42(8):855-859.
[41]HINTON G E.A Practical Guide to Training Restricted Boltzmann Machines [J].Momentum,2012,9(1):599-619.
[1] RAO Zhi-shuang, JIA Zhen, ZHANG Fan, LI Tian-rui. Key-Value Relational Memory Networks for Question Answering over Knowledge Graph [J]. Computer Science, 2022, 49(9): 202-207.
[2] TANG Ling-tao, WANG Di, ZHANG Lu-fei, LIU Sheng-yun. Federated Learning Scheme Based on Secure Multi-party Computation and Differential Privacy [J]. Computer Science, 2022, 49(9): 297-305.
[3] XU Yong-xin, ZHAO Jun-feng, WANG Ya-sha, XIE Bing, YANG Kai. Temporal Knowledge Graph Representation Learning [J]. Computer Science, 2022, 49(9): 162-171.
[4] WANG Jian, PENG Yu-qi, ZHAO Yu-fei, YANG Jian. Survey of Social Network Public Opinion Information Extraction Based on Deep Learning [J]. Computer Science, 2022, 49(8): 279-293.
[5] HAO Zhi-rong, CHEN Long, HUANG Jia-cheng. Class Discriminative Universal Adversarial Attack for Text Classification [J]. Computer Science, 2022, 49(8): 323-329.
[6] JIANG Meng-han, LI Shao-mei, ZHENG Hong-hao, ZHANG Jian-peng. Rumor Detection Model Based on Improved Position Embedding [J]. Computer Science, 2022, 49(8): 330-335.
[7] SUN Qi, JI Gen-lin, ZHANG Jie. Non-local Attention Based Generative Adversarial Network for Video Abnormal Event Detection [J]. Computer Science, 2022, 49(8): 172-177.
[8] HU Yan-yu, ZHAO Long, DONG Xiang-jun. Two-stage Deep Feature Selection Extraction Algorithm for Cancer Classification [J]. Computer Science, 2022, 49(7): 73-78.
[9] CHENG Cheng, JIANG Ai-lian. Real-time Semantic Segmentation Method Based on Multi-path Feature Extraction [J]. Computer Science, 2022, 49(7): 120-126.
[10] HOU Yu-tao, ABULIZI Abudukelimu, ABUDUKELIMU Halidanmu. Advances in Chinese Pre-training Models [J]. Computer Science, 2022, 49(7): 148-163.
[11] ZHOU Hui, SHI Hao-chen, TU Yao-feng, HUANG Sheng-jun. Robust Deep Neural Network Learning Based on Active Sampling [J]. Computer Science, 2022, 49(7): 164-169.
[12] SU Dan-ning, CAO Gui-tao, WANG Yan-nan, WANG Hong, REN He. Survey of Deep Learning for Radar Emitter Identification Based on Small Sample [J]. Computer Science, 2022, 49(7): 226-235.
[13] WANG Jun-feng, LIU Fan, YANG Sai, LYU Tan-yue, CHEN Zhi-yu, XU Feng. Dam Crack Detection Based on Multi-source Transfer Learning [J]. Computer Science, 2022, 49(6A): 319-324.
[14] CHU Yu-chun, GONG Hang, Wang Xue-fang, LIU Pei-shun. Study on Knowledge Distillation of Target Detection Algorithm Based on YOLOv4 [J]. Computer Science, 2022, 49(6A): 337-344.
[15] LIU Wei-ye, LU Hui-min, LI Yu-peng, MA Ning. Survey on Finger Vein Recognition Research [J]. Computer Science, 2022, 49(6A): 1-11.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!