Computer Science ›› 2022, Vol. 49 ›› Issue (11): 309-315.doi: 10.11896/jsjkx.211200006

• Computer Network • Previous Articles     Next Articles

Communication Efficient Asynchronous ADMM for General Form Consensus Optimization

WANG Dong-xia, LEI Yong-mei, ZHANG Ze-yu   

  1. School of Computer Engineering and Science,Shanghai University,Shanghai 200444,China
  • Received:2021-12-01 Revised:2022-05-09 Online:2022-11-15 Published:2022-11-03
  • About author:WANG Dong-xia,born in 1983,Ph.D candidate,is a member of China Computer Federation.Her main research interests include distributed machine learning,parallel and distributed computing.
    LEI Yong-mei,born in 1965,professor,Ph.D supervisor,is a member of China Computer Federation.Her main research interests include parallel and distributed computing,big data analysis,etc.
  • Supported by:
    National Natural Science Foundation of China(U1811461).

Abstract: The distributed alternating direction method of multipliers(ADMM) is one of the most widely used methods for solving large-scale machine learning applications.However,most distributed ADMM algorithms are based on full model updates.With the increasing of system scale and data volume,the communication cost has become the bottleneck for the distributed ADMM when big data are involved.In order to reduce the communication cost in a distributed environment,a general form consensus asynchronous distributed alternating direction method of multipliers(GFC-ADADMM) is proposed in this paper.First,in the GFC-ADADMM,the associated model parameters rather than full model parameters are transmitted among nodes to reduce the transmission load,and the associated model parameters are filtered according to the characteristics of high-dimensional sparse data sets to further reduce the transmission load.Second,the GFC-ADMM is implemented by an asynchronous allreduce framework,which combines the advantage of the asynchronous communication protocol and the allreduce communication mode.Third,combining the advantages of the stale synchronous parallel(SSP) computing model,allreduce communication model,and hybrid programming model,the asynchronous allreduce framework and MPI/OpenMP hybrid programming model are adopted to implement the GFC-ADADMM,which improves calculation efficiency and communication efficiency of the algorithm.Finally,the sparse logistic regression problem is solved by the GFC-ADADMM.Evaluation with large-scale datasets shows that compared with state-of-the-art asynchronous distributed ADMM algorithms,the GFC-ADADMM can reduce the total running time by 15%-63%,and has higher accuracy in convergence.

Key words: Distributed alternating direction method of multipliers, General form consensus optimization, Sparse allreduce, Hybrid programming model, Logistic regression

CLC Number: 

  • TP391
[1]LI Y,WANG X,FANG W,et al.A Distributed ADMM Approach for Collaborative Regression Learning in Edge Computing[J].Computers,Materials and Continua,2019,58(2):493-508.
[2]BALAMURUGAN P,POSINASETTY A,SHEVADE S.ADMM for Training Sparse Structural SVMs with Augmented l1 Regularizers[C]//Siam International Conference on Data Mining.2016.
[3]DHAR S,YI C,RAMAKRISHNAN N,et al.ADMM BasedScalable Machine Learning on Spark[C]//2015 IEEE International Conference on Big Data.Santa Clara:IEEE,2015:1174-1182.
[4]LI W,LIU Y,TIAN Z,et al.Communication-Censored Linea-rized ADMM for Decentralized Consensus Optimization[J].IEEE Transactions on Signal Processing,2020,6:18-34.
[5]LIU Y,YUAN K,WU G,et al.Decentralized Dynamic ADMM with Quantized and Censored Communications[C]//53rd Asilomar Conference on Signals,Systems,and Computers.2019.
[6]ISSAID C B,ELGABLI A,PARK J,et al.Communication Efficient Distributed Learning with Censored,Quantized,and Ge-neralized Group ADMM[J].arXiv:2009.06459,2020.
[7]ZHANG R,KWOK J.Asynchronous distributed ADMM forconsensus optimization[C]//International Conference on Machine Learning.2014:1701-1709.
[8]CTANG T H,HONG M,LIAO W C,et al.Asynchronous Distributed ADMM for Large-Scale Optimization—Part I:Algorithm and Convergence Analysis[J].IEEE Transactions on Signal Processing,2016,64(12):3118-3130.
[9]BOYD S,PARIKH N,CHU E,et al.Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers[J].Foundations & Trends in Machine Learning,2010,3(1):1-122.
[10]ZHU R,NIU D,LI Z.A Block-wise,Asynchronous and Distri-buted ADMM Algorithm for General Form Consensus Optimization[J].arXiv:1802.08882,2018.
[11]JIN H,JESPERSEN D,MEHROTRA P,et al.High Perfor-mance Computing Using MPI and OpenMP on Multi-core Parallel Systems[J].Parallel Computing,2011,37(9):562-575.
[12]CHORLEY M J,WALKER D W.Performance Analysis of aHybrid MPI/OpenMP Application on Multi-core Clusters[J].Journal of Computational Science,2010,1(3):168-174.
[13]GENKIN A,MADIGAN L D.Large-scale Bayesian Logistic Regression for Text Categorization[J].Technometrics,2007,49(3):291-304.
[14]ALGAMAL Z Y,LEE M H.A Two-stage Sparse Logistic Regression for Optimal Gene Selection in High-dimensional Microarray Data Classification[J].Advances in Data Analysis and Classification,2019,13(3):753-771.
[15]SHEVADE S K,KEERTHI S S.A Simple and Efficient Algorithm for Gene Selection Using Sparse Logistic Regression[J].Bioinformatics,2003,19(17):2246-2253.
[16]ELGABLI A,PARK J,BEDI A S,et al.GADMM:Fast andCommunication Efficient Framework for Distributed Machine Learning[J].Journal of Machine Learning Research,2020,21(76):1-39.
[17]WANG S,LEI Y.Fast Communication Structure for Asyn-chronous Distributed ADMM under Unbalance Process Arrival Pattern[C]//27th International Conference on Artificial Neural Networks.2018:362-371.
[18]XIE J,LEI Y.ADMMLIB:A Library of Communication-Effi-cient AD-ADMM for Distributed Machine Learning[C]//IFIP International Conference on Network and Parallel Computing.Cham:Springer,2019:322-326.
[19]WANG D,LEI Y,XIE J,et al.HSAC-ALADMM:An Asynchronous Lazy ADMM Algorithm Based on Hierarchical Sparse Allreduce Communication[J].The Journal of Supercomputing.2021,77(8):8111-8134.
[20]WANG D,LEI Y,ZHOU J.Hybrid MPI/OpenMP ParallelAsynchronous Distributed Alternating Direction Method of Multipliers[J].Computing,2021,103(12):2737-2762.
[21]HSIA C Y,ZHU Y,LIN C J.A Study on Trust Region Update Rules in Newton Methods for Large-scale Linear Classification[C]//Proceedings of the Ninth Asian Conference on Machine Learning.2017:33-48.
[1] LI Jing-tai, WANG Xiao-dan. XGBoost for Imbalanced Data Based on Cost-sensitive Activation Function [J]. Computer Science, 2022, 49(5): 135-143.
[2] HU Yan-mei, YANG Bo, DUO Bin. Logistic Regression with Regularization Based on Network Structure [J]. Computer Science, 2021, 48(7): 281-291.
[3] YU Meng-chi, MU Jia-peng, CAI Jian, XU Jian. Noisy Label Classification Learning Based on Relabeling Method [J]. Computer Science, 2020, 47(6): 79-84.
[4] BIAN Yu-ning, LU Li-kun, LI Ye-li, ZENG Qing-tao, SUN Yan-xiong. Implementation of Financial Venture Capital Score Card Model Based on Logistic Regression [J]. Computer Science, 2020, 47(11A): 116-118.
[5] LIU Meng-juan,ZENG Gui-chuan,YUE Wei,QIU Li-zhou,WANG Jia-chang. Review on Click-through Rate Prediction Models for Display Advertising [J]. Computer Science, 2019, 46(7): 38-49.
[6] ZHANG Shi-xiang, LI Wang-geng, LI Tong, ZHU Nan-nan. Improved CoreSets Construction Algorithm for Bayesian Logistic Regression [J]. Computer Science, 2019, 46(11A): 98-102.
[7] ZHANG Xiao-feng and ZHANG De-ping. Software Failure Prediction Model Based on Quasi-likelihood Method [J]. Computer Science, 2016, 43(Z11): 486-489.
[8] ZHANG Zheng-qing, ZHU Yi-jian, BAI Rui-rui, HUANG Yi-qing and YAN Jian-feng. Application of Service Bundling in Churn Predict System [J]. Computer Science, 2016, 43(Z11): 585-590.
[9] YANG Xu-hua and ZHONG Nan-yi. Forecasting of Hospital Outpatient Based on Deep Belief Network [J]. Computer Science, 2016, 43(Z11): 26-30.
[10] JIANG Jie,WANG Zhuo-fang,GONG Rong-sheng and CHEN Tie-ming. Imbalanced Data Classification Method and its Application Research for Intrusion Detection [J]. Computer Science, 2013, 40(4): 131-135.
[11] . [J]. Computer Science, 2008, 35(10): 197-199.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!