Asynchronous Distributed ADMM for Large-Scale Optimization-Part I: Algorithm and Convergence Analysis

被引:176
作者
Chang, Tsung-Hui [1 ]
Hong, Mingyi [2 ]
Liao, Wei-Cheng [3 ]
Wang, Xiangfeng [4 ]
机构
[1] Chinese Univ Hong Kong, Sch Sci & Engn, Shenzhen 518172, Peoples R China
[2] Iowa State Univ, Dept Ind & Mfg Syst Engn, Ames, IA 50011 USA
[3] Univ Minnesota, Dept Elect & Comp Engn, Minneapolis, MN 55455 USA
[4] E China Normal Univ, Shanghai Key Lab Trustworthy Comp, Sch Comp Sci & Software Engn, Shanghai 200062, Peoples R China
基金
美国国家科学基金会;
关键词
Distributed optimization; ADMM; asynchronous; consensus optimization; CONVEX;
D O I
10.1109/TSP.2016.2537271
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
080906 [电磁信息功能材料与结构]; 082806 [农业信息与电气工程];
摘要
Aiming at solving large-scale optimization problems, this paper studies distributed optimization methods based on the alternating direction method of multipliers (ADMM). By formulating the optimization problem as a consensus problem, the ADMM can be used to solve the consensus problem in a fully parallel fashion over a computer network with a star topology. However, traditional synchronized computation does not scale well with the problem size, as the speed of the algorithm is limited by the slowest workers. This is particularly true in a heterogeneous network where the computing nodes experience different computation and communication delays. In this paper, we propose an asynchronous distributed ADMM (AD-ADMM), which can effectively improve the time efficiency of distributed optimization. Our main interest lies in analyzing the convergence conditions of the AD-ADMM, under the popular partially asynchronous model, which is defined based on a maximum tolerable delay of the network. Specifically, by considering general and possibly non-convex cost functions, we show that the AD-ADMM is guaranteed to converge to the set of Karush-Kuhn-Tucker (KKT) points as long as the algorithm parameters are chosen appropriately according to the network delay. We further illustrate that the asynchrony of the ADMM has to be handled with care, as slightly modifying the implementation of the AD-ADMM can jeopardize the algorithm convergence, even under the standard convex setting.
引用
收藏
页码:3118 / 3130
页数:13
相关论文
共 36 条
[1]
[Anonymous], 2012, Scaling up Machine LearningParallel and Distributed Approaches
[2]
[Anonymous], 2003, NONLINEAR PROGRAMMIN
[3]
[Anonymous], 2011, Advances in Neural Information Processing Systems
[4]
[Anonymous], 2011, Advances in Neural Information Processing Systems
[5]
Bertsekas D. P., 1989, PARALLEL DISTRIBUTED, V23
[6]
Distributed optimization and statistical learning via the alternating direction method of multipliers [J].
Boyd S. ;
Parikh N. ;
Chu E. ;
Peleato B. ;
Eckstein J. .
Foundations and Trends in Machine Learning, 2010, 3 (01) :1-122
[7]
Convex Optimization for Big Data [J].
Cevher, Volkan ;
Becker, Stephen ;
Schmidt, Mark .
IEEE SIGNAL PROCESSING MAGAZINE, 2014, 31 (05) :32-43
[8]
Chang T.-H., 2016, ELECT COMPANION AS 1
[9]
Chang TH, 2016, INT CONF ACOUST SPEE, P4781, DOI 10.1109/ICASSP.2016.7472585
[10]
Asynchronous Distributed ADMM for Large-Scale Optimization-Part II: Linear Convergence Analysis and Numerical Performance [J].
Chang, Tsung-Hui ;
Liao, Wei-Cheng ;
Hong, Mingyi ;
Wang, Xiangfeng .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2016, 64 (12) :3131-3144