2015
DOI: 10.1109/tac.2014.2363299
|View full text |Cite
|
Sign up to set email alerts
|

Linear Convergence Rate of a Class of Distributed Augmented Lagrangian Algorithms

Abstract: Abstract-We study distributed optimization where nodes cooperatively minimize the sum of their individual, locally known, convex costs fi(x)'s, x ∈ R d is global. Distributed augmented Lagrangian (AL) methods have good empirical performance on several signal processing and learning applications, but there is limited understanding of their convergence rates and how it depends on the underlying network. This paper establishes globally linear (geometric) convergence rates of a class of deterministic and randomize… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
154
0

Year Published

2016
2016
2021
2021

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 132 publications
(155 citation statements)
references
References 47 publications
0
154
0
Order By: Relevance
“…If we ignore the cross term 2c[z k+1 − A d x ] e λ k+1 , then (14) tells us that as long asλ k changes from one iteration to the next one, the network of agents is, on average, getting closer to the optimal solution of D. The additional cross term depends on the consensus error e λ k+1 = λ k+1 −λ k+1 and is clearly a distinctive feature of Tracking-ADMM (as opposed to the parallel ADMM) arising from its distributed nature. In the parallel ADMM, a condition similar to (14) can be derived and leveraged to prove the algorithm convergence. However, as expected, this is not the case here: (14) is not sufficient to prove convergence of the proposed scheme and we are required to study the interplay between the minimization step in (6a) and the two consensus steps in (6b) and (6c).…”
Section: Proposition 1 (Local Optimality) Under Assumptions 1 and 2 mentioning
confidence: 99%
See 3 more Smart Citations
“…If we ignore the cross term 2c[z k+1 − A d x ] e λ k+1 , then (14) tells us that as long asλ k changes from one iteration to the next one, the network of agents is, on average, getting closer to the optimal solution of D. The additional cross term depends on the consensus error e λ k+1 = λ k+1 −λ k+1 and is clearly a distinctive feature of Tracking-ADMM (as opposed to the parallel ADMM) arising from its distributed nature. In the parallel ADMM, a condition similar to (14) can be derived and leveraged to prove the algorithm convergence. However, as expected, this is not the case here: (14) is not sufficient to prove convergence of the proposed scheme and we are required to study the interplay between the minimization step in (6a) and the two consensus steps in (6b) and (6c).…”
Section: Proposition 1 (Local Optimality) Under Assumptions 1 and 2 mentioning
confidence: 99%
“…In the parallel ADMM, a condition similar to (14) can be derived and leveraged to prove the algorithm convergence. However, as expected, this is not the case here: (14) is not sufficient to prove convergence of the proposed scheme and we are required to study the interplay between the minimization step in (6a) and the two consensus steps in (6b) and (6c).…”
Section: Proposition 1 (Local Optimality) Under Assumptions 1 and 2 mentioning
confidence: 99%
See 2 more Smart Citations
“…Generally speaking, a consensus-based algorithm is to directly integrate consensus theory into an optimization algorithm which only involves primal decision variables, and the distributed algorithms along this line subsume distributed subgradient [2], distributed primal-dual subgradient algorithms [3], distributed quasi-monotone subgradient algorithm [4], asynchronous distributed gradient [5], Newton-Raphson consensus [6], dual averaging [7], diffusion adaptation strategy [8], fast distributed gradient [9], and stochastic mirror descent [10]. On the other hand, the dual-decomposition-based algorithms aim at handling the alignment of all local decision variables by equality constraints, through introducing corresponding dual variables, and typical algorithms include alternating direction method of multipliers (ADMM) [11], augmented Lagrangian method [12], distributed dual proximal gradient [13], EXTRA [14], and distributed forwardbackward Bregman splitting [15].…”
Section: Introductionmentioning
confidence: 99%