2012 American Control Conference (ACC) 2012
DOI: 10.1109/acc.2012.6315289
|View full text |Cite
|
Sign up to set email alerts
|

Distributed dual averaging for convex optimization under communication delays

Abstract: In this paper we extend and analyze the distributed dual averaging algorithm [1] to handle communication delays and general stochastic consensus protocols. Assuming each network link experiences some fixed bounded delay, we show that distributed dual averaging converges and the error decays at a rate O(T −0.5 ) where T is the number of iterations. This bound is an improvement over [1] by a logarithmic factor in T for networks of fixed size. Finally, we extend the algorithm to the case of using general non-aver… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
73
0

Year Published

2012
2012
2021
2021

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 95 publications
(73 citation statements)
references
References 9 publications
0
73
0
Order By: Relevance
“…In [8] it is proven that restricting to doubly stochastic consensus protocols in distributed dual averaging is not necessary and it is still possible to converge to the optimum with a general row stochastic protocol P . However there are multiple reasons why using a row stochastic matrix may not be desirable.…”
Section: Push-sum Consensusmentioning
confidence: 99%
See 3 more Smart Citations
“…In [8] it is proven that restricting to doubly stochastic consensus protocols in distributed dual averaging is not necessary and it is still possible to converge to the optimum with a general row stochastic protocol P . However there are multiple reasons why using a row stochastic matrix may not be desirable.…”
Section: Push-sum Consensusmentioning
confidence: 99%
“…However there are multiple reasons why using a row stochastic matrix may not be desirable. The bias correction described in [8] requires knowledge of the stationary distribution of P in advance which is restrictive. Moreover, with a time-varying consensus protocol P (t), we may not even be able to specify the stationary distribution beyond its expectation and variance [9] or may only be able to achieve average consensus in expectation [10].…”
Section: Push-sum Consensusmentioning
confidence: 99%
See 2 more Smart Citations
“…(Here by private cost f i (·) we mean that the function f i (·) is known only by node i.) Existing distributed (sub)gradient algorithms, e.g., the algorithm proposed in [2] and extended and analyzed in [3], [2], [4], [5], [6], [7], and the one proposed in [8] and extended and analyzed in [9], [10], converge slowly. For example, assuming possibly non-differentiable, convex f i 's, with bounded gradients over the constraint set, algorithm [2] with constant step size ↵, after k iterations, has the error in the cost O(↵+1/(↵k)), which, for the optimized ↵, gives O(1/✏ 2 ) convergence time.…”
Section: Introductionmentioning
confidence: 99%