2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2016
DOI: 10.1109/icassp.2016.7472457
|View full text |Cite
|
Sign up to set email alerts
|

D-FW: Communication efficient distributed algorithms for high-dimensional sparse optimization

Abstract: International audienceWe propose distributed algorithms for high-dimensional sparse optimization. In many applications, the parameter is sparse but high-dimensional. This is pathological for existing distributed algorithms as the latter require an information exchange stage involving transmission of the full parameter, which may not be sparse during the intermediate steps of optimization. The novelty of this work is to develop communication efficient algorithms using the stochastic Frank-Wolfe (sFW) algorithm,… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
16
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
4
2

Relationship

1
5

Authors

Journals

citations
Cited by 10 publications
(16 citation statements)
references
References 15 publications
0
16
0
Order By: Relevance
“…Note that the bounded gradient assumption is a regular assumption in the convergence analysis of decentralized gradient methods (see, [4], [5], [18], [35], [36], [56], [57], [27], [62] for example), even in the convex setting [24] and also [10], though it is not required for centralized gradient descent.…”
Section: B Convergence Results Of Dgdmentioning
confidence: 99%
See 2 more Smart Citations
“…Note that the bounded gradient assumption is a regular assumption in the convergence analysis of decentralized gradient methods (see, [4], [5], [18], [35], [36], [56], [57], [27], [62] for example), even in the convex setting [24] and also [10], though it is not required for centralized gradient descent.…”
Section: B Convergence Results Of Dgdmentioning
confidence: 99%
“…In [62], the authors established convergence results similar to Prox-DGD under diminishing step sizes. The stochastic version of DeFW has also been developed in [27] for high-dimensional convex sparse optimization. The next one is projected stochastic gradient algorithm (Proj SGD) [4] for constrained, nonconvex, smooth consensus optimization with a convex constrained set.…”
Section: Related Work and Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…In the constrained setting, and in particular for distributed FW algorithms, the communicationefficient versions were only studied for specific problems such as sparse learning (Bellet et al, 2015;Lafond et al, 2016). In this paper, however, we develop Quantized Frank-Wolfe (QFW), a general communication-efficient distributed FW for both convex and non-convex objective functions.…”
Section: Introductionmentioning
confidence: 99%
“…for all t ≥ 1, where θ is an optimal solution to (1). Furthermore, if F is µ-strongly convex and the optimal solution θ lies in the interior of C, i.e., δ > 0 (cf.…”
mentioning
confidence: 99%