2015
DOI: 10.1137/130949993
|View full text |Cite
|
Sign up to set email alerts
|

Accelerated, Parallel, and Proximal Coordinate Descent

Abstract: Abstract. We propose a new randomized coordinate descent method for minimizing the sum of convex functions each of which depends on a small number of coordinates only. Our method (APPROX) is simultaneously Accelerated, Parallel, and PROXimal; this is the first time such a method is proposed. In the special case when the number of processors is equal to the number of coordinates, the method converges at the rate 2ωLR 2 /(k + 1) 2 , where k is the iteration counter,ω is a data-weighted average degree of separabi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

4
324
0
1

Year Published

2015
2015
2020
2020

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 239 publications
(329 citation statements)
references
References 24 publications
4
324
0
1
Order By: Relevance
“…Parallel methods were considered in [2,19,21], and more recently in [1,5,6,12,13,25,27,28]. A memory distributed method scaling to big data problems was recently developed in [22].…”
Section: Literaturementioning
confidence: 99%
See 2 more Smart Citations
“…Parallel methods were considered in [2,19,21], and more recently in [1,5,6,12,13,25,27,28]. A memory distributed method scaling to big data problems was recently developed in [22].…”
Section: Literaturementioning
confidence: 99%
“…The first assumption generalizes the ESO concept introduced in [21] and later used in [5,6,22,27,28] to nonuniform samplings. The second assumption requires that φ be strongly convex.…”
Section: Assumptionsmentioning
confidence: 99%
See 1 more Smart Citation
“…Nesterov's results have since been extended in several directions, for both randomized and deterministic update orders, (see, e.g. [2][3][4][5]). …”
Section: Introductionmentioning
confidence: 99%
“…Their code, called HOGWILD!, uses atomic operations to A. Aytekin, H. R. Feyzmahdavian avoid locking of loosely coupled memory locations in the minimization problem of sparse separable cost functions, and have achieved linear speedup in the number of processors. Following a similar sparsity and separability assumption, Fercoq and Richtárik have proposed an accelerated, parallel and proximal coordinate descent method to better utilize the available processors to achieve even further speedups [6].…”
Section: Introductionmentioning
confidence: 99%