2012
DOI: 10.48550/arxiv.1212.0873
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Parallel Coordinate Descent Methods for Big Data Optimization

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

4
204
0

Year Published

2015
2015
2019
2019

Publication Types

Select...
4
3

Relationship

5
2

Authors

Journals

citations
Cited by 40 publications
(208 citation statements)
references
References 0 publications
4
204
0
Order By: Relevance
“…If we run S2CD with the parameters set as in (13), then in each epoch the gradient of f is evaluated once (this is equivalent to n evaluations of ∇f i ), and the partial derivative of some function f i is evaluated 2m ≈ 52κ = O(κ) times. If we let C grad be the average cost of evaluating the gradient ∇f i and C pd be the average cost of evaluating the partial derivative ∇ j f i , then the total work of S2CD can be written as (nC grad + mC pd )k…”
Section: Complexity Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…If we run S2CD with the parameters set as in (13), then in each epoch the gradient of f is evaluated once (this is equivalent to n evaluations of ∇f i ), and the partial derivative of some function f i is evaluated 2m ≈ 52κ = O(κ) times. If we let C grad be the average cost of evaluating the gradient ∇f i and C pd be the average cost of evaluating the partial derivative ∇ j f i , then the total work of S2CD can be written as (nC grad + mC pd )k…”
Section: Complexity Resultsmentioning
confidence: 99%
“…The complexity can be improved to O( dβ τ µ log(1/ǫ)) in the case when τ coordinates are updated in each iteration, where β ∈ [1, τ ] is a problem-dependent constant[13]. This has been further studied for nonsmooth problems via smoothing[4], for arbitrary nonuniform distributions governing the selection of coordinates[15,12] and in the distributed setting[14,2,12].…”
mentioning
confidence: 99%
See 1 more Smart Citation
“…
In this work we study the parallel coordinate descent method (PCDM) proposed by Richtárik and Takáč [26] for minimizing a regularized convex function. We adopt elements from the work of Lu and Xiao [39], and combine them with several new insights, to obtain sharper iteration complexity results for PCDM than those presented in [26].
…”
mentioning
confidence: 99%
“…
In this work we study the parallel coordinate descent method (PCDM) proposed by Richtárik and Takáč [26] for minimizing a regularized convex function. We adopt elements from the work of Lu and Xiao [39], and combine them with several new insights, to obtain sharper iteration complexity results for PCDM than those presented in [26]. Moreover, we show that PCDM is monotonic in expectation, which was not confirmed in [26], and we also derive the first high probability iteration complexity result where the initial levelset is unbounded.
…”
mentioning
confidence: 99%