2017
DOI: 10.48550/arxiv.1711.06771
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Approximate Gradient Coding via Sparse Random Graphs

Abstract: Distributed algorithms are often beset by the straggler effect, where the slowest compute nodes in the system dictate the overall running time. Coding-theoretic techniques have been recently proposed to mitigate stragglers via algorithmic redundancy. Prior work in coded computation and gradient coding has mainly focused on exact recovery of the desired output. However, slightly inexact solutions can be acceptable in applications that are robust to noise, such as model training via gradient-based algorithms. In… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

2
91
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 27 publications
(93 citation statements)
references
References 14 publications
2
91
0
Order By: Relevance
“…• We provide a thorough analysis for these two schemes under shuffling and obtain expressions for their expected optimal decoding error. Our analysis for the FRC scheme extends the existing results [8,Thm. 6] for the heterogeneous straggler model.…”
Section: Introductionsupporting
confidence: 84%
See 2 more Smart Citations
“…• We provide a thorough analysis for these two schemes under shuffling and obtain expressions for their expected optimal decoding error. Our analysis for the FRC scheme extends the existing results [8,Thm. 6] for the heterogeneous straggler model.…”
Section: Introductionsupporting
confidence: 84%
“…Let A ∈ R n×r be the submatrix is formed by considering the columns of B that correspond to non-stragglers. This matrix A is termed as a non-straggler matrix [8]. For an (s − 1)-tolerant coding scheme, master is guaranteed to compute the exact gradient g when r ≥ n − s + 1 [2].…”
Section: System Modelmentioning
confidence: 99%
See 1 more Smart Citation
“…One of the applications of coding theory is to enable error correction [50]. Gradient coding is originally proposed as a straggler mitigation method [106], which is used to speed up synchronous distributed first-order methods [17,23,89]. Several works build upon it and extend it to the adversarial setup.…”
Section: Gradient Codingmentioning
confidence: 99%
“…The major difference between Algorithms 2 and DGD is (17): instead of updating the estimates by the sum of all agents updates, BGD uses an aggregation rule GradFilter (•). Generally speaking, GradFilter : R d × n → R d is a function that takes n vectors of d-dimension, and output a vector of d-dimension.…”
mentioning
confidence: 99%