ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2020
DOI: 10.1109/icassp40776.2020.9054164
|View full text |Cite
|
Sign up to set email alerts
|

Accelerating Distributed Deep Learning By Adaptive Gradient Quantization

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 16 publications
(8 citation statements)
references
References 2 publications
0
8
0
Order By: Relevance
“…We use test accuracy to measure the learning performance. We compare our proposed DQ-SGD with the following baselines: SignSGD [5], TernGrad [12], QSGD [2], Adaptive [11] and AdaQS [9].…”
Section: Methodsmentioning
confidence: 99%
See 3 more Smart Citations
“…We use test accuracy to measure the learning performance. We compare our proposed DQ-SGD with the following baselines: SignSGD [5], TernGrad [12], QSGD [2], Adaptive [11] and AdaQS [9].…”
Section: Methodsmentioning
confidence: 99%
“…Several studies try to construct adaptive quantization schemes through engineering heuristics or empirical observations. However, they do not come up with a solid theoretical analysis [9], [10], [11], which even results in contradicted conclusions. More specifically, MQGrad [10], and AdaQS [11] suggest using few quantization bits in early epochs and gradually increase the number of bits in later epochs; while the scheme proposed by Anders [9] states that more quantization bits should be used for the gradient with a larger root-mean-squared (RMS) value, choosing to use more bits in the early training stage and fewer bits in the later stage.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…In order to solve the problem of low model accuracy or low compression rate in the existing quantization schemes, Guo et al [ 11 ] proposed a novel adaptive quantization scheme(AdaQS). AdaQS algorithm according to the ratio of the mean and standard deviation of the gradient (MSDR) automatically determine the quantitative level, and achieve a balance between model accuracy and quantitative level.…”
Section: Related Workmentioning
confidence: 99%