2019 2nd Workshop on Energy Efficient Machine Learning and Cognitive Computing for Embedded Applications (EMC2) 2019
DOI: 10.1109/emc249363.2019.00014
|View full text |Cite
|
Sign up to set email alerts
|

Accelerated CNN Training through Gradient Approximation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 4 publications
0
3
0
Order By: Relevance
“…of the feature map gradient tensors. (Wang and Nelaturu, 2019). offers a scaling approach to approximate the filter gradient tensors.…”
Section: Comparison To Related Workmentioning
confidence: 99%
“…of the feature map gradient tensors. (Wang and Nelaturu, 2019). offers a scaling approach to approximate the filter gradient tensors.…”
Section: Comparison To Related Workmentioning
confidence: 99%
“…Given probing vectors drawn from (7), we have to modify the scaling factor of the multi-channel randomized trace estimator (6) to ensure it is unbiased,…”
Section: Single Channel Case Let Us Start By Writing the Action Of A ...mentioning
confidence: 99%
“…Unfortunately, restricting memory usage without introducing significant computational overhead remains a challenge and can lead to difficult to manage additional complexity. Examples include (optimal) checkpointing [1,2], where the state is periodically stored and recomputed during the backward pass, invertible networks [3][4][5], where the state can be derived from the output, and certain approximation methods where computations are made with limited precision arithmetic [6] or where unbiased estimates are made of the gradient using certain approximations [7,8], e.g., via randomized automatic differentiation (RAD, [9]) or via direct feedback alignment (DFA, [10][11][12]).…”
Section: Introductionmentioning
confidence: 99%