2018
DOI: 10.1109/tcad.2018.2857080
|View full text |Cite
|
Sign up to set email alerts
|

Weighted Quantization-Regularization in DNNs for Weight Memory Minimization Toward HW Implementation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
6
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
3
2

Relationship

2
7

Authors

Journals

citations
Cited by 31 publications
(6 citation statements)
references
References 12 publications
0
6
0
Order By: Relevance
“…Exploding gradient problems: In addition to methods for solving the vanishing gradient problem, some other methods are exploited to tackle the exploding gradient problem, such as gradient clipping [38] and weight regularization [39]. There are two types of gradient clipping: the value clipping method is to clip the gradient that exceeds a preset threshold, and the norm clipping one adjusts the gradient according to its norm [38].…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Exploding gradient problems: In addition to methods for solving the vanishing gradient problem, some other methods are exploited to tackle the exploding gradient problem, such as gradient clipping [38] and weight regularization [39]. There are two types of gradient clipping: the value clipping method is to clip the gradient that exceeds a preset threshold, and the norm clipping one adjusts the gradient according to its norm [38].…”
Section: Related Workmentioning
confidence: 99%
“…These regularization methods add a norm term to the loss function to softly constrain the parameter range. If the exploding gradient occurs (i.e., the norm of the parameter becomes very large), the regularization term can "pull back" the weight to a relatively flat region (i.e., the region which is close to zero), thus limit the occurrence of exploding gradients to some extent [39]. Nevertheless, the regularization term still remains unresolved issues on efficiency and stability.…”
Section: Related Workmentioning
confidence: 99%
“…Attempting to close the gap between the computational intensity of DNNs and the available computing power, a wide variety of hardware accelerators for DNNs and other AI workloads have emerged in recent years. A considerable amount of research has improved the efficiency of DNNs and reduced their memory consumption by applying methods such as pruning [5], [6], quantization [7]- [9], and factorization [10], [11]. Alternatively, a network architecture that is expected to work efficiently on the target device can be designed and trained directly.…”
Section: Introductionmentioning
confidence: 99%
“…In the algorithm side, light-weighted neural networks are used to optimize DNNs algorithm, which uses minimal neural layers and channels to complete the signal detection and classification [4]. In addition, the weight quantization and pruning optimization techniques have demonstrated superior performance [5] [6]. Both of them can improve the computation performance by reducing the scale of parameters hence to decrease computation operations and memory requirements.…”
Section: Introductionmentioning
confidence: 99%