1990
DOI: 10.1117/12.20700
|View full text |Cite
|
Sign up to set email alerts
|

Weight discretization paradigm for optical neural networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
35
0

Year Published

1998
1998
2023
2023

Publication Types

Select...
4
4
2

Relationship

1
9

Authors

Journals

citations
Cited by 68 publications
(35 citation statements)
references
References 0 publications
0
35
0
Order By: Relevance
“…Note that pre-training with continuous weights is not necessary, but is often included for faster convergence. The discretization of the weights described by Fiesler et al 16] uses d discretization levels with d= 2 , 3 , 5 , 7 , o r 9 . T h e discretization levels are symmetric around zero, except ford = 2 and are equidistant:…”
Section: Discussionmentioning
confidence: 99%
“…Note that pre-training with continuous weights is not necessary, but is often included for faster convergence. The discretization of the weights described by Fiesler et al 16] uses d discretization levels with d= 2 , 3 , 5 , 7 , o r 9 . T h e discretization levels are symmetric around zero, except ford = 2 and are equidistant:…”
Section: Discussionmentioning
confidence: 99%
“…However, direct quantization of the trained floating-point weights does not yield good results. Therefore, we employ the weight quantization strategy similar to the algorithms proposed in [7,13] to retrain the weights after the direct quantization. Also, internal signals (output values of the units) are uniformly quantized from 0 to 1.…”
Section: Fixed-point Dnn Design For Phoneme Recognitionmentioning
confidence: 99%
“…In order to solve this problem, methods of reducing the size of a CNN model or designing an efficient CNN structure have emerged as core fields in recent deep neural network research. Representative methods include pruning [9]- [12], quantization [13]- [15], knowledge distillation [16]- [18], weight sharing [19]- [22], and efficient structural design methods, e.g., Depthwise Separable Convolution [23]- [26]. These methods are widely used to compress the size of the CNN model.…”
Section: Introductionmentioning
confidence: 99%