2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2014
DOI: 10.1109/icassp.2014.6855060
|View full text |Cite
|
Sign up to set email alerts
|

X1000 real-time phoneme recognition VLSI using feed-forward deep neural networks

Abstract: Deep neural networks show very good performance in phoneme and speech recognition applications when compared to previously used GMM (Gaussian Mixture Model)-based ones. However, efficient implementation of deep neural networks is difficult because the network size needs to be very large when high recognition accuracy is demanded. In this work, we develop a digital VLSI for phoneme recognition using deep neural networks and assess the design in terms of throughput, chip size, and power consumption. The develope… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
27
0

Year Published

2015
2015
2024
2024

Publication Types

Select...
4
4

Relationship

1
7

Authors

Journals

citations
Cited by 37 publications
(27 citation statements)
references
References 15 publications
0
27
0
Order By: Relevance
“…After the fine tuning of the parameters, the training with the boundary model was applied. Although the drop out [17] will improve the absolute performance of DNNs, we believe that it does not critically affect the relative results among the methods, as discussed in [7]. Table 3 specifies the WAs with different number of bits for discretization.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…After the fine tuning of the parameters, the training with the boundary model was applied. Although the drop out [17] will improve the absolute performance of DNNs, we believe that it does not critically affect the relative results among the methods, as discussed in [7]. Table 3 specifies the WAs with different number of bits for discretization.…”
Section: Methodsmentioning
confidence: 99%
“…A fixed-point DNN, whose weights, bias parameters and middle layer inputs are linearly quantized to n bits, enables fast processing on a Very Large Scale Integration [7] or a CPU with Supplemental Streaming SIMD Extensions 3 (SSSE3) instruction set [8]. The parameters of fixed-point DNNs are trained by iterating the quantization of weights to n bits and usual back propagation [7,9].…”
Section: Introductionmentioning
confidence: 99%
“…Kim et al use reduced precision for a custom neural network circuit design [22]. However, this implementation lacks the configurability to run di↵erent networks at di↵er-ent precisions.…”
Section: Related Workmentioning
confidence: 98%
“…The performances of the networks are shown in TABLE III. (N, K) is (16,4) for the structured sparse networks. Floatingpoint networks show the lowest error rate when normalizers are not applied.…”
Section: B Cifar-10mentioning
confidence: 99%