2019
DOI: 10.1631/fitee.1800469
|View full text |Cite
|
Sign up to set email alerts
|

Binary neural networks for speech recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
13
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 23 publications
(13 citation statements)
references
References 30 publications
0
13
0
Order By: Relevance
“…Dai et al [20] proposed a hidden-layer LSTM and grow-and-prune training method to address the problems of model redundancy and runtime delay. Qian et al [21] introduced binary neural network for acoustic modeling in speech recognition. Mori et al [22] performed Tensor-Train decomposition on the weight matrix of the recurrent network to reduce the number of ASR parameters.…”
Section: Related Workmentioning
confidence: 99%
“…Dai et al [20] proposed a hidden-layer LSTM and grow-and-prune training method to address the problems of model redundancy and runtime delay. Qian et al [21] introduced binary neural network for acoustic modeling in speech recognition. Mori et al [22] performed Tensor-Train decomposition on the weight matrix of the recurrent network to reduce the number of ASR parameters.…”
Section: Related Workmentioning
confidence: 99%
“…An ultimate goal for many data and resource intensive deep learning based AI applications, including ASR systems, is to derive "lossless" model compression approaches that allow high performance and low-footprint speech recognition systems to be constructed while incurring minimum performance degradation. To this end, one efficient solution is to use low-bit deep neural network (DNN) quantization techniques [17,18,19,20], which has drawn increasing interest in the machine learning and speech technology community in recent years. By replacing floating point weights with low precision values, the resulting quantization methods can significantly reduce the model size and inference time without modifying the model architectures.…”
Section: Introductionmentioning
confidence: 99%
“…Another powerful family of techniques recently drawing increasing interest across the machine learning, computer vision and speech technology communities to solve this problem is to use low-bit DNN quantization techniques [31]- [37], [52], [57], [58], [62], [74], [75]. By replacing floating point based DNN parameters with low precision values, for example, binary numbers, model sizes can be dramatically reduced without changing the DNN architecture [32], [57], [73].…”
Section: Introductionmentioning
confidence: 99%
“…1) To the best of our knowledge, this paper presents the first work in the speech technology community to apply mixed precision DNN quantization techniques to both LSTM-RNN and Transformer based NNLMs. In contrast, prior researches within the speech community in this direction largely focused on uniform precision based quantization of convolutional neural networks (CNNs) acoustic models [62] and LSTM-RNN language models [57], [58], [75].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation