2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2018
DOI: 10.1109/icassp.2018.8462334
|View full text |Cite
|
Sign up to set email alerts
|

Accelerating Recurrent Neural Network Language Model Based Online Speech Recognition System

Abstract: This paper presents methods to accelerate recurrent neural network based language models (RNNLMs) for online speech recognition systems. Firstly, a lossy compression of the past hidden layer outputs (history vector) with caching is introduced in order to reduce the number of LM queries. Next, RNNLM computations are deployed in a CPU-GPU hybrid manner, which computes each layer of the model on a more advantageous platform. The added overhead by data exchanges between CPU and GPU is compensated through a frame-w… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
11
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
2
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 18 publications
(11 citation statements)
references
References 15 publications
0
11
0
Order By: Relevance
“…RNNs have been used for a variety of tasks, both regression and classification, such as natural language processing (Li and Xu 2018), speech recognition (Lee et al 2018), in clinical application (Tomašev et al 2019), and more recently, activity recognition from accelerometer data (Edel and Köppe 2016; Guan and Plötz 2017) and modeling of long-term human activity (Kim et al 2017). Because PAEE is influenced by past activities (lag effect), RNNs could be a suitable modeling candidate for tackling the challenge of PAEE estimation.…”
Section: Modeling Architecturementioning
confidence: 99%
“…RNNs have been used for a variety of tasks, both regression and classification, such as natural language processing (Li and Xu 2018), speech recognition (Lee et al 2018), in clinical application (Tomašev et al 2019), and more recently, activity recognition from accelerometer data (Edel and Köppe 2016; Guan and Plötz 2017) and modeling of long-term human activity (Kim et al 2017). Because PAEE is influenced by past activities (lag effect), RNNs could be a suitable modeling candidate for tackling the challenge of PAEE estimation.…”
Section: Modeling Architecturementioning
confidence: 99%
“…Toshniwal et al (2018) proposed a single end-to-end speech recognition system that works on 9 different Indian languages. Lee et al (2018) presented methods to accelerate RNNs language models for online speech recognition systems.…”
Section: Fig 1 Basic Structure Of Rnnsmentioning
confidence: 99%
“…Despite using feed-forward neural LMs in decoding, empirical results showed significant relative improvements both in speed and accuracy. Other relevant contributions addressing one-pass decoding with neural LMs have focused on heuristics to reduce the number of queries to the model and catching network states [20], alternative one-pass decoding strategies such as on-the-fly rescoring [21], improving CPU-GPU communications [22] and, more recently, combining Gated Recurrent Units with more efficient objective functions, such as Noise Contrastive Estimation [23]. Certainly different from these contributions, other authors have explored the idea of converting neural LMs, either recurrent or not, into ngram models that can thus be smoothly integrated into a conventional decoder [24], [25].…”
Section: Introductionmentioning
confidence: 99%