2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2015
DOI: 10.1109/icassp.2015.7178826
|View full text |Cite
|
Sign up to set email alerts
|

Constructing long short-term memory based deep recurrent neural networks for large vocabulary speech recognition

Abstract: Long short-term memory (LSTM) recurrent neural networks (RNNs) have been shown to give state-of-the-art performance on many speech recognition tasks, as they are able to provide the learned dynamically changing contextual window of all sequence history. On the other hand, the convolutional neural networks (CNNs) have brought significant improvements to deep feed-forward neural networks (FFNNs), as they are able to better reduce spectral variation in the input signal. In this paper, a network architecture calle… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
118
1
3

Year Published

2016
2016
2023
2023

Publication Types

Select...
3
2
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 260 publications
(122 citation statements)
references
References 47 publications
0
118
1
3
Order By: Relevance
“…Several works report improvements using CNNs, TDNNs and LSTMs with respect to DNNs, especially for large vocabularies [17,1,13,18,7,11]. We do not find this behaviour in our task.…”
Section: Resultscontrasting
confidence: 58%
“…Several works report improvements using CNNs, TDNNs and LSTMs with respect to DNNs, especially for large vocabularies [17,1,13,18,7,11]. We do not find this behaviour in our task.…”
Section: Resultscontrasting
confidence: 58%
“…To understand class-specific spectral characteristics in the EEG recordings, we analyzed band powers in five frequency ranges: delta (0-4 Hz), theta (4-8 Hz), alpha (8)(9)(10)(11)(12)(13)(14), low beta (14)(15)(16)(17)(18)(19)(20), high beta (20)(21)(22)(23)(24)(25)(26)(27)(28)(29)(30) and low gamma .…”
Section: Visualizations Of the Spectral Differences Between Normalmentioning
confidence: 99%
“…For example, deep ConvNets reduced the error rates on the ImageNet image‐recognition challenge, where 1.2 million images must be classified into 1000 different classes, from above 26% to below 4% within 4 years [He et al, ; Krizhevsky et al, ]. ConvNets also reduced error rates in recognizing speech, for example, from English news broadcasts [Sainath et al, ,; Sercu et al, ]; however, in this field, hybrid models combining ConvNets with other machine‐learning components, notably recurrent networks, and deep neural networks without convolutions are also competitive [Li and Wu, ; Sainath et al, ; Sak et al, ]. Deep ConvNets also contributed to the spectacular success of AlphaGo, an artificial intelligence that beat the world champion in the game of Go [Silver et al, ].…”
Section: Introductionmentioning
confidence: 99%