Interspeech 2018 2018
DOI: 10.21437/interspeech.2018-2226
|View full text |Cite
|
Sign up to set email alerts
|

Articulatory and Stacked Bottleneck Features for Low Resource Speech Recognition

Abstract: In this paper, we discuss the benefits of using articulatory and stacked bottleneck features (SBF) for low resource speech recognition. Articulatory features (AF) which capture the underlying attributes of speech production are found to be robust to channel and speaker variations. However, building an efficient articulatory classifier to extract AF requires an enormous amount of data. In low resource acoustic modeling, we propose to train the bidirectional long short-term memory (BLSTM) articulatory classifier… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
4
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 20 publications
0
4
0
Order By: Relevance
“…The IITM Speech Lab systems [17] used articulatory and stacked bottlneck (SBN) features in the Acoustic Model by creating an bidirectional long short term memory (BLSTM) articulatory feature classifier using the pooled data. The best system with Word Error Rates of 24.29%, 30.33%, 17.9% for Gujarati, Telugu, Tamil respectively was a TDNN-based system with articulatory and SBN features.…”
Section: Resultsmentioning
confidence: 99%
“…The IITM Speech Lab systems [17] used articulatory and stacked bottlneck (SBN) features in the Acoustic Model by creating an bidirectional long short term memory (BLSTM) articulatory feature classifier using the pooled data. The best system with Word Error Rates of 24.29%, 30.33%, 17.9% for Gujarati, Telugu, Tamil respectively was a TDNN-based system with articulatory and SBN features.…”
Section: Resultsmentioning
confidence: 99%
“…Assamese, Bengali, Kurmanji, Lithuanian, Pashto, Turkish, and Vietnamese in the IARPA Babel [51] BLSTM & TDNN AF and SBF Multilingual corpus assistance and feature stitching Gujarati, Tamil, and Telugu [52] Initially, BN features were regarded as dependent on language, until the speech team of Brno University of Technology proposed a language-independent BN feature extraction framework [39]. The framework is a five-layer MLP with a sigmoid hidden unit, a linear bottleneck, and several output layers.…”
Section: Language-independent Bn Feature Multilingual Corpus Assistancementioning
confidence: 99%
“…Among them, the WER of Gujarati language dropped from 14.61% to 14.11%, and the WER of Telugu language dropped from 21.44% to 19.80%. The team concluded from the experimental results that the proposed combination of AF and SBF is an improvement over traditional features [52].…”
Section: Language-independent Bn Feature Multilingual Corpus Assistancementioning
confidence: 99%
See 1 more Smart Citation