6th Workshop on Spoken Language Technologies for Under-Resourced Languages (SLTU 2018) 2018
DOI: 10.21437/sltu.2018-3
|View full text |Cite
|
Sign up to set email alerts
|

Interspeech 2018 Low Resource Automatic Speech Recognition Challenge for Indian Languages

Abstract: India has more than 1500 1 languages, with 30 of them spoken by more than one million native speakers. Most of them are low-resource and could greatly benefit from speech and language technologies. Building speech recognition support for these low-resource languages requires innovation in handling constraints on data size, while also exploiting the unique properties and similarities among Indian languages. With this goal, we organized a low-resource Automatic Speech Recognition challenge for Indian languages a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
14
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
4
1

Relationship

3
7

Authors

Journals

citations
Cited by 33 publications
(14 citation statements)
references
References 10 publications
0
14
0
Order By: Relevance
“…Building on (Conneau et al, 2020a), pre-training with a diverse set of languages improves performance on languages not present in the pre-training data, which indicates that multilingual models may be useful for building speech technologies for very low resource languages, provided that some data for finetuning and evaluation is collected. Although these experiments do not include comparisons with monolingual models, the models beat previous SOTA models in terms of performance, which in turn are better performing than monolingual models (Srivastava et al, 2018). (Wang et al, 2021) proposes to combine supervised learning with labeled data and the CTC loss and SSL with the WAV2VEC2.0 loss in a Multitask Learning (MTL) scheme which they call UniSpeech.…”
Section: Models That Use Unlabeled Datamentioning
confidence: 99%
“…Building on (Conneau et al, 2020a), pre-training with a diverse set of languages improves performance on languages not present in the pre-training data, which indicates that multilingual models may be useful for building speech technologies for very low resource languages, provided that some data for finetuning and evaluation is collected. Although these experiments do not include comparisons with monolingual models, the models beat previous SOTA models in terms of performance, which in turn are better performing than monolingual models (Srivastava et al, 2018). (Wang et al, 2021) proposes to combine supervised learning with labeled data and the CTC loss and SSL with the WAV2VEC2.0 loss in a Multitask Learning (MTL) scheme which they call UniSpeech.…”
Section: Models That Use Unlabeled Datamentioning
confidence: 99%
“…Further, the selected data is split into a train (Trn) and test (Tst) without sentence overlap and with out-of-vocabulary (OOV) rates at about 30% between train and test sets. The Tel, Tam and Guj data are taken from Interspeech 2018 low resource automatic speech recognition challenge for Indian languages, for which, the data was provided by SpeechOcean.com and Microsoft [17]. The train and test sets are considered as-is for this challenge, however, the blind test set is modified with speed perturbations randomly between 1.1 to 1.4 (with increments of 0.05), and/or adding one noise randomly from white, babble and three noises chosen in the Musan dataset [18] considering the signal-to-noise ratio randomly between 18dB to 30dB at step of 1dB.…”
Section: Characteristics Of the Datasetmentioning
confidence: 99%
“…Tundra is a comprehensive dataset with more than 14 languages built from audiobooks [17]. Perhaps the closest dataset to the one presented in this paper is the Low Resource Automatic Speech Recognition Challenge for Indian Languages [18]. It has over 150 hours of Tamil, Telugu and Gujarati languages.…”
Section: Related Workmentioning
confidence: 99%