Interspeech 2018 2018
DOI: 10.21437/interspeech.2018-1302
|View full text |Cite
|
Sign up to set email alerts
|

BUT System for Low Resource Indian Language ASR

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
11
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 26 publications
(12 citation statements)
references
References 14 publications
1
11
0
Order By: Relevance
“…The best ASR system, trained with the BSRD v2 dataset and using a 15-gram character-based language model, achieved a CER and a WER of 10.49% and 25.45%, respectively. This is similar to the result found in [9] (15.2% of WER) for Tamil language, and comparable to the improvement reported by [13] and [14] using a multilingual setting training and multimodal data augmentation, respectively. From these tables, one clearly verifies the advantage of incorporating the language model onto the acoustic models obtained in Section VIII, with the WER dropping from 71.62% to 30.50% for the BRSD v1 set, and a 21.96% improvement for the BRSD v2 set.…”
Section: Complete Asr Experimentssupporting
confidence: 89%
See 1 more Smart Citation
“…The best ASR system, trained with the BSRD v2 dataset and using a 15-gram character-based language model, achieved a CER and a WER of 10.49% and 25.45%, respectively. This is similar to the result found in [9] (15.2% of WER) for Tamil language, and comparable to the improvement reported by [13] and [14] using a multilingual setting training and multimodal data augmentation, respectively. From these tables, one clearly verifies the advantage of incorporating the language model onto the acoustic models obtained in Section VIII, with the WER dropping from 71.62% to 30.50% for the BRSD v1 set, and a 21.96% improvement for the BRSD v2 set.…”
Section: Complete Asr Experimentssupporting
confidence: 89%
“…Most languages in the world lack the amount of text, speech, and/or linguistic resources required to build large models based on deep neural networks. There has been an increasing research interest on how to build a high-accuracy ASR system for languages with insufficient annotated data (using from hours to a few dozen hours of annotated speech), such as Brazilian Portuguese [8]; Indian languages (Gujarati, Tamil, and Telugu) [9], [10]; and Seneca (an Indigenous North-American language) [11].…”
Section: Development Context and General Diagram Of The Systemmentioning
confidence: 99%
“…The Jilebi system [11] which performed the best across all languages was a multilingual TDNN based system with transfer learning. The system used an n-gram based LM for decoding and an RNN-based LM for rescoring.…”
Section: Resultsmentioning
confidence: 99%
“…This indicates that developing ASR systems for Indic languages with their richer phoneme/character sets and vocabulary is more challenging than English. Jilebi (Pulugundla et al, 2018) 14.0 13.9 14.7 Cogknit (Fathima et al, 2018) 17.7 16.0 17.1 CSALT-LEAP (Srivastava et al, 2018) -16.3 17.6 ISI-Billa (Billa, 2018) 19.3 19.6 20.9 MTL-SOL (Sailor and Hain, 2020) 18.4 16.3 18.6 Reed (Sen et al, 2021) 16.1 19.9 20.2 CNN + Context temporal features (Sen et al, 2020) 18…”
Section: Ablation Studies On Fine-tuning and Decodingmentioning
confidence: 99%