ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2020
DOI: 10.1109/icassp40776.2020.9053808
|View full text |Cite
|
Sign up to set email alerts
|

Improving the Performance of Transformer Based Low Resource Speech Recognition for Indian Languages

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
14
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 28 publications
(15 citation statements)
references
References 14 publications
0
14
0
Order By: Relevance
“…Their extensive experimentation with classical SOTA techniques shows that transformer based models perform better than i.e., Mono-DNN, Mono-LSTM, SHL-MDNN, SHL-MLSTM, and SHL-MLSTM-residual models. (Shetty and NJ, 2020) also use transformer based models to show performance gains when using language ID. They conduct experiments in three Indian languages i.e., Gujarati, Tamil, and Telugu.…”
Section: Models That Use Only Labeled Datamentioning
confidence: 99%
“…Their extensive experimentation with classical SOTA techniques shows that transformer based models perform better than i.e., Mono-DNN, Mono-LSTM, SHL-MDNN, SHL-MLSTM, and SHL-MLSTM-residual models. (Shetty and NJ, 2020) also use transformer based models to show performance gains when using language ID. They conduct experiments in three Indian languages i.e., Gujarati, Tamil, and Telugu.…”
Section: Models That Use Only Labeled Datamentioning
confidence: 99%
“…A few multi-lingual approaches are recently proposed in the E2E framework [23,24,25]. Authors in [23] propose the multilingual RNN-T model with language specific adapters and datasampling to handle data imbalance.…”
Section: Relation To Prior Workmentioning
confidence: 99%
“…Audio-to-byte E2E system is proposed in [24] where bytes are used as target units instead of grapheme or word piece units, as bytes are suitable to scale to multiple languages. A transformer based multi-lingual E2E model, along with methods to incorporate language information is proposed in [25]. Although multi-lingual methods are attractive to address the problem of low-resource languages, the transfer learning methods, besides being simple and effective, have the benefit of not needing the high-resource language data, but only the models trained on them.…”
Section: Relation To Prior Workmentioning
confidence: 99%
“…In this context, many works are based on E2E frameworks to explore different ASR scenarios and improve the performance accordingly. For example, code-switch [6][7][8][9] and low-resource task [10][11][12] etc.…”
Section: Introductionmentioning
confidence: 99%