Interspeech 2021 2021
DOI: 10.21437/interspeech.2021-627
|View full text |Cite
|
Sign up to set email alerts
|

Text Augmentation for Language Models in High Error Recognition Scenario

Abstract: We examine the effect of data augmentation for training of language models for speech recognition. We compare augmentation based on global error statistics with one based on per-word unigram statistics of ASR errors and observe that it is better to only pay attention the global substitution, deletion and insertion rates. This simple scheme also performs consistently better than label smoothing and its sampled variants. Additionally, we investigate into the behavior of perplexity estimated on augmented data, bu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(1 citation statement)
references
References 10 publications
0
1
0
Order By: Relevance
“…Rosenberg et al [24] also train a Tacotron model on Librispeech corpus and out-of-domain speech to explore the effect of acoustic diversity. Text augmentation has been used [25,26] to improve language model scores and reduce WER. The aforementioned research has been adapted for languages with sufficient resources.…”
Section: Data Augmentation In Asrmentioning
confidence: 99%
“…Rosenberg et al [24] also train a Tacotron model on Librispeech corpus and out-of-domain speech to explore the effect of acoustic diversity. Text augmentation has been used [25,26] to improve language model scores and reduce WER. The aforementioned research has been adapted for languages with sufficient resources.…”
Section: Data Augmentation In Asrmentioning
confidence: 99%