2022
DOI: 10.48550/arxiv.2204.07553
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Improving Rare Word Recognition with LM-aware MWER Training

Abstract: Language models (LMs) significantly improve the recognition accuracy of end-to-end (E2E) models on words rarely seen during training, when used in either the shallow fusion or the rescoring setups. In this work, we introduce LMs in the learning of hybrid autoregressive transducer (HAT) models in the discriminative training framework, to mitigate the training versus inference gap regarding the use of LMs. For the shallow fusion setup, we use LMs during both hypotheses generation and loss computation, and the LM… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 24 publications
0
2
0
Order By: Relevance
“…In this study, the feed-forward-network f is a linear layer without any activation functions. The rescoring model is trained using a standard minimum word error rate (MWER) loss calculated on n-best hypotheses from the first-pass: loss [28,29,30]:…”
Section: Rescoring Modelmentioning
confidence: 99%
See 1 more Smart Citation
“…In this study, the feed-forward-network f is a linear layer without any activation functions. The rescoring model is trained using a standard minimum word error rate (MWER) loss calculated on n-best hypotheses from the first-pass: loss [28,29,30]:…”
Section: Rescoring Modelmentioning
confidence: 99%
“…The hyperparameters for the model architecture are outlined in Table 2. Both BERT models were first pre-trained on the MC4 dataset [30], and then domain-adapted using internal in-domain data. The BERT models were used to initialize the baseline RescoreBERT models, followed by MWER-based training on the general training data [10].…”
Section: Rescoring Modelsmentioning
confidence: 99%