ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2019
DOI: 10.1109/icassp.2019.8683481
|View full text |Cite
|
Sign up to set email alerts
|

Improvements to N-gram Language Model Using Text Generated from Neural Language Model

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
12
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 22 publications
(12 citation statements)
references
References 18 publications
0
12
0
Order By: Relevance
“…A language model is used to perform the decoding process on the network output for the speech-to-text conversion process. The decoding process is performed by calculating the probability of appearance of each word based on the exact word order [19]. This probability is calculated from the word chunks based on word order in the N-grams model.…”
Section: Language Modelmentioning
confidence: 99%
“…A language model is used to perform the decoding process on the network output for the speech-to-text conversion process. The decoding process is performed by calculating the probability of appearance of each word based on the exact word order [19]. This probability is calculated from the word chunks based on word order in the N-grams model.…”
Section: Language Modelmentioning
confidence: 99%
“…Sometimes these are called approximative models as they try to capture the knowledge of the neural model through their augmented training corpus. Suzuki et al [8] uses a domain balanced mixture of the training corpora to train a shallow RNNLM for text generation, and improve speech recognition results for Japanese, Korean and English. Wang et al [9] report using general domain pre-trained Transformer [10] to augment text corpora used to train LMs.…”
Section: Introduction and Problem Statementmentioning
confidence: 99%
“…Subword unit based ASR has been demonstrated to improve WER for several morphologically rich languages [12,13]. Suzuki et al [8] use subword approach for data augmentation to enrich text corpora to train BNLM, but compose these subwords back into words to prepare the final LM, unlike our approach that retokenizes words into subword units in the final LM.…”
Section: Introduction and Problem Statementmentioning
confidence: 99%
“…Although by splitting the decoding into two parts we can leverage knowledge of the NNLMs and demonstrate significant Word Error Rate Reduction (WERR), it also introduces considerable processing delay [4,5,12]. Therefore, techniques exploiting the capabilities of NNLMs in a single-pass decoding approach have received particular attention recently [9,13]. A possible technique is to augment the in-domain training data with a large text corpus generated by an NNLM [1,3].…”
Section: Introductionmentioning
confidence: 99%
“…In a related work, Suzuki et al [13] use a domain balanced mixture of the training corpora to train a shallow RNNLM for text generation and improve speech recognition results for Japanese, Korean and English tasks. For Korean subword-based language models are also utilized, but only for text generation, since in the language model of the ASR system subwords are merged.…”
Section: Introductionmentioning
confidence: 99%