2012 IEEE Spoken Language Technology Workshop (SLT) 2012
DOI: 10.1109/slt.2012.6424216
|View full text |Cite
|
Sign up to set email alerts
|

Topic n-gram count language model adaptation for speech recognition

Abstract: We introduce novel language model (LM) adaptation approaches using the latent Dirichlet allocation (LDA) model. Observed n-grams in the training set are assigned to topics using soft and hard clustering. In soft clustering, each n-gram is assigned to topics such that the total count of that n-gram for all topics is equal to the global count of that n-gram in the training set. Here, the normalized topic weights of the n-gram are multiplied by the global n-gram count to form the topic n-gram count for the respec… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2013
2013
2019
2019

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 11 publications
(6 citation statements)
references
References 15 publications
0
6
0
Order By: Relevance
“…Also, notice that in this setup, expected WB achieved consistent improvements over fractional WB. A similar experiment was carried out with other training corpora and seed LMs, including LDA-based seeds like in [16] with analogous results.…”
Section: Split-and-merge Technique For Lm Adaptationmentioning
confidence: 80%
“…Also, notice that in this setup, expected WB achieved consistent improvements over fractional WB. A similar experiment was carried out with other training corpora and seed LMs, including LDA-based seeds like in [16] with analogous results.…”
Section: Split-and-merge Technique For Lm Adaptationmentioning
confidence: 80%
“…In this regard, there are different ways in which the interpolation weight can be selected: it can be set empirically by minimizing the perplexity in a development stage with data not seen during training [Clarkson, 1999, Tur and; it can also be estimated under some optimization algorithm, such as Expectation Maximization [Daumé et al, 2010] or Maximum A Posteriori (MAP) adaptation [Wang and Stolcke, 2007]; or it can be set dynamically depending on the current situation of the interaction (the topic of the speech, a specific speaker, etc.) [Haidar andO'Shaughnessy, 2012, Seymore andRosenfeld, 1997].…”
Section: Mixture-based Modelsmentioning
confidence: 99%
“…for natural language processing and text analysis applications such as speech recognition, parsing, spelling, etc. [7,51,9,33,41]. In addition, N-gram is applied for whole-genome protein sequences [26] and for computer virus detection [17,24].…”
Section: N-gram Analysismentioning
confidence: 99%