Proceedings of the Workshop on Speech and Natural Language - HLT '91 1992
DOI: 10.3115/1075527.1075552
|View full text |Cite
|
Sign up to set email alerts
|

Improvements in stochastic language modeling

Abstract: We describe two attempt to improve our stochastic language models. In the first, we identify a systematic overestimation in the traditional backoff model, and use statisticalreasoning to correct it. Our modification results in up to 6% reduction in the perplexity of various tasks. Although the improvement is modest, it is achieved with hardly any increasein the complexity of the model. Both analysis and empirical data suggestthat the moditieation is most suitable when training data is sparse. In the second att… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
22
0

Year Published

1996
1996
2020
2020

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 30 publications
(22 citation statements)
references
References 6 publications
0
22
0
Order By: Relevance
“…As another, more detailed, example, in Rosenfeld and Huang (1992) we report on our early work on trigger models. We used a trigger utility measure, closely related to mutual information, to select some 620 000 triggers.…”
Section: Linear Interpolationmentioning
confidence: 99%
“…As another, more detailed, example, in Rosenfeld and Huang (1992) we report on our early work on trigger models. We used a trigger utility measure, closely related to mutual information, to select some 620 000 triggers.…”
Section: Linear Interpolationmentioning
confidence: 99%
“…That is, it is more likely that the triggered sequence appears after its trigger. (Lau, Rosenfeld, & Roukos, 1993;Rosenfeld, 1994;Rosenfeld & Huang, 1992) have implemented this trigger pair based approach, leading to an improvement in the performance on Large Vocabulary Continuous Speech Recognition systems adapted to different topics.…”
Section: Previous Workmentioning
confidence: 99%
“…This can be done by using an interpolation, by enhancing the probabilities of the corresponding n-gram when a word is triggered, etc. (Rosenfeld & Huang, 1992). Such models usually take into account only relationships between two words, although they can be applied to longer triggers, which could lead to more accurate models (for the same reason a higher order Markov model is more accurate).…”
Section: Trigger Modelmentioning
confidence: 99%
“…This selection is usually performed using a fixed size sliding window. Once the triggers have been selected, they are used to refine n-gram models (Chen & Chan, 2003;Rosenfeld & Huang, 1992). However, several triggers may match the history and have to be combined.…”
Section: Trigger Modelmentioning
confidence: 99%
See 1 more Smart Citation