Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2014
DOI: 10.3115/v1/p14-1072
|View full text |Cite
|
Sign up to set email alerts
|

Kneser-Ney Smoothing on Expected Counts

Abstract: Widely used in speech and language processing, Kneser-Ney (KN) smoothing has consistently been shown to be one of the best-performing smoothing methods. However, KN smoothing assumes integer counts, limiting its potential uses-for example, inside Expectation-Maximization. In this paper, we propose a generalization of KN smoothing that operates on fractional counts, or, more precisely, on distributions over counts. We rederive all the steps of KN smoothing to operate on count distributions instead of integral c… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
33
0

Year Published

2015
2015
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 21 publications
(33 citation statements)
references
References 21 publications
0
33
0
Order By: Relevance
“…Recently, better estimation methods during the maximization step of EM have been proposed to avoid the over-fitting in WA, such as using Kneser-Ney Smoothing to back-off the expected counts (Zhang and Chiang, 2014) or integrating the smoothed l 0 prior to the estimation of probability (Vaswani et al, 2012). Our work differs from theirs by addressing the over-fitting directly in the EM algorithm by adopting a leave-one-out approach.…”
Section: Related Workmentioning
confidence: 99%
“…Recently, better estimation methods during the maximization step of EM have been proposed to avoid the over-fitting in WA, such as using Kneser-Ney Smoothing to back-off the expected counts (Zhang and Chiang, 2014) or integrating the smoothed l 0 prior to the estimation of probability (Vaswani et al, 2012). Our work differs from theirs by addressing the over-fitting directly in the EM algorithm by adopting a leave-one-out approach.…”
Section: Related Workmentioning
confidence: 99%
“…Namely, methods like the Witten-Bell smoothing [20] and the KneserNey smoothing [21] are still used in state-of-the-art work [22,23,24]. In general, these methods combine high-order models with lower-order ones.…”
Section: N-gram Smoothing Methodsmentioning
confidence: 99%
“…To avoid overfitting, we find that it is necessary to apply an expected smoothing approach in training. We choose expected Kneser-Ney smoothing technique (Zhang and Chiang, 2014) as it is simple and achieves state-of-the-art performance on the language modeling problem.…”
Section: Biasing Language Modelsmentioning
confidence: 99%