1995
DOI: 10.1109/34.476512
|View full text |Cite
|
Sign up to set email alerts
|

On the estimation of 'small' probabilities by leaving-one-out

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
64
0

Year Published

1996
1996
2015
2015

Publication Types

Select...
3
3
3

Relationship

1
8

Authors

Journals

citations
Cited by 66 publications
(64 citation statements)
references
References 13 publications
0
64
0
Order By: Relevance
“…Our framework then enriches the user language model with social network information. We select four popular smoothing methods to demonstrate such effect, namely additive smoothing, absolute smoothing (Ney et al, 1995), Jelinek-Mercer smoothing (Jelinek and Mercer, 1980) and Dirichlet smoothing (MacKay and Peto, 1994). The results of using only the base model (i.e.…”
Section: Baseline Methodsmentioning
confidence: 99%
“…Our framework then enriches the user language model with social network information. We select four popular smoothing methods to demonstrate such effect, namely additive smoothing, absolute smoothing (Ney et al, 1995), Jelinek-Mercer smoothing (Jelinek and Mercer, 1980) and Dirichlet smoothing (MacKay and Peto, 1994). The results of using only the base model (i.e.…”
Section: Baseline Methodsmentioning
confidence: 99%
“…For this purpose, one should note that, in this kind of deterministic models, the likelihood of the training sample is maximized if the stochastic model assigns to every tree t in the sample a probability equal to its relative frequency in Ω [8]. So, these probabilities must be calculated as the ratio of the number of occurrences of a transition to the number of occurrences of the state to which this transition leads.…”
Section: Stochastic K-testable Tree Modelsmentioning
confidence: 99%
“…Therefore, the use of smoothing techniques becomes necessary if one wants to use these models for parsing. Two classical techniques of this type are linear interpolation and backing-off [8].…”
Section: Smoothingmentioning
confidence: 99%
“…We include the plain smoothing of Additive (also known as Add-δ) smoothing and Absolute Discounting decrease the probability of seen words by subtracting a constant (Ney et al, 1995). We also implement several classic strategies smoothed from the whole collection as background information: Jelinek-Mercer (J-M) applies a linear interpolation, and Dirichlet employs a prior on collection influence .…”
Section: Algorithms For Comparisonmentioning
confidence: 99%