2018
DOI: 10.1214/18-ejs1485
|View full text |Cite
|
Sign up to set email alerts
|

Noise contrastive estimation: Asymptotic properties, formal comparison with MC-MLE

Abstract: A statistical model is said to be un-normalised when its likelihood function involves an intractable normalising constant. Two popular methods for parameter inference for these models are MC-MLE (Monte Carlo maximum likelihood estimation), and NCE (noise contrastive estimation); both methods rely on simulating artificial data-points to approximate the normalising constant. While the asymptotics of MC-MLE have been established under general hypotheses (Geyer, 1994), this is not so for NCE. We establish consiste… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 16 publications
0
3
0
Order By: Relevance
“…where where have used that for ν → ∞, m = νn becomes arbitrarily large too, so that the average over the y i becomes an expectation with respect to q(y). A more general result was established by Riou-Durand and Chopin (2018). Moreover, they further considered the case of finite ν, but large sample sizes n, and showed that the variance of the noise-contrastive estimator is always smaller than the variance of the Monte Carlo MLE estimator (Geyer, 1994) where the partition function is approximated with a sample average, assuming in both cases that the auxiliary/reference distributions were fixed.…”
Section: Estimating Energy-based Modelsmentioning
confidence: 90%
“…where where have used that for ν → ∞, m = νn becomes arbitrarily large too, so that the average over the y i becomes an expectation with respect to q(y). A more general result was established by Riou-Durand and Chopin (2018). Moreover, they further considered the case of finite ν, but large sample sizes n, and showed that the variance of the noise-contrastive estimator is always smaller than the variance of the Monte Carlo MLE estimator (Geyer, 1994) where the partition function is approximated with a sample average, assuming in both cases that the auxiliary/reference distributions were fixed.…”
Section: Estimating Energy-based Modelsmentioning
confidence: 90%
“…, and taking the limit of → ∞ in (38), thus, gives where where have used that for → ∞ , m = n becomes arbitrarily large too, so that the average over the i becomes an expectation with respect to q( ) . A more general result was established by Riou-Durand and Chopin (2018). Moreover, they further considered the case of finite , but large sample sizes n, and showed that the variance of the noise-contrastive estimator is always smaller than the variance of the Monte Carlo MLE estimator (Geyer 1994) where the partition function is approximated with a sample average, assuming in both cases that the auxiliary/reference distributions were fixed.…”
Section: Estimating Energy-based Modelsmentioning
confidence: 91%
“…With respect to a fixed Q, it remains an open question about what formally are the nature of the challenges posed by a poorly chosen Q, which could be statistical and/or algorithmic. Various previous works have analyzed the asymptotic behavior of NCE and its variants (Gutmann & Hyvärinen, 2012;Riou-Durand et al, 2018;Uehara et al, 2020), but these do not provide guidance on the finite sample behavior of NCE or its common variants. The improvements to NCE in prior works are all borne out by the empirical observations of NCE practitioners, rather than motivated by theory, which is precisely the aim of this work.…”
Section: Related Workmentioning
confidence: 99%