2017
DOI: 10.48550/arxiv.1703.05160
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

A New Unbiased and Efficient Class of LSH-Based Samplers and Estimators for Partition Function Computation in Log-Linear Models

Abstract: Log-linear models are arguably the most successful class of graphical models for large-scale applications because of their simplicity and tractability. Learning and inference with these models require calculating the partition function, which is a major bottleneck and intractable for large state spaces. Importance Sampling (IS) and MCMC-based approaches are lucrative. However, the condition of having a "good" proposal distribution is often not satisfied in practice.In this paper, we add a new dimension to effi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
21
0

Year Published

2017
2017
2020
2020

Publication Types

Select...
6
1

Relationship

5
2

Authors

Journals

citations
Cited by 13 publications
(21 citation statements)
references
References 6 publications
0
21
0
Order By: Relevance
“…In this paper, we change this. Our work provides a truly constant time adaptive sampling scheme utilizing the recent advances in Locality Sensitive Sampling [14,15]. More impressively, we provide an efficient implementation of our proposal on CPU, which outperforms TensorFlow's implementation of softmax and other negative sampling strategies on some of the best available GPUs (V100) in terms of wall-clock training time.…”
Section: Negative Samplingmentioning
confidence: 93%
See 1 more Smart Citation
“…In this paper, we change this. Our work provides a truly constant time adaptive sampling scheme utilizing the recent advances in Locality Sensitive Sampling [14,15]. More impressively, we provide an efficient implementation of our proposal on CPU, which outperforms TensorFlow's implementation of softmax and other negative sampling strategies on some of the best available GPUs (V100) in terms of wall-clock training time.…”
Section: Negative Samplingmentioning
confidence: 93%
“…In this section, we briefly describe the recent development of using locality sensitive hashing for sampling and estimation [14,15,16,17]. Locality Sensitive Hashing [18,19] is a widely used paradigm for large scale similarity search and nearest neighbor search.…”
Section: Lsh Based Hash Tablesmentioning
confidence: 99%
“…Most existing works for reducing the softmax inference complexity are based on post-approximation of a fixed softmax that has been trained in a standard procedure. Locality Sensitive Hashing (LSH) has been demonstrated as a powerful technique under this category [9,[15][16][17]. Small word graph is another powerful technique for this problem [11].…”
Section: Related Workmentioning
confidence: 99%
“…These methods operate offline since efficient adaptive sampling on streaming data is a challenging problem. Recently, locality-sensitive hashing has been used as a fast adaptive sampler for the KDE problem [5,35]. In particular, the hashing-based estimator (HBE) introduced by [5] has strong theoretical guarantees for KDE, even in high dimensions.…”
Section: Related Workmentioning
confidence: 99%
“…While LSH was originally introduced for the high dimensional nearest-neighbor search problem, the technique has also recently been applied to unbiased statistical estimation via adaptive sampling for a variety of functions [5,35]. Our KDE method will use the RACE algorithm, which views LSH as a slightly different kind of statistical estimator [25].…”
Section: Repeated Array-of-counts Estimator (Race)mentioning
confidence: 99%