2017
DOI: 10.1093/imaiai/iax015
|View full text |Cite
|
Sign up to set email alerts
|

Sketching for large-scale learning of mixture models

Abstract: Learning parameters from voluminous data can be prohibitive in terms of memory and computational requirements. We propose a "compressive learning" framework where we estimate model parameters from a sketch of the training data. This sketch is a collection of generalized moments of the underlying probability distribution of the data. It can be computed in a single pass on the training set, and is easily computable on streams or distributed datasets. The proposed framework shares similarities with compressive se… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
148
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
4
3
1

Relationship

2
6

Authors

Journals

citations
Cited by 53 publications
(151 citation statements)
references
References 80 publications
0
148
0
Order By: Relevance
“…To recover the centroids C from y, the state-of-the-art algorithm is compressed learning via orthogonal matching pursuit with replacement (CL-OMPR) [5,6]. It aims to solve arg min…”
Section: A Sketched Clusteringmentioning
confidence: 99%
See 2 more Smart Citations
“…To recover the centroids C from y, the state-of-the-art algorithm is compressed learning via orthogonal matching pursuit with replacement (CL-OMPR) [5,6]. It aims to solve arg min…”
Section: A Sketched Clusteringmentioning
confidence: 99%
“…That is, {x t } T t=1 are assumed to be drawn i.i.d. from the GMM distribution (5). To recover the centroids C [c 1 , .…”
Section: A High-dimensional Inference Frameworkmentioning
confidence: 99%
See 1 more Smart Citation
“…Note that this type of error is often considered when introducing quantization [7,19]. Ideally, one has η = 0, however in some cases it can be considerably simpler to prove that the LRIP holds with a non-zero η [20]. The reader would note that the classical RIP is often expressed with a constant α = (1 − t) −1 where t < 1 is a small as possible.…”
Section: Deterministic Operatormentioning
confidence: 99%
“…Interestingly, κ also defines a Reproducing Kernel Hilbert Space in which two probability density functions (pdfs) can be compared with a Maximum Mean Discrepancy (MMD) metric [22][23][24][25][26]. Equipped with the MMD metric, the Generalized Method of Moments [27] in (3) is equivalent to an infinite-dimensional Compressed Sensing [11] problem, where the "sparse" pdf underlying the data (e.g., approximated by few Diracs) is reconstructed from a small number of compressive, random linear pdf measurements: the sketch [26]. The method to solve (3) is thus inspired by the OMP(R) CS recovery algorithm, i.e., Orthogonal Matching Pursuit (with Replacement) [28,29].…”
Section: Introductionmentioning
confidence: 99%