2022
DOI: 10.1109/tpami.2021.3139612
|View full text |Cite
|
Sign up to set email alerts
|

Fast and Accurate Least-Mean-Squares Solvers for High Dimensional Data

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
52
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
6
3
1

Relationship

1
9

Authors

Journals

citations
Cited by 26 publications
(52 citation statements)
references
References 23 publications
0
52
0
Order By: Relevance
“…learner encoder url kdda kddb kdd12 For small d, this opens up sketching approaches to learning whose superlinearity in input dimension otherwise makes them inaccessible to wide, sparse datasets, such as coreset construction for nearest-neighbor queries [28]. Further, in the case of linear models, an n × d design can be represented faithfully as a (d + 1) × d one by Carathéodory's Theorem [29], which for small d can greatly simplify linear learning (e.g., tuning regularization parameters no longer requires multiple passes over the data).…”
Section: Discussionmentioning
confidence: 99%
“…learner encoder url kdda kddb kdd12 For small d, this opens up sketching approaches to learning whose superlinearity in input dimension otherwise makes them inaccessible to wide, sparse datasets, such as coreset construction for nearest-neighbor queries [28]. Further, in the case of linear models, an n × d design can be represented faithfully as a (d + 1) × d one by Carathéodory's Theorem [29], which for small d can greatly simplify linear learning (e.g., tuning regularization parameters no longer requires multiple passes over the data).…”
Section: Discussionmentioning
confidence: 99%
“…Other related work considers generalizations of k-median and k-means by either adding capacity constraints [7,29,52,85], generalizing the notion of centers to subspaces [19,41,42], time series [54] or sets [61] or considering more general objective functions [5,15]. Coresets have also been studied for many other problems: we cite non-comprehensively decision trees [60], kernel methods [59,62,83], determinant maximization [57], diversity maximization [58], shape fitting problems [2,22], linear regression [9,53,88], logistic regression [56,81], Gaussian mixtures [70], dependency networks [79], or low-rank approximation [71]. The interested reader is referred to [3,40,80] and similar surveys for more pointers to coreset literature.…”
Section: Overview Of Our Techniquesmentioning
confidence: 99%
“…In the past years, coreset techniques have been widely applied to many optimization problems, such as: clustering [11,22,25], logistic regression [26,37,42,47,42], Bayesian methods [9,8], linear regression [17,19,12,27,47], robust optimization [18], Gaussian mixture model [32], and active learning [14,43]. Recently, [33] also proposed the notion of "accurate" coresets, which do not introduce any approximation error when compressing the input dataset. Coresets are also applied to speed up large-scale or distributed machine learning algorithms [41,34,35,6].…”
Section: Related Workmentioning
confidence: 99%