2020
DOI: 10.48550/arxiv.2006.00840
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Universal Robust Regression via Maximum Mean Discrepancy

Abstract: Many datasets are collected automatically, and are thus easily contaminated by outliers. In order to overcome this issue, there was recently a regain of interest in robust estimation methods. However, most of these methods are designed for a specific purpose, such as estimation of the mean, or linear regression. We propose estimators based on Maximum Mean Discrepancy (MMD) optimization as a universal framework for robust regression. We provide non-asymptotic error bounds, and show that our estimators are robus… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
15
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
3
2

Relationship

2
3

Authors

Journals

citations
Cited by 5 publications
(15 citation statements)
references
References 19 publications
0
15
0
Order By: Relevance
“…The dataset consists of m = 2 16 points, from which a minibatch of 2 10 points is sampled at random at every iteration. Depending on the considered experiment, either n samples are generated using MC or n, n 3 4 , n 2 3 or n 1 2 are simulated using RQMC at every iteration. The optimisation algorithm is run for 3,000 iterations for every setting.…”
Section: Inference For Bivariate Beta Distributionsmentioning
confidence: 99%
See 2 more Smart Citations
“…The dataset consists of m = 2 16 points, from which a minibatch of 2 10 points is sampled at random at every iteration. Depending on the considered experiment, either n samples are generated using MC or n, n 3 4 , n 2 3 or n 1 2 are simulated using RQMC at every iteration. The optimisation algorithm is run for 3,000 iterations for every setting.…”
Section: Inference For Bivariate Beta Distributionsmentioning
confidence: 99%
“…As could be reasonably expected, the RQMC-based estimator with n points is significantly more expensive than a MC with n points, but it is also much more accurate in l 2 error. Similarly, the RQMC-based estimators with n 2 3 or n 1 2 are less accurate in l 2 error but usually cheaper than MC with n points. More interestingly, we see that the RQMC estimator with n 3 4 points is both cheaper and more accurate than the MC estimator with n points for the Wasserstein and Sinkhorn divergences.…”
Section: Inference For Bivariate Beta Distributionsmentioning
confidence: 99%
See 1 more Smart Citation
“…See [13,19] for the two-sample test problem, e.g. As a tool for parametric estimation, even though it was implicitly used in specific examples in machine learning [15], MMD has been studied as a general method for inference only recently [2,4,10,11]. In the latter papers, it appeared that MMD criteria lead to consistent estimators that are unconditionally robust to model misspecification.…”
Section: Contextmentioning
confidence: 99%
“…Z 2 = z 2 is N (sz 2 , 1 − s 2 ), we can easily calculate ψ(z) := E exp(−Z 2 1 /(γ 2 /2))|Z 2 = z . Indeed, setting τ 2 := {1/γ 2 + 1/(1 − s 2 )} −1 , we have…”
mentioning
confidence: 99%

Estimation of copulas via Maximum Mean Discrepancy

Alquier,
Chérief-Abdellatif,
Derumigny
et al. 2020
Preprint
Self Cite