2021
DOI: 10.48550/arxiv.2102.12919
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Distribution-Free Robust Linear Regression

Jaouad Mourtada,
Tomas Vaškevičius,
Nikita Zhivotovskiy

Abstract: We study random design linear regression with no assumptions on the distribution of the covariates and with a heavy-tailed response variable. When learning without assumptions on the covariates, we establish boundedness of the conditional second moment of the response variable as a necessary and sufficient condition for achieving deviation-optimal excess risk rate of convergence. In particular, combining the ideas of truncated least squares, median-of-means procedures and aggregation theory, we construct a non… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
3

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 60 publications
(156 reference statements)
0
3
0
Order By: Relevance
“…We will use the following classical estimator (Györfi et al, 2002, Thm 11.3); it is quoted here with an improved bound which is proven in (Mourtada et al, 2021 (w ⊤ (x, 1) − f * (x)) 2 dQ(x).…”
Section: A Simplified Version Of Our Estimatormentioning
confidence: 99%
“…We will use the following classical estimator (Györfi et al, 2002, Thm 11.3); it is quoted here with an improved bound which is proven in (Mourtada et al, 2021 (w ⊤ (x, 1) − f * (x)) 2 dQ(x).…”
Section: A Simplified Version Of Our Estimatormentioning
confidence: 99%
“…The paper of Hardt, Recht, and Singer [19] on the stability of gradient descent methods has generated a wave of interest in this direction. Recent works use various notions of stability in their analysis: some authors are motivated by the analysis of gradient descent algorithms [31,29,15,4], while others use the notion of average stability to obtain the in-expectation O(1/n) rate for regularized regression [28,17,46] and some more specific improper learning procedures [36,37]. One of the key open questions left in [19] is related to the lack of high probability generalization bounds.…”
Section: Introductionmentioning
confidence: 99%
“…We first observe that an important downside of the small method is the necessity to have Q 2ξ (E, a) > 0, i.e, the random vector a must satisfy a small ball condition, otherwise the conclusion of the theorem is empty, the right hand side becomes negative. The small ball condition has been analysed in a recent line of work dedicated to remove such condition in many different problems in mathematical statistics and mathematical signal processing [14,19,21,25]. Indeed, the estimates provided by the small ball method quickly deteriorates when Q 2ξ (E, a) is not an absolute constant.…”
Section: Introductionmentioning
confidence: 99%