2022
DOI: 10.48550/arxiv.2202.06054
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

When do Models Generalize? A Perspective from Data-Algorithm Compatibility

Abstract: Benign overfitting demonstrates that overparameterized models can perform well on test data while fitting noisy training data. However, it only considers the final min-norm solution in linear regression, which ignores the algorithm information and the corresponding training procedure. In this paper, we generalize the idea of benign overfitting to the whole training trajectory instead of the min-norm solution and derive a time-variant bound based on the trajectory analysis. Starting from the timevariant bound, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 11 publications
0
1
0
Order By: Relevance
“…Overparameterized linear model. There has been several recent progress in theoretical understanding of overparameterized linear model under different scenarios, where the main goal is to provide non-asymptotic generalization guarantees, such as studies of linear regression [5], ridge regression [36], constant-stepsize SGD [44], decaying-stepsize SGD [41], GD [42], Gaussian Mixture models [39]. This paper aims to derive the non-asymptotic excess risk bound for MAML under mixed linear model, which can be independent of data dimension d and still converge as the iteration T enlarges.…”
Section: Related Workmentioning
confidence: 99%
“…Overparameterized linear model. There has been several recent progress in theoretical understanding of overparameterized linear model under different scenarios, where the main goal is to provide non-asymptotic generalization guarantees, such as studies of linear regression [5], ridge regression [36], constant-stepsize SGD [44], decaying-stepsize SGD [41], GD [42], Gaussian Mixture models [39]. This paper aims to derive the non-asymptotic excess risk bound for MAML under mixed linear model, which can be independent of data dimension d and still converge as the iteration T enlarges.…”
Section: Related Workmentioning
confidence: 99%