2017
DOI: 10.48550/arxiv.1703.09580
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Early Stopping without a Validation Set

Abstract: Early stopping is a widely used technique to prevent poor generalization performance when training an over-expressive model by means of gradient-based optimization. To find a good point to halt the optimizer, a common practice is to split the dataset into a training and a smaller validation set to obtain an ongoing estimate of the generalization performance. We propose a novel early stopping criterion based on fast-to-compute local statistics of the computed gradients and entirely removes the need for a held-o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
25
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
3
3

Relationship

0
6

Authors

Journals

citations
Cited by 12 publications
(25 citation statements)
references
References 12 publications
0
25
0
Order By: Relevance
“…The central idea behind early stopping (stop training or optimization) [23] is that there exists a critical regime during the training of a learning model where the model ceases to generalize (perform better) on unseen data points while being able to do improve performance on given training data. Identifying this point of negative or zero return is also attractive from a computational perspective and is the goal of various early stopping rules or methods in machine learning [24], [25], [26]. A conventional and widely popular early stopping method in machine learning is the one based on validation data, which we name as Validation-based method.…”
Section: Early Stopping Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…The central idea behind early stopping (stop training or optimization) [23] is that there exists a critical regime during the training of a learning model where the model ceases to generalize (perform better) on unseen data points while being able to do improve performance on given training data. Identifying this point of negative or zero return is also attractive from a computational perspective and is the goal of various early stopping rules or methods in machine learning [24], [25], [26]. A conventional and widely popular early stopping method in machine learning is the one based on validation data, which we name as Validation-based method.…”
Section: Early Stopping Methodsmentioning
confidence: 99%
“…A conventional and widely popular early stopping method in machine learning is the one based on validation data, which we name as Validation-based method. Although very effective in practice, especially with large training datasets where holding off a small part of the training data has no effect in the learning process, there are drawbacks to Validation-based early stopping [26]. The validation performance may have a large stochastic error depending on the size of the validation set and may introduce biases leading to poor generalization estimates.…”
Section: Early Stopping Methodsmentioning
confidence: 99%
See 3 more Smart Citations