2022
DOI: 10.1016/j.inffus.2021.11.005
|View full text |Cite
|
Sign up to set email alerts
|

A comprehensive survey on regularization strategies in machine learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
57
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 141 publications
(58 citation statements)
references
References 83 publications
1
57
0
Order By: Relevance
“…There are numerous strategies to tackle this issue, such as early stopping (stop the training before the model incorporates noise); expand the training set (more data implies a more accurate model); feature selection (this identifies the most relevant features to be learned, ignoring redundant ones); ensemble methods (that aggregate the output of a set of classifiers, selecting the best output by a voting process); and, finally, regularisation (that, in general, limits the amount of variance in the model by penalising input parameters with large coefficients). Regularisation methods have received great attention from recent studies [ 61 ], since they are solely related to the algorithms, and not to data quality or classifier competitions. One particular regularisation method worth mentioning in this review is Dropout , that ignores randomly chosen neurons (with a certain probability) during the training phase, so that a reduced network is obtained as a result.…”
Section: Background Knowledgementioning
confidence: 99%
“…There are numerous strategies to tackle this issue, such as early stopping (stop the training before the model incorporates noise); expand the training set (more data implies a more accurate model); feature selection (this identifies the most relevant features to be learned, ignoring redundant ones); ensemble methods (that aggregate the output of a set of classifiers, selecting the best output by a voting process); and, finally, regularisation (that, in general, limits the amount of variance in the model by penalising input parameters with large coefficients). Regularisation methods have received great attention from recent studies [ 61 ], since they are solely related to the algorithms, and not to data quality or classifier competitions. One particular regularisation method worth mentioning in this review is Dropout , that ignores randomly chosen neurons (with a certain probability) during the training phase, so that a reduced network is obtained as a result.…”
Section: Background Knowledgementioning
confidence: 99%
“…Batch size defines the number of inputs that will be propagated through the network each time. Batch normalization, as one of the common regularization strategies, aims to deal with noise data, the limited size of the training data, and the complexity of classifiers to avoid overfitting [49]. Using a smaller batch size requires less memory and results in faster training; however, setting the batch size too small will result in less accuracy for the estimate of the gradient.…”
Section: Hyperparameter Tuningmentioning
confidence: 99%
“…The (2, 1)-norm favors a small number of nonzero rows in the matrix W , therefore ensuring that the common features (most effective centers) will be selected. It should be noted that, Regularization techniques [10,11] proved to improve the generalization ability and therefore the performance of a model. A comprehensive study and a state-of-the-art review of the regularization strategies in machine learning is given in [10].…”
Section: Introductionmentioning
confidence: 99%
“…It should be noted that, Regularization techniques [10,11] proved to improve the generalization ability and therefore the performance of a model. A comprehensive study and a state-of-the-art review of the regularization strategies in machine learning is given in [10]. It is being used in different classification problems such as, image recognition [12], Underwater Acoustic Data Classification [11] e.t.c.…”
Section: Introductionmentioning
confidence: 99%