2019
DOI: 10.1007/978-3-030-32520-6_12
|View full text |Cite
|
Sign up to set email alerts
|

Preventing Overfitting by Training Derivatives

Abstract: Derivative training is a well-known method to improve the accuracy of neural networks. In the forward pass, not only the output values are computed, but also their derivatives, and their deviations from the target derivatives are included in the cost function, which is minimized with respect to the weights by a gradient-based algorithm. So far, this method has been implemented for relatively low-dimensional tasks. In this study, we apply the approach to the problem of image analysis. We consider the task of re… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 69 publications
0
5
0
Order By: Relevance
“…Over-fitting and underfitting are common fitting problems for model training (Dahan and Keller 2021;Graham et al 2020;Yuan et al 2020). Over-fitting takes the local features of training data set as the whole features, and the reason is that too many parameters of model make the training error small and the test error large (Yildirim and Ozkale 2021;Avrutskiy 2020). When under-fitting occurs, the learning ability of model is weak and the basic features of training data set cannot be learned (Handelman et al 2019;Ahmed and Isa 2017;Van Calster and Vickers 2015).…”
Section: Evaluation Metrics For Machine Learningmentioning
confidence: 99%
“…Over-fitting and underfitting are common fitting problems for model training (Dahan and Keller 2021;Graham et al 2020;Yuan et al 2020). Over-fitting takes the local features of training data set as the whole features, and the reason is that too many parameters of model make the training error small and the test error large (Yildirim and Ozkale 2021;Avrutskiy 2020). When under-fitting occurs, the learning ability of model is weak and the basic features of training data set cannot be learned (Handelman et al 2019;Ahmed and Isa 2017;Van Calster and Vickers 2015).…”
Section: Evaluation Metrics For Machine Learningmentioning
confidence: 99%
“…The capacity of a model is defined by its ability to fit various functions. The model architecture, number of hidden layers, weight size, and hypothesis space size affect the capacity of the model [10,39,40]. In addition, weight decay is a term added to the objective function to penalize the high-order weights [10,27], In the early stopping technique, training should be stopped before weights become large to avoid hidden units being in their non-linear ranges, resulting in high capacity [29].…”
Section: Adapting Capacitymentioning
confidence: 99%
“…Data augmentation is an effective approach for magnifying small datasets by inverting, trimming, rotating, zooming, and adjusting brightness, sharpness, and contrast. These approaches rely on existing samples to create new samples from the entire training dataset [10,35].…”
Section: Data Enrichmentmentioning
confidence: 99%
“…The first idea of approximation of a differential equation by MLP‐NN was proposed by IE Lagaris with one hidden‐layer network . Later Avrutskiy develops the idea by using more than one hidden layer MLP‐NN . The basic idea of this paper is inspired by these works.…”
Section: Neural Network Approximation Of Fpementioning
confidence: 99%
“…In , the single hidden layer network is applied to find the solution of differential equations. Recently, more than one‐hidden‐layer structure network is also investigated for both enhancing the function approximation power of the neural network as well as finding the solution of PDEs . In a special form of the solution for differential equations was chosen to satisfy the initial and boundary conditions.…”
Section: Introductionmentioning
confidence: 99%