2009
DOI: 10.1118/1.3213517
|View full text |Cite
|
Sign up to set email alerts
|

Noise injection for training artificial neural networks: A comparison with weight decay and early stopping

Abstract: The purpose of this study was to investigate the effect of a noise injection method on the "overfitting" problem of artificial neural networks (ANNs) in two-class classification tasks. The authors compared ANNs trained with noise injection to ANNs trained with two other methods for avoiding overfitting: weight decay and early stopping. They also evaluated an automatic algorithm for selecting the magnitude of the noise injection. They performed simulation studies of an exclusive-or classification task with trai… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

3
73
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
9

Relationship

0
9

Authors

Journals

citations
Cited by 164 publications
(76 citation statements)
references
References 30 publications
3
73
0
Order By: Relevance
“…Random Noise injection to the weights or the hidden units has been utilized in many neural network researches for many years [15], [19]- [21]. Kurita et al [9] injected noise into the hidden layers of an MLP and showed the network ability to get automatically structurized by simply adding the noises and therefore improving the generalization ability of the network.…”
Section: Related Workmentioning
confidence: 99%
“…Random Noise injection to the weights or the hidden units has been utilized in many neural network researches for many years [15], [19]- [21]. Kurita et al [9] injected noise into the hidden layers of an MLP and showed the network ability to get automatically structurized by simply adding the noises and therefore improving the generalization ability of the network.…”
Section: Related Workmentioning
confidence: 99%
“…Therefore, to avoid over-fitting in the proposed RNN model, three standard techniques were used. These were Gaussian noise injection into the training data [32], using a ReLU activation function in the hidden layers [20], and subsequently applying the dropout technique [29]. Dropout leads to big improvements in the prediction performance of the model.…”
Section: Mitigating Overfittingmentioning
confidence: 99%
“…The first element of the inverse-forward ANNs is the inverse ANN which is trained with noisy data and is responsible to filter the noise, while the second element is the forward ANN which eliminates the necessity of FEM. Training with noise-free FEM data of the inverse ANN would result in overfitting and its subsequent failure (Arbabi et al, 2015a;Zur et al, 2009). The inverse ANN trained with noisy data (Gaussian random noise) is most sensitive to the general trend of the experimental data without being influenced by small deviations from FEM caused by uncertainties involved in the experimental data.…”
Section: Discussionmentioning
confidence: 99%
“…Our previous study showed that ANNs are very sensitive to any deviations from the underlying computational model that is used for their training (Arbabi et al, 2015a). To alleviate this problem, the training data of ANNs can be contaminated with some level of noise (Arbabi et al, 2015a;Derks et al, 1997;Zur et al, 2009) to increase the robustness of ANN. Therefore, we trained the ANN using the input concentration vs. time curves contaminated with different levels of Gaussian noise, i.e.…”
Section: Inverse-forward Artificial Neural Networkmentioning
confidence: 99%