2018
DOI: 10.48550/arxiv.1808.03578
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Dropout is a special case of the stochastic delta rule: faster and more accurate deep learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 0 publications
0
2
0
Order By: Relevance
“…Dropout in a U-Net architecture can perform as a special case of the delta rule in which we introduce noise in the transmission of information [38] by randomly masking weights of the network. Dropout is presented as an especial case of delta rule called stochastic delta rule [39] in which each weight in the model is assigned as a random variable from a Gaussian distribution with the mean µ w ij and standard deviation of σ w ij [38]. Dropout, as an special case of stochastic delta rule, introduces a form of regularization that aids in escaping poor local minima.…”
Section: U-net With Dropoutmentioning
confidence: 99%
See 1 more Smart Citation
“…Dropout in a U-Net architecture can perform as a special case of the delta rule in which we introduce noise in the transmission of information [38] by randomly masking weights of the network. Dropout is presented as an especial case of delta rule called stochastic delta rule [39] in which each weight in the model is assigned as a random variable from a Gaussian distribution with the mean µ w ij and standard deviation of σ w ij [38]. Dropout, as an special case of stochastic delta rule, introduces a form of regularization that aids in escaping poor local minima.…”
Section: U-net With Dropoutmentioning
confidence: 99%
“…By randomly deactivating a subset of neurons during each training iteration, dropout prevents the network from relying too heavily on specific neurons or features. This selective deactivation encourages the remaining neurons to compensate and learn more robust representations, leading to a broader exploration of the weight space and increasing the odds of finding the optimum solution [38]. Additionally, keeping the dropout in the inference process will introduce stochasticity by generating results from a randomly selected sub-network and will result in an approximation of posterior distribution [40].…”
Section: U-net With Dropoutmentioning
confidence: 99%