2021
DOI: 10.1109/tnnls.2020.3026784
|View full text |Cite
|
Sign up to set email alerts
|

Revisiting Internal Covariate Shift for Batch Normalization

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
13
0
1

Year Published

2021
2021
2023
2023

Publication Types

Select...
7
2
1

Relationship

1
9

Authors

Journals

citations
Cited by 63 publications
(26 citation statements)
references
References 8 publications
1
13
0
1
Order By: Relevance
“…The strengthening of Ioffe and Szegedy's argument is obtained from a study by (Awais et al, 2021). A series of experiments in this study show that ICS reduction is a major factor in increasing the convergence of a deep learning network, not only by batch normalization but also by all other methods that contribute to ICS reduction.…”
Section: Plant Classificationsupporting
confidence: 56%
“…The strengthening of Ioffe and Szegedy's argument is obtained from a study by (Awais et al, 2021). A series of experiments in this study show that ICS reduction is a major factor in increasing the convergence of a deep learning network, not only by batch normalization but also by all other methods that contribute to ICS reduction.…”
Section: Plant Classificationsupporting
confidence: 56%
“…The number of possible implementations of model architectures is also large enough that finding the perfect working model for a problem cannot be achieved. Models can be improved by the addition or removal of layers and different activation functions within the layers, as well as by utilizing regularization techniques such as batch normalization [ 34 ] or dropout. The deep learning techniques employed in this research are based on many evaluations and modifications to the architectures, as well as an examination of the related literature.…”
Section: Resultsmentioning
confidence: 99%
“…It comprises of 7 Conv-ReLU-BN blocks, where Conv refers to the standard Convolution operation, ReLU refers to the Rectified Linear Activation layer and BN refers to Batch Normalization layer, used to keep the weights in within a threshold reducing internal covariate shift [45,46], thus stabilizing the training process. Dropout layers [47] were used prior to the last two dense layers, though it hasn't been represented in the Fig.…”
Section: Model Architecturementioning
confidence: 99%