2015 IEEE International Conference on Computer Vision (ICCV) 2015
DOI: 10.1109/iccv.2015.123
|View full text |Cite
|
Sign up to set email alerts
|

Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification

Abstract: Rectified activation units (rectifiers) are essential for state-of-the-art neural networks. In this work, we study rectifier neural networks for image classification from two aspects. First, we propose a Parametric Rectified Linear Unit (PReLU) that generalizes the traditional rectified unit. PReLU improves model fitting with nearly zero extra computational cost and little overfitting risk. Second, we derive a robust initialization method that particularly considers the rectifier nonlinearities. This method en… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

19
9,063
3
65

Year Published

2017
2017
2023
2023

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 15,592 publications
(9,748 citation statements)
references
References 34 publications
19
9,063
3
65
Order By: Relevance
“…Recent state-of-the-art neural networks have increased depth, ranging from 16 (Simonyan & Zisserman, 2014) to 152 (He, Zhang, Ren, & Sun, 2015) layers (combined with some architectural advances). While the brain is clearly not shallow, its depth is limited to substantially fewer computational layers considering feed-forward processing (Lamme & Roelfsema, 2000).…”
Section: Box 1 Deep Neural Networkmentioning
confidence: 99%
“…Recent state-of-the-art neural networks have increased depth, ranging from 16 (Simonyan & Zisserman, 2014) to 152 (He, Zhang, Ren, & Sun, 2015) layers (combined with some architectural advances). While the brain is clearly not shallow, its depth is limited to substantially fewer computational layers considering feed-forward processing (Lamme & Roelfsema, 2000).…”
Section: Box 1 Deep Neural Networkmentioning
confidence: 99%
“…We selected PReLU as the activation function [20]. It introduces a very small number of parameters on the basis of the ReLU [21].…”
Section: Going Faster and Preventing Overfittingmentioning
confidence: 99%
“…From the possible batch sizes of (16,20,24,32,64), 32 was selected in view of the performance of our graphics processing unit (GPU) and test accuracy. The time limit of the initial training epochs was 400, but on the basis of our prior experimental results the best precision was reached within 80 epochs, so the number of epochs used during training was only 80.…”
Section: Experimental Setup For Classification Of Labeled Pixelsmentioning
confidence: 99%
“…The reported results for visual concept classification [9,12,15,23] indicate that human performance is (far) better than the respective state of the art for automated visual concept classification at that time. Although He et al claimed in 2015 that human-level performance has been surpassed by their approach [8], this claim remains questionable since only two human annotators were involved in the underlying study of Russakovsky et al [18]. There are also some methodological issues in the reported experiments of the other studies, for example, experimental settings are not well defined, the employment of crowdsourcing is critical, or the number of images is too small which prevents drawing conclusions for rare classes.…”
Section: Human and Machine Performance In Visual And Auditory Recognimentioning
confidence: 57%
“…Russakovsky et al compared this submission with two human annotators and discovered that the neuronal network outperformed one of them. He et al [8] claimed their system to be the first one that surpassed human-level performance (5.1%) on ImageNet data by achieving an error rate of 4.94%.…”
Section: Human and Machine Performance In Visual And Auditory Recognimentioning
confidence: 99%