2018
DOI: 10.2478/acss-2018-0007
|View full text |Cite
|
Sign up to set email alerts
|

Smooth Non-increasing Square Spatial Extents of Filters in Convolutional Layers of CNNs for Image Classification Problems

Abstract: The present paper considers an open problem of setting hyperparameters for convolutional neural networks aimed at image classification. Since selecting filter spatial extents for convolutional layers is a topical problem, it is approximately solved by accumulating statistics of the neural network performance. The network architecture is taken on the basis of the MNIST database experience. The eight-layered architecture having four convolutional layers is nearly best suitable for classifying small and medium si… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2018
2018
2019
2019

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(5 citation statements)
references
References 41 publications
0
5
0
Order By: Relevance
“…The represented method of generating an infinitely scalable dataset relies on pseudorandomization of the polygon vertices' number by (1) and coordinates of the vertices by (2) -(4). Inequalities (5) and (6) help in making a polygon of an appropriate form and size, unless the polygon is a triangle. On rare occasions, the triangle can be generated very small or thin reminding an arrow (see Fig.…”
Section: Discussionmentioning
confidence: 99%
See 3 more Smart Citations
“…The represented method of generating an infinitely scalable dataset relies on pseudorandomization of the polygon vertices' number by (1) and coordinates of the vertices by (2) -(4). Inequalities (5) and (6) help in making a polygon of an appropriate form and size, unless the polygon is a triangle. On rare occasions, the triangle can be generated very small or thin reminding an arrow (see Fig.…”
Section: Discussionmentioning
confidence: 99%
“…So, the question is whether it is possible to test segmentation network architectures much faster in order to find optimal solutions that could be imparted to real-world semantic image segmentation tasks. Such solutions are the number of convolutional layers and their parameters (the size and number of filters) [5,8], max pooling layers [9], parameters of deconvolutional layers (the size and number of filters), and, probably, dropout layers [10]. For example, a plausible purpose is to research on training data of smaller and simpler images so that real-world tasks could inherit close-to-optimal network architectures from them.…”
Section: Problem Statementmentioning
confidence: 99%
See 2 more Smart Citations
“…Therefore, a pretrained CNN to be included into the Faster R-CNN object detector should not contain a DropOut layer, especially if the CNN is trained much longer with the DropOut layer. An exclusion may be for very simple original image classification datasets (like MNIST [13,14] or EEACL26 [2,4,5]), where overfitting is prevented with two or even more DropOut layers and they do not retard the training process.…”
Section: Discussionmentioning
confidence: 99%