2018
DOI: 10.1109/access.2018.2877890
|View full text |Cite
|
Sign up to set email alerts
|

Benchmark Analysis of Representative Deep Neural Network Architectures

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

13
354
1
7

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 681 publications
(375 citation statements)
references
References 12 publications
13
354
1
7
Order By: Relevance
“…4) Effect of the feature extractor: The effect of the feature extractor for Faster R-CNN is very limited on the AP, except for a high value of IoU threshold (0.9) on the Stanford dataset, as can be seen in Figure 19 and Figure 20. Nevertheless, in terms of inference speed, the Inception-v2 feature extractor is significantly faster than Resnet50 (Figures 21 and 22), which is consistent with the findings of Bianco et al [31] who also showed that Inception-v2 (aka BN-inception) is less computationally complex. 5) Effect of the input size: Figures 21 and 22 show a significant gain in YOLOv3's AP when moving from a 320x320 input size to 416x416, but the performance stagnates when we move further to 608x608, which means that the 416x416 resolution is sufficient to detect the objects of the two datasets.…”
supporting
confidence: 87%
See 1 more Smart Citation
“…4) Effect of the feature extractor: The effect of the feature extractor for Faster R-CNN is very limited on the AP, except for a high value of IoU threshold (0.9) on the Stanford dataset, as can be seen in Figure 19 and Figure 20. Nevertheless, in terms of inference speed, the Inception-v2 feature extractor is significantly faster than Resnet50 (Figures 21 and 22), which is consistent with the findings of Bianco et al [31] who also showed that Inception-v2 (aka BN-inception) is less computationally complex. 5) Effect of the input size: Figures 21 and 22 show a significant gain in YOLOv3's AP when moving from a 320x320 input size to 416x416, but the performance stagnates when we move further to 608x608, which means that the 416x416 resolution is sufficient to detect the objects of the two datasets.…”
supporting
confidence: 87%
“…On the other hand, the main hyperparameter for Faster R-CNN is the feature extractor. We tested two different feature extractors: Inception-v2 [30] (also called BN-inception in the literature [31]) and Resnet50 [32]. These settings make a total of 5 classifiers that we trained and tested on the two datasets described above, which amounts to 10 experiments that we summarize in Table VI.…”
Section: B Hyperparametersmentioning
confidence: 99%
“…The incorporation of specific platform constraints to such approaches involves modeling how the network architecture relates with the optimization target. As a first step for modeling the performance of embedded CNNs, recent studies have carried out systematic benchmarking on several hardware systems [6,[27][28][29]. Gaining in specificity, an energy estimation methodology for CNN accelerators has been introduced in [30,31].…”
Section: Related Workmentioning
confidence: 99%
“…We also compare to well established standard CNN architectures [29,28,13,37,26]. For the accuracies, FLOPs and parameters of standard CNNs, we use the benchmark analysis of Bianco et al [1]. Compared to other standard CNN architecures, the accuracy of our networks are superior to MobileNetv2 [26], GoogleNet [29] and VGG [28].…”
Section: Multi-stage Shiftingmentioning
confidence: 99%