2017 IEEE International Symposium on Circuits and Systems (ISCAS) 2017
DOI: 10.1109/iscas.2017.8050341
|View full text |Cite
|
Sign up to set email alerts
|

Towards closing the energy gap between HOG and CNN features for embedded vision

Abstract: Computer vision enables a wide range of applications in robotics/drones, self-driving cars, smart Internet of Things, and portable/wearable electronics. For many of these applications, local embedded processing is preferred due to privacy and/or latency concerns. Accordingly, energy-efficient embedded vision hardware delivering real-time and robust performance is crucial. While deep learning is gaining popularity in several computer vision algorithms, a significant energy consumption difference exists compared… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
30
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
4
2
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 46 publications
(30 citation statements)
references
References 22 publications
0
30
0
Order By: Relevance
“…In our view, these limitations can be easily overcome using one of the following: (a) training of multiple object detectors, one for each position or (b) through the use of CNN‐based detectors, such as Ren et al (). While CNN‐based detectors are insensitive to changes in orientation, it is worth pointing out that such algorithms require a larger training dataset (more images), higher hardware specifications, are less portable (file size), and require more fine‐tuning (Suleiman et al, ).…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…In our view, these limitations can be easily overcome using one of the following: (a) training of multiple object detectors, one for each position or (b) through the use of CNN‐based detectors, such as Ren et al (). While CNN‐based detectors are insensitive to changes in orientation, it is worth pointing out that such algorithms require a larger training dataset (more images), higher hardware specifications, are less portable (file size), and require more fine‐tuning (Suleiman et al, ).…”
Section: Discussionmentioning
confidence: 99%
“…They also require large training samples (one to two orders of magnitude higher than the one proposed here), more fine‐tuning of training parameters, specialized hardware (e.g. graphics processing unit, GPU), and usually perform at lower image processing rates (images per second) than simpler models (Suleiman, Chen, Emer, & Sze, ).…”
Section: Methodsmentioning
confidence: 99%
“…While the current state of the art in object detection relies on convolutional neural networks (CNN, e.g., Ren et al, 2015), these can be overpowered for standard biological applications. They also require large training samples (one to two orders of magnitude higher than the one proposed here), more fine-tuning of training parameters, specialized hardware (e.g., GPU), and usually perform at lower image processing rates (images per second) than simpler models (Suleiman et al, 2017).…”
Section: Training Object Detectorsmentioning
confidence: 99%
“…In our view, these limitations can be overcome using one of the following: (1) training of multiple object detectors, one for each position or (2) through the use of CNN-based detectors, such as Ren et al (2015). While CNN based detectors are insensitive to changes in orientation, it is worth pointing out that such algorithms require a larger training dataset (more images), higher hardware specifications, are less portable (file size), and require more fine tuning (Suleiman et al, 2017).…”
Section: Limitations Of the Frameworkmentioning
confidence: 99%
“…is is particularly important in certain cases such as cameras on-board UAVs where he altitudes that UAVs y/hover, and the resulting resolution of objects in the image, make the detection of small objects even more challenging. However, increasing the CNN receptive eld makes the learning process di cult since the problem dimensionality is increased and at the same time the computational cost also increases exponentially [9]. To tackle this issue it is necessary to reduce the data that needs to be processed from the original input image.…”
Section: Introductionmentioning
confidence: 99%