Proceedings of the 20th ACM Conference on Embedded Networked Sensor Systems 2022
DOI: 10.1145/3560905.3568300
|View full text |Cite
|
Sign up to set email alerts
|

Ultra-Low Power DNN Accelerators for IoT

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2022
2022
2025
2025

Publication Types

Select...
6

Relationship

0
6

Authors

Journals

citations
Cited by 9 publications
(6 citation statements)
references
References 16 publications
0
6
0
Order By: Relevance
“…Giordano et al [47] benchmark a single architecture for image classification on several different platforms. Moss et al [48] evaluate different image classification architectures on a single platform, MAX78000. Unlike these works, we describe the full deployment pipeline in the context of object detection, from architecture exploration to quantization and hardware-optimized implementation.…”
Section: Related Workmentioning
confidence: 99%
“…Giordano et al [47] benchmark a single architecture for image classification on several different platforms. Moss et al [48] evaluate different image classification architectures on a single platform, MAX78000. Unlike these works, we describe the full deployment pipeline in the context of object detection, from architecture exploration to quantization and hardware-optimized implementation.…”
Section: Related Workmentioning
confidence: 99%
“…However, when similar models are executed using such MCU integrating multiple convolutional engines, we can expect the same latency baseline of Google Coral, although lower in magnitude. Recent works in the literature show that the MAX78000 base execution units typically require two-dimensional inputs and in case less computing resources are required, data is filled with zeros, thus resulting in a latency baseline mostly independent from the operation size (for instance, a network with 4 input channels, 3 × 3 kernel size, padding of one, and 4 to 64 output channels, leads to a constant baseline latency of ∼ 150 µs, by far larger than SPLVP) [18]. The MLP models investigated in this work have a small size (the biggest Avila model counts 270 neurons).…”
Section: ) Other Acceleratorsmentioning
confidence: 99%
“…In general, complex accelerators designed to support large Convolutional Neural Networks (CNNs) typically provide substantial latency for very small models because the internal logic is typically underutilized; moreover, they require interfacing with middle-end processors and operating systems. For instance, in the MAX78000 SoC, convolutional and linear filters with input data size of 16 × 16 provide a latency of ∼ 75 µs, irrespective of filter size, while a single two-dimensional convolutional layer with four output channels requires ∼ 150 µs, irrespective of the number of input channels [18]. Similarly, the Google Edge TPU platform suffers from extreme underutilization of its Processing Elements (PEs) and inefficient sequential scheduling of Fully Connected (FC) layers [19].…”
Section: Introductionmentioning
confidence: 99%
“…The generated attention map, AttMap l , is fed into a classifier along with the feature map of the current video frame F l i+j in order to detect any semantic variation. The classifier component represented by the function Z Class is a classifier with two outputs, which is parameterized by θ Class in equation (7). It generates a class label using the attention map AttMap l and the feature map from the current video frame F l i+j .…”
Section: B Temporal Early Exit Module (Teem)mentioning
confidence: 99%
“…Object detection in static images has achieved remarkable successes in recent years using CNNs [3]. However, beyond individual images, video object detection has emerged as a new challenge, particularly when deployed on various embedded devices with limited computation and energy resources [4], [5], [6], [7]. This is due to the high computational cost introduced by applying existing image object detection networks in real-time on numerous individual video frames [8].…”
Section: Introductionmentioning
confidence: 99%