2021
DOI: 10.3390/electronics10060689
|View full text |Cite
|
Sign up to set email alerts
|

On-Device Deep Learning Inference for System-on-Chip (SoC) Architectures

Abstract: As machine learning becomes ubiquitous, the need to deploy models on real-time, embedded systems will become increasingly critical. This is especially true for deep learning solutions, whose large models pose interesting challenges for target architectures at the “edge” that are resource-constrained. The realization of machine learning, and deep learning, is being driven by the availability of specialized hardware, such as system-on-chip solutions, which provide some alleviation of constraints. Equally importa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 25 publications
(30 reference statements)
0
2
0
Order By: Relevance
“…DNNs are structures that can be massively parallelized. This property has been exploited through the construction of a specific System on Chip (SoC) [37] that can execute GFLOPS and TFLOPS [38] keeping controlled both the consumption and the clock frequency necessary for the operation of the chip. In terms of system scalability, the neural models implemented are trained by focusing on their ability to generalize, seeking to maintain its quality under different scenarios making a massive deployment more transparent.…”
Section: A Fundamentals Of Deep Neural Networkmentioning
confidence: 99%
“…DNNs are structures that can be massively parallelized. This property has been exploited through the construction of a specific System on Chip (SoC) [37] that can execute GFLOPS and TFLOPS [38] keeping controlled both the consumption and the clock frequency necessary for the operation of the chip. In terms of system scalability, the neural models implemented are trained by focusing on their ability to generalize, seeking to maintain its quality under different scenarios making a massive deployment more transparent.…”
Section: A Fundamentals Of Deep Neural Networkmentioning
confidence: 99%
“…Lastly, a DSP is an auxiliary processor specialized in processing numerical operations at high speed and repeatedly performing multiply-accumulate operations [3], so it is appropriate for accelerating the inference of DNNs in edge devices. Usually, a DSP is mounted on a system-on-chip (SoC) along with a CPU and a GPU [4].…”
Section: Introductionmentioning
confidence: 99%