2022
DOI: 10.3390/electronics11152316
|View full text |Cite
|
Sign up to set email alerts
|

AI-Driven Performance Modeling for AI Inference Workloads

Abstract: Deep Learning (DL) is moving towards deploying workloads not only in cloud datacenters, but also to the local devices. Although these are mostly limited to inference tasks, it still widens the range of possible target architectures significantly. Additionally, these new targets usually come with drastically reduced computation performance and memory sizes compared to the traditionally used architectures—and put the key optimization focus on the efficiency as they often depend on batteries. To help developers q… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
1
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
1
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 44 publications
0
1
0
Order By: Relevance
“…To solve the above problems, some researchers came up with a kernel additive method; they predict each kernel operation, such as convolution, dense, and LSTM, individually and sum up all kernel values to predict the overall performance of the DL model [9,16,19,21,23,25]. Yu et al [24] used the wave-scaling technique to predict the inference latency of the DL model on GPU, but this technique requires access to a GPU in order to make the prediction.…”
Section: Related Workmentioning
confidence: 99%
“…To solve the above problems, some researchers came up with a kernel additive method; they predict each kernel operation, such as convolution, dense, and LSTM, individually and sum up all kernel values to predict the overall performance of the DL model [9,16,19,21,23,25]. Yu et al [24] used the wave-scaling technique to predict the inference latency of the DL model on GPU, but this technique requires access to a GPU in order to make the prediction.…”
Section: Related Workmentioning
confidence: 99%
“…None of these tools are specifically made for compiling code to microcontroller targets, but many of them support microcontroller chips. Sponner et al have done a good review and a benchmark about these tools targeting embedded platforms in [155]. From these tools the TVM project is probably the most interesting.…”
Section: Edge Ai Software For Microcontrollersmentioning
confidence: 99%