2022 IEEE High Performance Extreme Computing Conference (HPEC) 2022
DOI: 10.1109/hpec55821.2022.9926290
|View full text |Cite
|
Sign up to set email alerts
|

Towards Fast GPU-based Sparse DNN Inference: A Hybrid Compute Model

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(1 citation statement)
references
References 16 publications
0
1
0
Order By: Relevance
“…• Edge Devices: Recently edge platforms (mobile phones, wearables, IoT sensors) have also become feasible deployment targets for DNNs, owing to advances in research on DNN architectures that maintain accuracy while reducing resource consumption (such as quantized (Wu et al, 2016) and sparse networks (Xu et al, 2018)). These advances notwithstanding, those platforms reproduce many of the phenomena of their larger scale analogues (Suo et al, 2021).…”
Section: Discussionmentioning
confidence: 99%
“…• Edge Devices: Recently edge platforms (mobile phones, wearables, IoT sensors) have also become feasible deployment targets for DNNs, owing to advances in research on DNN architectures that maintain accuracy while reducing resource consumption (such as quantized (Wu et al, 2016) and sparse networks (Xu et al, 2018)). These advances notwithstanding, those platforms reproduce many of the phenomena of their larger scale analogues (Suo et al, 2021).…”
Section: Discussionmentioning
confidence: 99%