2020 IEEE High Performance Extreme Computing Conference (HPEC) 2020
DOI: 10.1109/hpec43674.2020.9286182
|View full text |Cite
|
Sign up to set email alerts
|

Beyond Floating-Point Ops: CNN Performance Prediction with Critical Datapath Length

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2

Relationship

1
5

Authors

Journals

citations
Cited by 7 publications
(3 citation statements)
references
References 14 publications
0
3
0
Order By: Relevance
“…The FLOPs is a widely used indicator to measure the computational complexity of the model. The entire number of network operations that can be summed up into a single floating‐point hardware operation is the definition of FLOPs (Langerman et al., 2020; Molchanov et al., 2016). The original ResNeSt is much lighter than Res2Net due to fewer parameters (37.28 M) and computations (211.15G).…”
Section: Resultsmentioning
confidence: 99%
“…The FLOPs is a widely used indicator to measure the computational complexity of the model. The entire number of network operations that can be summed up into a single floating‐point hardware operation is the definition of FLOPs (Langerman et al., 2020; Molchanov et al., 2016). The original ResNeSt is much lighter than Res2Net due to fewer parameters (37.28 M) and computations (211.15G).…”
Section: Resultsmentioning
confidence: 99%
“…Previous works examining the relationship between efficiency measures showed that different cost indicators do not correlate well with each other during neural network training (Dehghani et al, 2021). In particular, it has been hypothesized that discrepancies between FLOPs and wallclock inference latency is primarily are primarily compute bounded by kernel execution or memory-bound by data movement as opposed to framework bottlenecks (Langerman et al, 2020). These previous works have largely focused on convolutional neural networks (CNNs) in computer vision.…”
Section: Efficiency Metrics and Cost Indicatorsmentioning
confidence: 99%
“…Transformers have been shown to be more efficient in terms of inference time when compared to convolutional networks [56]. This effect is not directly correlated with the number of parameters, but is rather more influenced by the network structure [79].…”
Section: Inference Timementioning
confidence: 99%