Proceedings of the 48th International Conference on Parallel Processing 2019
DOI: 10.1145/3337821.3337873
|View full text |Cite
|
Sign up to set email alerts
|

Cynthia

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2020
2020
2025
2025

Publication Types

Select...
5
1
1

Relationship

1
6

Authors

Journals

citations
Cited by 37 publications
(4 citation statements)
references
References 20 publications
0
4
0
Order By: Relevance
“…Inference (17) Training (10) Both (1) Hardware accelerator (13) Algorithm co-design (8) New numeric formats ( 6) FIGURE 6. Included papers by categories.…”
Section: ) Pruningmentioning
confidence: 99%
See 1 more Smart Citation
“…Inference (17) Training (10) Both (1) Hardware accelerator (13) Algorithm co-design (8) New numeric formats ( 6) FIGURE 6. Included papers by categories.…”
Section: ) Pruningmentioning
confidence: 99%
“…However, distributed training is less costefficient than training using one machine. Distributed training involves parameter synchronization overhead and can lead to under-utilized cloud VMs waiting for parameter updates from each other [13]. Likewise, the combination of increased inference frequency and near real-time requirement can be resolved by the provision of more cloud VMs.…”
Section: Introductionmentioning
confidence: 99%
“…Both DNN training and inference are computation intensive. Efficient DNN training has been extensively studied [33], [34], [35], [36], [37], [38] and we focus on DNN inference in this paper. DNN-based mobile applications typically impose strict SLA requirements such as constrained end-toend latency [8], [22], [39].…”
Section: Deep Learningmentioning
confidence: 99%
“…[25] shows how to do cluster resource allocation and scheduling for ML training jobs by developing and using an empirical performance model to determine number of workers and parameter servers to use. Similarly, Cynthia [26] uses an analytical performance model for cost efficient cloud resource provisioning. In contrast, our approach can directly start training without the need for apriori modeling.…”
Section: Resnetmentioning
confidence: 99%