2020
DOI: 10.48550/arxiv.2006.16423
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Efficient Algorithms for Device Placement of DNN Graph Operators

Abstract: Modern machine learning workloads use large models, with complex structures, that are very expensive to execute. The devices that execute complex models are becoming increasingly heterogeneous as we see a flourishing of domain-specific accelerators being offered as hardware accelerators in addition to CPUs. These trends necessitate distributing the workload across multiple devices. Recent work has shown that significant gains can be obtained with model parallelism, i.e, partitioning a neural network's computat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 6 publications
0
1
0
Order By: Relevance
“…FlexFlow [22] uses automatic search to discover the best operator parallelization strategy in the graph. Building on this direction of auto-parallelization, these recent papers [39,60] use optimal synthesis and reinforcement learning to find optimized device placement to further improve parallelism without the need for manual intervention. However, these general systems are not specifically designed for highly sparse recommendation models.…”
Section: Related Workmentioning
confidence: 99%
“…FlexFlow [22] uses automatic search to discover the best operator parallelization strategy in the graph. Building on this direction of auto-parallelization, these recent papers [39,60] use optimal synthesis and reinforcement learning to find optimized device placement to further improve parallelism without the need for manual intervention. However, these general systems are not specifically designed for highly sparse recommendation models.…”
Section: Related Workmentioning
confidence: 99%