2023
DOI: 10.1109/tpds.2023.3281931
|View full text |Cite
|
Sign up to set email alerts
|

A Survey on Auto-Parallelism of Large-Scale Deep Learning Training

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(2 citation statements)
references
References 50 publications
0
2
0
Order By: Relevance
“…The open-source movement has democratized access to cutting-edge tools, with platforms such as TensorFlow [38] and PyTorch [39] leading the way. This democratization has led to the emergence of MLOps, a practice that seeks to standardize and automate the AI model lifecycle [40,41].…”
Section: Open-source Architectures and Mlopsmentioning
confidence: 99%
“…The open-source movement has democratized access to cutting-edge tools, with platforms such as TensorFlow [38] and PyTorch [39] leading the way. This democratization has led to the emergence of MLOps, a practice that seeks to standardize and automate the AI model lifecycle [40,41].…”
Section: Open-source Architectures and Mlopsmentioning
confidence: 99%
“…Recently, in order to eliminate the complexity of model parallelism, Auto Parallelism is being studied. Auto Parallelism automatically implements model parallelism by considering the hardware and model structure at the system level [22]. This shows novel parallelism performance in all general-purpose deep-learning models, but it is necessary to design model parallelism directly to secure resource use efficiency in large-scale artificial intelligence models.…”
Section: Related Workmentioning
confidence: 99%