50th International Conference on Parallel Processing 2021
DOI: 10.1145/3472456.3472497
|View full text |Cite
|
Sign up to set email alerts
|

Hippie: A Data-Paralleled Pipeline Approach to Improve Memory-Efficiency and Scalability for Large DNN Training

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1
1

Relationship

1
6

Authors

Journals

citations
Cited by 13 publications
(2 citation statements)
references
References 19 publications
0
2
0
Order By: Relevance
“…To tackle the issue mentioned above, a model named Hippie was proposed [26]. Hippie is a hybrid parallel training framework that combines pipeline and data parallelism to increase the memory economy and scalability of massive DNN training.…”
Section: Related Workmentioning
confidence: 99%
“…To tackle the issue mentioned above, a model named Hippie was proposed [26]. Hippie is a hybrid parallel training framework that combines pipeline and data parallelism to increase the memory economy and scalability of massive DNN training.…”
Section: Related Workmentioning
confidence: 99%
“…DAPPLE [12] and Megatron-LM [29] modify the pipeline schedule and reduce the usage of activation memory. Synchronous PMP works suffer from bubbles, Hippie [48] utilizes bubbles for communications with half step delayed parameter updating, and Chimera [22] uses a bidirectional pipeline to reduce the bubble but doubles the model memory.…”
Section: Background and Related Workmentioning
confidence: 99%