Timezone: »
Many state-of-the-art ML results have been obtained by scaling up the number of parameters in existing models. However, parameters and activations for such large models often do not fit in the memory of a single accelerator device; this means that it is necessary to distribute training of large models over multiple accelerators. In this work, we propose PipeDream-2BW, a system that supports memory-efficient pipeline parallelism. PipeDream-2BW uses a novel pipelining and weight gradient coalescing strategy, combined with the double buffering of weights, to ensure high throughput, low memory footprint, and weight update semantics similar to data parallelism. In addition, PipeDream-2BW automatically partitions the model over the available hardware resources, while respecting hardware constraints such as memory capacities of accelerators and interconnect topologies. PipeDream-2BW can accelerate the training of large GPT and BERT language models by up to 20x with similar final model accuracy.
Author Information
Deepak Narayanan (Stanford)
Amar Phanishayee (Microsoft Research)
Kaiyu Shi (AISpeech)
Xie Chen (Microsoft)
Matei Zaharia (Stanford and Databricks)
Related Events (a corresponding poster, oral, or spotlight)
-
2021 Spotlight: Memory-Efficient Pipeline-Parallel DNN Training »
Thu. Jul 22nd 12:20 -- 12:25 PM Room
More from the Same Authors
-
2021 : Have the Cake and Eat It Too? Higher Accuracy and Less Expense when Using Multi-label ML APIs Online »
Lingjiao Chen · James Zou · Matei Zaharia -
2021 : Machine Learning API Shift Assessments: Change is Coming! »
Lingjiao Chen · James Zou · Matei Zaharia -
2023 : Improve Model Inference Cost with Image Gridding »
Shreyas Krishnaswamy · Lisa Dunlap · Lingjiao Chen · Matei Zaharia · James Zou · Joseph Gonzalez -
2022 : What Can Data-Centric AI Learn from Data Engineering? »
Matei Zaharia -
2022 Workshop: Knowledge Retrieval and Language Models »
Maithra Raghu · Urvashi Khandelwal · Chiyuan Zhang · Matei Zaharia · Alexander Rush -
2022 Poster: Efficient Online ML API Selection for Multi-Label Classification Tasks »
Lingjiao Chen · Matei Zaharia · James Zou -
2022 Spotlight: Efficient Online ML API Selection for Multi-Label Classification Tasks »
Lingjiao Chen · Matei Zaharia · James Zou -
2021 Poster: Boosting the Throughput and Accelerator Utilization of Specialized CNN Inference Beyond Increasing Batch Size »
Jack Kosaian · Amar Phanishayee · Matthai Philipose · Debadeepta Dey · Rashmi Vinayak -
2021 Spotlight: Boosting the Throughput and Accelerator Utilization of Specialized CNN Inference Beyond Increasing Batch Size »
Jack Kosaian · Amar Phanishayee · Matthai Philipose · Debadeepta Dey · Rashmi Vinayak -
2020 Poster: The Non-IID Data Quagmire of Decentralized Machine Learning »
Kevin Hsieh · Amar Phanishayee · Onur Mutlu · Phillip Gibbons -
2019 Poster: LIT: Learned Intermediate Representation Training for Model Compression »
Animesh Koratana · Daniel Kang · Peter Bailis · Matei Zaharia -
2019 Oral: LIT: Learned Intermediate Representation Training for Model Compression »
Animesh Koratana · Daniel Kang · Peter Bailis · Matei Zaharia