Timezone: »

Memory-Efficient Pipeline-Parallel DNN Training
Deepak Narayanan · Amar Phanishayee · Kaiyu Shi · Xie Chen · Matei Zaharia

Thu Jul 22 05:20 AM -- 05:25 AM (PDT) @

Many state-of-the-art ML results have been obtained by scaling up the number of parameters in existing models. However, parameters and activations for such large models often do not fit in the memory of a single accelerator device; this means that it is necessary to distribute training of large models over multiple accelerators. In this work, we propose PipeDream-2BW, a system that supports memory-efficient pipeline parallelism. PipeDream-2BW uses a novel pipelining and weight gradient coalescing strategy, combined with the double buffering of weights, to ensure high throughput, low memory footprint, and weight update semantics similar to data parallelism. In addition, PipeDream-2BW automatically partitions the model over the available hardware resources, while respecting hardware constraints such as memory capacities of accelerators and interconnect topologies. PipeDream-2BW can accelerate the training of large GPT and BERT language models by up to 20x with similar final model accuracy.

Author Information

Deepak Narayanan (Stanford)
Amar Phanishayee (Microsoft Research)
Kaiyu Shi (AISpeech)
Xie Chen (Microsoft)
Matei Zaharia (Stanford and Databricks)

Related Events (a corresponding poster, oral, or spotlight)

More from the Same Authors