Training a Trillion Parameters with Pipeline Parallelism

DeepSpeed includes new support for pipeline parallelism! DeepSpeed’s training engine provides hybrid 3D parallelism for training models with over a trillion parameters. In addition to scaling to the extreme, we have demonstrated that hybrid parallelism accelerates training on clusters with low-bandwidth network by up to 7x.

  • For a brief overview and results including trillion-parameter capabilities, see our press release.
  • To get started with pipeline parallel training in DeepSpeed, we recommend our tutorial.
  • See our AlexNet example in DeepSpeedExamples.
  • Read our API documentation on readthedocs.

Updated: