Microsoft DeepSpeed* Optimization on Large Models

Browse upcoming events and watch past recordings.

author-image

By

As models get larger, to ensure they fit into device memory, libraries and techniques must be enabled to help reduce the memory size.

This webinar provides:

  • The basic steps to enable Microsoft DeepSpeed* on Intel Gaudi AI accelerators
  • Information on how ZeRO1 and ZeRO2 memory optimizers and activation checkpointing can be used to reduce memory use on a large model.
  • Instructions on how to use Intel Gaudi AI software APIs to detect the peak memory of any model and provide guidance on when to use these techniques.
  • A Q&A session.

Additional Resources

Featured Speakers
 

picture of greg serochi

Greg Serochi

Developer advocate and applications engineering