Microsoft DeepSpeed* Optimization on Large Models

Browse upcoming events and watch past recordings.

As models get larger, to ensure they fit into device memory, libraries and techniques must be enabled to help reduce the memory size.

This webinar provides:

  • The basic steps to enable Microsoft DeepSpeed* on Intel Gaudi AI accelerators
  • Information on how ZeRO1 and ZeRO2 memory optimizers and activation checkpointing can be used to reduce memory use on a large model.
  • Instructions on how to use Intel Gaudi AI software APIs to detect the peak memory of any model and provide guidance on when to use these techniques.
  • A Q&A session.

Additional Resources

Featured Speakers
 

picture of greg serochi

Greg Serochi

Developer advocate and applications engineering
 

 

1