As models get larger, to ensure they fit into device memory, libraries and techniques must be enabled to help reduce the memory size.
This webinar provides:
- The basic steps to enable Microsoft DeepSpeed* on Intel Gaudi AI accelerators
- Information on how ZeRO1 and ZeRO2 memory optimizers and activation checkpointing can be used to reduce memory use on a large model.
- Instructions on how to use Intel Gaudi AI software APIs to detect the peak memory of any model and provide guidance on when to use these techniques.
- A Q&A session.
Additional Resources
- Documentation
- DeepSpeed User Guide
- System access:
- Model Repository
- User Forum