Accelerating Large-Scale Mixture-of-Experts Training in PyTorch with NeMo Automodel #777
Pinned
bernardwin
announced in
Announcements
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Training large-scale Mixture-of-Experts (MoE) models efficiently is hard. NVIDIA NeMo Automodel makes it easy with accelerated performance, production-ready recipes and reproducible benchmarks for popular MoE architectures.
Learn how this open-source library combines native PyTorch parallelisms with optimizations like NVIDIA Transformer Engine and DeepEP:
https://developer.nvidia.com/blog/accelerating-large-scale-mixture-of-experts-training-in-pytorch/
Beta Was this translation helpful? Give feedback.
All reactions