Skip to content

Commit acef54f

Browse files
docs: Update Frontpage README with new features (#731)
Signed-off-by: Sahil Jain <[email protected]> Signed-off-by: Sahil Jain <[email protected]> Signed-off-by: Terry Kong <[email protected]> Co-authored-by: Terry Kong <[email protected]> Co-authored-by: Terry Kong <[email protected]>
1 parent 9ed38d5 commit acef54f

File tree

1 file changed

+11
-7
lines changed

1 file changed

+11
-7
lines changed

README.md

Lines changed: 11 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -37,29 +37,33 @@ What you can expect:
3737
- **Comprehensive documentation** that is both detailed and user-friendly, with practical examples.
3838

3939
## 📣 News
40+
* [7/25/2025] [Release v0.3.0!](https://github.com/NVIDIA-NeMo/RL/releases/tag/v0.3.0)
41+
* 📊 View the release run metrics on [Google Colab](https://colab.research.google.com/drive/15kpesCV1m_C5UQFStssTEjaN2RsBMeZ0?usp=sharing) to get a head start on your experimentation.
4042
* [5/14/2025] [Reproduce DeepscaleR with NeMo RL!](docs/guides/grpo-deepscaler.md)
4143
* [5/14/2025] [Release v0.2.1!](https://github.com/NVIDIA-NeMo/RL/releases/tag/v0.2.1)
4244
* 📊 View the release run metrics on [Google Colab](https://colab.research.google.com/drive/1o14sO0gj_Tl_ZXGsoYip3C0r5ofkU1Ey?usp=sharing) to get a head start on your experimentation.
4345

4446
## Features
4547

46-
_Available now_ | 🔜 _Coming in v0.3_
48+
_Available now_ | 🔜 _Coming in v0.4_
4749

4850
-**Fast Generation** - vLLM backend for optimized inference.
49-
-**HuggingFace Integration** - Works with 1-32B models (Qwen2.5, Llama).
50-
-**Distributed Training** - Fully Sharded Data Parallel (FSDP) support and Ray-based infrastructure.
51+
-**HuggingFace Integration** - Works with 1-70B models (Qwen, Llama).
52+
-**Distributed Training** - Fully Sharded Data Parallel (FSDP2) support and Ray-based infrastructure.
5153
-**Environment Support** - Support for multi-environment training.
5254
-**Learning Algorithms** - GRPO (Group Relative Policy Optimization), SFT (Supervised Fine-Tuning), and DPO (Direct Preference Optimization).
5355
-**Multi-Turn RL** - Multi-turn generation and training for RL with tool use, games, etc.
54-
-**Large Model Support** - Native PyTorch support for models up to 32B parameters.
55-
-**Advanced Parallelism** - PyTorch native FSDP2, TP, and SP for efficient training.
56+
-**Large Model Support** - Native PyTorch support for models up to 70B parameters.
57+
-**Advanced Parallelism** - PyTorch native FSDP2, TP, CP, and SP for efficient training.
58+
-**(even) Larger Model Support with Long(er) Sequences** - Advanced parallelisms with Megatron Core (TP/PP/CP/SP/EP).
5659
-**Worker Isolation** - Process isolation between RL Actors (no worries about global state).
5760
-**Environment Isolation** - Dependency isolation between components.
58-
-**(even) Larger Model Support with Long(er) Sequence** - Support advanced parallelism in training with Megatron Core.
5961
-**Megatron Inference** - (static) Megatron Inference for day-0 support for new megatron models.
62+
-**MoE Models** - Support for DeepseekV3 and Qwen-3 MoE models
63+
-**Sequence Packing** - Sequence packing in both DTensor and MCore for huge training perf gains
64+
6065

6166
- 🔜 **Improved Native Performance** - Improve training time for Native Pytorch Models.
62-
- 🔜 **MoE Models** - Support DeepseekV3 and Llama4.
6367
- 🔜 **Megatron Inference** - (dynamic) Megatron Inference for fast day-0 support for new megatron models.
6468

6569
## Prerequisites

0 commit comments

Comments
 (0)