You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
<summary><b><ahref=https://huggingface.co/nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B-BF16>NVIDIA-Nemotron-3-Nano-30B-A3B</a> is out with full reproducible script and recipes! Checkout <ahref=https://github.com/NVIDIA-NeMo/Megatron-Bridge/tree/nano-v3>NeMo Megatron-Bridge</a>, <ahref=https://github.com/NVIDIA-NeMo/Automodel/blob/main/examples/llm_finetune/nemotron/nemotron_nano_v3_squad.yaml>NeMo AutoModel</a>, <ahref=https://github.com/NVIDIA-NeMo/RL>NeMo-RL</a> and <ahref=https://catalog.ngc.nvidia.com/orgs/nvidia/containers/nemo?version=25.11.nemotron_3_nano>NGC container</a> to try them!(2025-12-15)
16
+
<summary><b><ahref=https://huggingface.co/nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B-BF16>NVIDIA-Nemotron-3-Nano-30B-A3B</a> is out with full reproducible script and recipes! Check out <ahref=https://github.com/NVIDIA-NeMo/Megatron-Bridge/tree/nano-v3>NeMo Megatron-Bridge</a>, <ahref=https://github.com/NVIDIA-NeMo/AutoModel/blob/main/examples/llm_finetune/nemotron/nemotron_nano_v3_squad.yaml>NeMo AutoModel</a>, <ahref=https://github.com/NVIDIA-NeMo/RL>NeMo-RL</a> and <ahref=https://catalog.ngc.nvidia.com/orgs/nvidia/containers/nemo?version=25.11.nemotron_3_nano>NGC container</a> to try them!</b> (2025-12-15)
17
17
</details>
18
18
19
19
20
-
21
20
<detailsopen>
22
-
<summary><b>Pivot notice: This repo will pivot to focus on speech models collections only. Please refer to <ahref=https://github.com/NVIDIA-NeMo>NeMo Framework Github Org</a> for the complete list of repos under NeMo Framework</b></summary>
23
-
NeMo 2.0, with its support for LLMs and VLMs will be deprecated by 25.11, and replaced by <a href=https://github.com/NVIDIA-NeMo/Megatron-Bridge>NeMo Megatron-Bridge</a> and <a href=https://github.com/NVIDIA-NeMo/Automodel>NeMo Automodel</a>. More details can be find in the <a href=https://github.com/NVIDIA-NeMo>NeMo Framework github org readme</a>. (2025-10-10)
21
+
<summary><b>⚠️ Pivot notice: This repo will pivot to focus on speech models collections only. Please refer to <ahref=https://github.com/NVIDIA-NeMo>NeMo Framework Github Org</a> for the complete list of repos under NeMo Framework</b></summary>
22
+
NeMo 2.0, with its support for LLMs and VLMs will be deprecated by 25.11, and replaced by <a href=https://github.com/NVIDIA-NeMo/Megatron-Bridge>NeMo Megatron-Bridge</a> and <a href=https://github.com/NVIDIA-NeMo/AutoModel>NeMo AutoModel</a>. More details can be found in the <a href=https://github.com/NVIDIA-NeMo>NeMo Framework GitHub org readme</a>. (2025-10-10)
23
+
24
+
Following collections are deprecated and will be removed in a later release, please use previous versions if you are using them:
25
+
- nlp
26
+
- llm
27
+
- vlm
28
+
- vision
24
29
</details>
25
30
26
31
<detailsclosed>
27
32
<summary><b>Pretrain and finetune :hugs:Hugging Face models via AutoModel</b></summary>
28
-
Nemo Framework's latest feature AutoModel enables broad support for :hugs:Hugging Face models, with 25.04 focusing on
33
+
NeMo Framework's latest feature AutoModel enables broad support for :hugs:Hugging Face models, with 25.04 focusing on
29
34
30
35
31
36
- <a href=https://huggingface.co/transformers/v3.5.1/model_doc/auto.html#automodelforcausallm>AutoModelForCausalLM</a> in the <ahref="https://huggingface.co/models?pipeline_tag=text-generation&sort=trending">Text Generation</a> category
@@ -35,7 +40,7 @@ More Details in Blog: <a href=https://developer.nvidia.com/blog/run-hugging-face
35
40
</details>
36
41
37
42
<detailsclosed>
38
-
<summary><b>Training on Blackwell using Nemo</b></summary>
43
+
<summary><b>Training on Blackwell using NeMo</b></summary>
39
44
NeMo Framework has added Blackwell support, with <a href=https://docs.nvidia.com/nemo-framework/user-guide/latest/performance/performance_summary.html>performance benchmarks on GB200 & B200</a>. More optimizations to come in the upcoming releases.(2025-05-19)
40
45
</details>
41
46
@@ -82,7 +87,7 @@ More Details in Blog: <a href=https://developer.nvidia.com/blog/run-hugging-face
82
87
State-of-the-Art Multimodal Generative AI Model Development with NVIDIA NeMo
83
88
</a> (2024-11-06)
84
89
</summary>
85
-
NVIDIA recently announced significant enhancements to the NeMo platform, focusing on multimodal generative AI models. The update includes NeMo Curator and the Cosmos tokenizer, which streamline the data curation process and enhance the quality of visual data. These tools are designed to handle large-scale data efficiently, making it easier to develop high-quality AI models for various applications, including robotics and autonomous driving. The Cosmos tokenizers, in particular, efficiently map visual data into compact, semantic tokens, which is crucial for training large-scale generative models. The tokenizer is available now on the <a href=http://github.com/NVIDIA/cosmos-tokenizer/NVIDIA/cosmos-tokenizer>NVIDIA/cosmos-tokenizer</a> GitHub repo and on <a href=https://huggingface.co/nvidia/Cosmos-Tokenizer-CV8x8x8>Hugging Face</a>.
90
+
NVIDIA recently announced significant enhancements to the NeMo platform, focusing on multimodal generative AI models. The update includes NeMo Curator and the Cosmos tokenizer, which streamline the data curation process and enhance the quality of visual data. These tools are designed to handle large-scale data efficiently, making it easier to develop high-quality AI models for various applications, including robotics and autonomous driving. The Cosmos tokenizers, in particular, efficiently map visual data into compact, semantic tokens, which is crucial for training large-scale generative models. The tokenizer is available now on the <a href=https://github.com/NVIDIA/cosmos-tokenizer>NVIDIA/cosmos-tokenizer</a> GitHub repo and on <a href=https://huggingface.co/nvidia/Cosmos-Tokenizer-CV8x8x8>Hugging Face</a>.
86
91
<br><br>
87
92
</details>
88
93
<details>
@@ -216,22 +221,14 @@ NVIDIA NeMo 2.0 introduces several significant improvements over its predecessor
216
221
217
222
Overall, these enhancements make NeMo 2.0 a powerful, scalable, and user-friendly framework for AI model development.
218
223
219
-
> [!IMPORTANT]
220
-
> NeMo 2.0 is currently supported by the LLM (large language model) and VLM (vision language model) collections.
221
-
222
224
### Get Started with NeMo 2.0
223
225
224
226
- Refer to the [Quickstart](https://docs.nvidia.com/nemo-framework/user-guide/latest/nemo-2.0/quickstart.html) for examples of using NeMo-Run to launch NeMo 2.0 experiments locally and on a slurm cluster.
225
227
- For more information about NeMo 2.0, see the [NeMo Framework User Guide](https://docs.nvidia.com/nemo-framework/user-guide/latest/nemo-2.0/index.html).
226
-
-[NeMo 2.0 Recipes](https://github.com/NVIDIA/NeMo/blob/main/nemo/collections/llm/recipes) contains additional examples of launching large-scale runs using NeMo 2.0 and NeMo-Run.
227
228
- For an in-depth exploration of the main features of NeMo 2.0, see the [Feature Guide](https://docs.nvidia.com/nemo-framework/user-guide/latest/nemo-2.0/features/index.html#feature-guide).
228
229
- To transition from NeMo 1.0 to 2.0, see the [Migration Guide](https://docs.nvidia.com/nemo-framework/user-guide/latest/nemo-2.0/migration/index.html#migration-guide) for step-by-step instructions.
229
230
230
-
### Get Started with Cosmos
231
-
232
-
NeMo Curator and NeMo Framework support video curation and post-training of the Cosmos World Foundation Models, which are open and available on [NGC](https://catalog.ngc.nvidia.com/orgs/nvidia/teams/cosmos/collections/cosmos) and [Hugging Face](https://huggingface.co/collections/nvidia/cosmos-6751e884dc10e013a0a0d8e6). For more information on video datasets, refer to [NeMo Curator](https://developer.nvidia.com/nemo-curator). To post-train World Foundation Models using the NeMo Framework for your custom physical AI tasks, see the [Cosmos Diffusion models](https://github.com/NVIDIA/Cosmos/blob/main/cosmos1/models/diffusion/nemo/post_training/README.md) and the [Cosmos Autoregressive models](https://github.com/NVIDIA/Cosmos/blob/main/cosmos1/models/autoregressive/nemo/post_training/README.md).
233
-
234
-
## LLMs and MMs Training, Alignment, and Customization
231
+
## Training and Customization
235
232
236
233
All NeMo models are trained with
237
234
[Lightning](https://github.com/Lightning-AI/lightning). Training is
NeMo ASR and TTS models can be optimized for inference and deployed for
274
253
production use cases with [NVIDIA Riva](https://developer.nvidia.com/riva).
275
254
276
-
## NeMo Framework Launcher
277
-
278
-
> [!IMPORTANT]
279
-
> NeMo Framework Launcher is compatible with NeMo version 1.0 only. [NeMo-Run](https://github.com/NVIDIA/NeMo-Run) is recommended for launching experiments using NeMo 2.0.
280
-
281
-
[NeMo Framework
282
-
Launcher](https://github.com/NVIDIA/NeMo-Megatron-Launcher) is a
283
-
cloud-native tool that streamlines the NeMo Framework experience. It is
284
-
used for launching end-to-end NeMo Framework training jobs on CSPs and
285
-
Slurm clusters.
286
-
287
-
The NeMo Framework Launcher includes extensive recipes, scripts,
288
-
utilities, and documentation for training NeMo LLMs. It also includes
289
-
the NeMo Framework [Autoconfigurator](https://github.com/NVIDIA/NeMo-Megatron-Launcher#53-using-autoconfigurator-to-find-the-optimal-configuration),
290
-
which is designed to find the optimal model parallel configuration for
291
-
training on a specific cluster.
292
-
293
-
To get started quickly with the NeMo Framework Launcher, please see the
@@ -396,7 +351,7 @@ To install nemo_toolkit from such a wheel, use the following installation method
396
351
pip install "nemo_toolkit[all]"
397
352
```
398
353
399
-
If a more specific version is desired, we recommend a Pip-VCS install. From [NVIDIA/NeMo](github.com/NVIDIA/NeMo), fetch the commit, branch, or tag that you would like to install.
354
+
If a more specific version is desired, we recommend a Pip-VCS install. From [NVIDIA/NeMo](https://github.com/NVIDIA/NeMo), fetch the commit, branch, or tag that you would like to install.
400
355
To install nemo_toolkit from this Git reference `$REF`, use the following installation method:
401
356
402
357
```bash
@@ -415,18 +370,16 @@ following domain-specific commands:
415
370
```bash
416
371
pip install nemo_toolkit['all'] # or pip install "nemo_toolkit['all']@git+https://github.com/NVIDIA/NeMo@${REF:-'main'}"
417
372
pip install nemo_toolkit['asr'] # or pip install "nemo_toolkit['asr']@git+https://github.com/NVIDIA/NeMo@$REF:-'main'}"
418
-
pip install nemo_toolkit['nlp'] # or pip install "nemo_toolkit['nlp']@git+https://github.com/NVIDIA/NeMo@${REF:-'main'}"
419
373
pip install nemo_toolkit['tts'] # or pip install "nemo_toolkit['tts']@git+https://github.com/NVIDIA/NeMo@${REF:-'main'}"
420
-
pip install nemo_toolkit['vision'] # or pip install "nemo_toolkit['vision']@git+https://github.com/NVIDIA/NeMo@${REF:-'main'}"
421
374
pip install nemo_toolkit['multimodal'] # or pip install "nemo_toolkit['multimodal']@git+https://github.com/NVIDIA/NeMo@${REF:-'main'}"
422
375
```
423
376
424
377
### NGC PyTorch container
425
378
426
-
**NOTE: The following steps are supported beginning with 24.04 (NeMo-Toolkit 2.3.0)**
379
+
**NOTE: The following steps are supported beginning with 25.09 (NeMo-Toolkit 2.6.0)**
427
380
428
381
We recommended that you start with a base NVIDIA PyTorch container:
429
-
nvcr.io/nvidia/pytorch:25.01-py3.
382
+
nvcr.io/nvidia/pytorch:25.09-py3.
430
383
431
384
If starting with a base NVIDIA PyTorch container, you must first launch
Bria Builds Responsible Generative AI for Enterprises Using NVIDIA NeMo, Picasso
515
-
</a> (2024/03/06)
516
-
</summary>
517
-
Bria, a Tel Aviv startup at the forefront of visual generative AI for enterprises now leverages the NVIDIA NeMo Framework.
518
-
The Bria.ai platform uses reference implementations from the NeMo Multimodal collection, trained on NVIDIA Tensor Core GPUs, to enable high-throughput and low-latency image generation.
519
-
Bria has also adopted NVIDIA Picasso, a foundry for visual generative AI models, to run inference.
0 commit comments