Skip to content

Commit 8ed3cd9

Browse files
authored
chore: add nanov3 lora sft recipe to doc (#1860)
Signed-off-by: ruit <[email protected]>
1 parent dacac7e commit 8ed3cd9

File tree

1 file changed

+2
-1
lines changed

1 file changed

+2
-1
lines changed

docs/guides/sft.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -191,7 +191,7 @@ Upon completion of the training process, you can refer to our [evaluation guide]
191191

192192
## LoRA Configuration
193193

194-
NeMo RL supports LoRA (Low-Rank Adaptation) for parameter-efficient fine-tuning. LoRA reduces trainable parameters by using low-rank matrices for weight updates while keeping the base model frozen.
194+
NeMo RL supports LoRA (Low-Rank Adaptation) for parameter-efficient fine-tuning, including Nano‑v3 models. LoRA reduces trainable parameters by using low-rank matrices for weight updates while keeping the base model frozen.
195195

196196
Notes:
197197
- LoRA is supported with DTensor v2 and Megatron backends. Uses the DTensor backend by default. DTensor v1 does not support LoRA (ensure `policy.dtensor_cfg._v2=true` when using DTensor).
@@ -234,6 +234,7 @@ policy:
234234
```bash
235235
uv run examples/run_sft.py policy.dtensor_cfg.lora_cfg.enabled=true
236236
```
237+
For the Nano‑v3 SFT LoRA recipe, see:[sft-nanov3-30BA3B-2n8g-fsdp2-lora.yaml](../../examples/configs/recipes/llm/sft-nanov3-30BA3B-2n8g-fsdp2-lora.yaml).
237238

238239
### Megatron Configuration Parameters
239240

0 commit comments

Comments
 (0)