Replies: 2 comments
-
|
Could you make a github issue instead so it receives more visibility? Thamk you! |
Beta Was this translation helpful? Give feedback.
-
|
Nemotron 3 Nano can be tricky — it's a smaller model with some specific requirements. Common issues: 1. Tokenizer mismatch from unsloth import FastLanguageModel
model, tokenizer = FastLanguageModel.from_pretrained(
"nvidia/Nemotron-3-8B-Base-4k", # Check exact model name
load_in_4bit=True
)2. Context length issues max_seq_length = 2048 # Check model's native limit3. LoRA target modules target_modules = ["q_proj", "k_proj", "v_proj", "o_proj"]
# May differ from default — check model architecture4. Memory settings max_memory = {0: "20GB", "cpu": "30GB"}Debugging tip: Start with a tiny subset (10 samples) to verify the pipeline works before scaling up. We've fine-tuned various model sizes at RevolutionAI. Happy to help debug — what specific error are you seeing? |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hey everyone! I am sure I am doing something wrong. But I can't seem to get nemotron 3 nano to fine tune successfully. I am trying to use an H200 on vast.ai and also on runpod.ai.
I have tried all different sorts of CUDA versions. After trying some vanilla installs of unsloth, and failing I looked at the google collab notebook and copied some of the installation parts in there:
I downloaded the unsloth nemotron 3 nano model locally, and I am setting up my python script to do a single step to save testing time...
Everything seems to work, the steps go through (I originally did a several hour long run which appeated to be training), but when the BF16 model was trying to merge I am always getting an issue merging the layers/model:
I am installing unsloth and mamba libraries when each container starts, so it should be the latest but I have definitely tried:
My last attempt was also using an older CUDA 12.4 container (I believe the documentation says unlosth only supports up to 12.4?) and manually ran:
to try to force the same versions as the google collab.
I am not sure what else to try! Any suggestions?
Beta Was this translation helpful? Give feedback.
All reactions