You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: examples/model_configs/peft_model.yaml
+1-1Lines changed: 1 addition & 1 deletion
Original file line number
Diff line number
Diff line change
@@ -1,6 +1,6 @@
1
1
model_parameters:
2
2
model_name: "deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B"# pretrained=model_name,trust_remote_code=boolean,revision=revision_to_use,model_parallel=True ... For a PEFT model, the pretrained model should be the one trained with PEFT and the base model below will contain the original model on which the adapters will be applied.
3
-
tokenizer: null # name of tokenier to use if defferent from the model's default
3
+
tokenizer: null # name of tokenizer to use if different from the model's default
4
4
subfolder: null # subfolder in the model's directory to use
5
5
dtype: "float16"# Specifying the model to be loaded in 4 bit uses BitsAndBytesConfig. The other option is to use "8bit" quantization.
0 commit comments