File tree Expand file tree Collapse file tree 2 files changed +41
-1
lines changed
examples/llm_pte_finetuning Expand file tree Collapse file tree 2 files changed +41
-1
lines changed Original file line number Diff line number Diff line change 1+ tokenizer :
2+ _component_ : torchtune.models.qwen2.qwen2_tokenizer
3+ path : /tmp/Qwen2-0.5B-Instruct/vocab.json
4+ merges_file : /tmp/Qwen2-0.5B-Instruct/merges.txt
5+ max_seq_len : 512
6+
7+ dataset :
8+ _component_ : torchtune.datasets.alpaca_cleaned_dataset
9+ seed : null
10+ shuffle : True
11+ batch_size : 1
12+
13+ loss :
14+ _component_ : torch.nn.CrossEntropyLoss
15+
16+ model :
17+ _component_ : torchtune.models.qwen2.lora_qwen2_0_5b
18+ lora_attn_modules : ['q_proj', 'k_proj', 'v_proj']
19+ apply_lora_to_mlp : False
20+ lora_rank : 32
21+ lora_alpha : 64
22+ learning_rate : 5e-3
23+
24+ checkpointer :
25+ _component_ : torchtune.training.FullModelHFCheckpointer
26+ checkpoint_dir : /tmp/Qwen2-0.5B-Instruct
27+ checkpoint_files : [
28+ model.safetensors
29+ ]
30+ recipe_checkpoint : null
31+ output_dir : /tmp/qwen_0.5B_ft-output
32+ model_type : QWEN2
33+ resume_from_checkpoint : False
34+ save_adapter_weights_only : False
35+
36+ device : cpu
37+ dtype : fp32
38+
39+ enable_activation_checkpointing : True
40+ compile : False
Original file line number Diff line number Diff line change @@ -84,7 +84,7 @@ def main() -> None:
8484 # params run from [param_start, outputs_end]
8585 grad_start = et_mod .run_method ("__et_training_gradients_index_forward" , [])[0 ]
8686 param_start = et_mod .run_method ("__et_training_parameters_index_forward" , [])[0 ]
87- learning_rate = 5e-3
87+ learning_rate = cfg . model . learning_rate
8888 f .seek (0 )
8989 losses = []
9090 for i , batch in tqdm (enumerate (train_dataloader ), total = num_training_steps ):
You can’t perform that action at this time.
0 commit comments