Skip to content

Commit d7fae96

Browse files
felipemello1Felipe Mello
andauthored
expose and set lora_dropout = 0.0 (#1492)
Co-authored-by: Felipe Mello <[email protected]>
1 parent 221031a commit d7fae96

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

48 files changed

+90
-28
lines changed

recipes/configs/code_llama2/7B_lora_single_device.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -23,6 +23,7 @@ model:
2323
apply_lora_to_output: False
2424
lora_rank: 8
2525
lora_alpha: 16
26+
lora_dropout: 0.0
2627

2728
# Tokenizer
2829
tokenizer:

recipes/configs/code_llama2/7B_qlora_single_device.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -23,6 +23,7 @@ model:
2323
apply_lora_to_output: False
2424
lora_rank: 8
2525
lora_alpha: 16
26+
lora_dropout: 0.0
2627

2728
# Tokenizer
2829
tokenizer:

recipes/configs/gemma/2B_lora.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -34,6 +34,7 @@ model:
3434
apply_lora_to_mlp: True
3535
lora_rank: 64
3636
lora_alpha: 16
37+
lora_dropout: 0.0
3738

3839
checkpointer:
3940
_component_: torchtune.training.FullModelHFCheckpointer

recipes/configs/gemma/2B_lora_single_device.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -33,6 +33,7 @@ model:
3333
apply_lora_to_mlp: True
3434
lora_rank: 64
3535
lora_alpha: 16
36+
lora_dropout: 0.0
3637

3738
checkpointer:
3839
_component_: torchtune.training.FullModelHFCheckpointer

recipes/configs/gemma/2B_qlora_single_device.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -33,6 +33,7 @@ model:
3333
apply_lora_to_mlp: True
3434
lora_rank: 64
3535
lora_alpha: 16
36+
lora_dropout: 0.0
3637

3738
checkpointer:
3839
_component_: torchtune.training.FullModelHFCheckpointer

recipes/configs/gemma/7B_lora.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -34,6 +34,7 @@ model:
3434
apply_lora_to_mlp: True
3535
lora_rank: 64
3636
lora_alpha: 16
37+
lora_dropout: 0.0
3738

3839
checkpointer:
3940
_component_: torchtune.training.FullModelHFCheckpointer

recipes/configs/gemma/7B_lora_single_device.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -33,6 +33,7 @@ model:
3333
apply_lora_to_mlp: True
3434
lora_rank: 8
3535
lora_alpha: 16
36+
lora_dropout: 0.0
3637

3738
checkpointer:
3839
_component_: torchtune.training.FullModelHFCheckpointer

recipes/configs/gemma/7B_qlora_single_device.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -33,6 +33,7 @@ model:
3333
apply_lora_to_mlp: True
3434
lora_rank: 64
3535
lora_alpha: 16
36+
lora_dropout: 0.0
3637

3738
checkpointer:
3839
_component_: torchtune.training.FullModelHFCheckpointer

recipes/configs/llama2/13B_lora.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -27,6 +27,7 @@ model:
2727
apply_lora_to_output: True
2828
lora_rank: 8
2929
lora_alpha: 16
30+
lora_dropout: 0.0
3031

3132
checkpointer:
3233
_component_: torchtune.training.FullModelHFCheckpointer

recipes/configs/llama2/13B_qlora_single_device.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -23,6 +23,7 @@ model:
2323
apply_lora_to_output: False
2424
lora_rank: 8
2525
lora_alpha: 16
26+
lora_dropout: 0.0
2627

2728
tokenizer:
2829
_component_: torchtune.models.llama2.llama2_tokenizer

0 commit comments

Comments
 (0)