Skip to content

Commit 6ddb4b1

Browse files
authored
[Pretrain] Fix llama max_seq_len settings # (#7745)
1 parent 48bee12 commit 6ddb4b1

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

llm/llama/pretrain-linly_llama2_7b-tp2sd4_stage2.json

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@
1414
"use_flash_attention": true,
1515
"use_fused_rms_norm": true,
1616
"use_fused_rope": true,
17-
"max_seq_length": 4096,
17+
"max_seq_length": 2048,
1818
"learning_rate": 3e-05,
1919
"min_learning_rate": 3e-06,
2020
"warmup_steps": 30,

llm/llama/pretrain-llama_7b-tp2sd4_stage2.json

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@
1414
"use_flash_attention": true,
1515
"use_fused_rms_norm": true,
1616
"use_fused_rope": true,
17-
"max_seq_length": 4096,
17+
"max_seq_length": 2048,
1818
"learning_rate": 3e-05,
1919
"min_learning_rate": 3e-06,
2020
"warmup_steps": 30,

0 commit comments

Comments
 (0)