Skip to content

Commit 31ab716

Browse files
authored
remove sequence parallel in llama4 (#40084)
1 parent a1a4fcd commit 31ab716

File tree

1 file changed

+0
-3
lines changed

1 file changed

+0
-3
lines changed

src/transformers/models/llama4/configuration_llama4.py

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -251,9 +251,6 @@ class Llama4TextConfig(PretrainedConfig):
251251
"layers.*.self_attn.k_proj": "colwise",
252252
"layers.*.self_attn.v_proj": "colwise",
253253
"layers.*.self_attn.o_proj": "rowwise",
254-
"layers.*.input_layernorm.weight": "sequence_parallel",
255-
"layers.*.post_attention_layernorm.weight": "sequence_parallel",
256-
"norm.weight": "sequence_parallel",
257254
"layers.*.feed_forward.shared_expert.gate_proj": "local_colwise",
258255
"layers.*.feed_forward.shared_expert.up_proj": "local_colwise",
259256
"layers.*.feed_forward.shared_expert.down_proj": "local_rowwise",

0 commit comments

Comments
 (0)