Skip to content

Commit f1485d2

Browse files
committed
fix: If not using ROPE, context is "infinite"
Branch: GraniteFour Signed-off-by: Gabe Goodhart <[email protected]>
1 parent 63f1ed8 commit f1485d2

File tree

1 file changed

+2
-1
lines changed

1 file changed

+2
-1
lines changed

convert_hf_to_gguf.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6553,7 +6553,6 @@ def set_gguf_parameters(self):
65536553
## General Params ##
65546554
self.gguf_writer.add_embedding_length(self.d_model)
65556555
self.gguf_writer.add_block_count(self.block_count)
6556-
self.gguf_writer.add_context_length(self.hparams.get("max_position_embeddings", 0))
65576556
self.gguf_writer.add_vocab_size(self.hparams["vocab_size"])
65586557
self.gguf_writer.add_feed_forward_length(self.hparams["intermediate_size"])
65596558

@@ -6584,6 +6583,8 @@ def set_gguf_parameters(self):
65846583
## If Bamba, use rope, otherwise don't
65856584
use_rope = "BambaForCausalLM" in self.hparams["architectures"]
65866585
self.gguf_writer.add_rope_scaling_finetuned(use_rope)
6586+
if not use_rope:
6587+
self.gguf_writer.add_context_length(2**20)
65876588

65886589
## Validation ##
65896590
d_head = self.find_hparam(["d_head"], optional=True) or 64

0 commit comments

Comments
 (0)