Skip to content

Commit 2ddfa27

Browse files
committed
Use layer_types instead of full_attn_idxs
1 parent 0941092 commit 2ddfa27

File tree

1 file changed

+4
-1
lines changed

1 file changed

+4
-1
lines changed

convert_hf_to_gguf.py

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6971,7 +6971,10 @@ def _add_feed_forward_length(self):
69716971

69726972
def set_gguf_parameters(self):
69736973
# set num_key_value_heads only for attention layers
6974-
self.hparams["num_key_value_heads"] = [(self.hparams["num_key_value_heads"] if x in self.hparams["full_attn_idxs"] else 0) for x in range(self.block_count)]
6974+
self.hparams["num_key_value_heads"] = [
6975+
self.hparams["num_key_value_heads"] if layer_type == "full_attention" else 0
6976+
for layer_type in self.hparams["layer_types"]
6977+
]
69756978

69766979
super().set_gguf_parameters()
69776980
self.gguf_writer.add_vocab_size(self.hparams["vocab_size"])

0 commit comments

Comments
 (0)