We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 0941092 commit 2ddfa27Copy full SHA for 2ddfa27
convert_hf_to_gguf.py
@@ -6971,7 +6971,10 @@ def _add_feed_forward_length(self):
6971
6972
def set_gguf_parameters(self):
6973
# set num_key_value_heads only for attention layers
6974
- self.hparams["num_key_value_heads"] = [(self.hparams["num_key_value_heads"] if x in self.hparams["full_attn_idxs"] else 0) for x in range(self.block_count)]
+ self.hparams["num_key_value_heads"] = [
6975
+ self.hparams["num_key_value_heads"] if layer_type == "full_attention" else 0
6976
+ for layer_type in self.hparams["layer_types"]
6977
+ ]
6978
6979
super().set_gguf_parameters()
6980
self.gguf_writer.add_vocab_size(self.hparams["vocab_size"])
0 commit comments