Skip to content

Commit a846d02

Browse files
committed
remove todo
1 parent 2dee7cf commit a846d02

File tree

2 files changed

+2
-6
lines changed

2 files changed

+2
-6
lines changed

gguf-py/gguf/tensor_mapping.py

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -583,6 +583,7 @@ class TensorNameMap:
583583
),
584584

585585
MODEL_TENSOR.SSM_NORM: (
586+
"model.layers.{bid}.mamba.norm", # falcon-h1
586587
"backbone.layers.{bid}.mixer.norm", # mamba2
587588
),
588589

@@ -1177,10 +1178,6 @@ class TensorNameMap:
11771178
"resampler.attn.out_proj",
11781179
),
11791180

1180-
MODEL_TENSOR.SSM_NORM: (
1181-
"model.layers.{bid}.mamba.norm",
1182-
),
1183-
11841181
MODEL_TENSOR.V_RESMPL_KV: (
11851182
"resampler.kv_proj",
11861183
),

src/llama-arch.cpp

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1955,8 +1955,7 @@ bool llm_arch_is_recurrent(const llm_arch & arch) {
19551955
}
19561956

19571957
bool llm_arch_is_hybrid(const llm_arch & arch) {
1958-
// TODO: There are currently no hybrid models! Once there are, this will be
1959-
// the place to identify them
1958+
// List all mamba-attention hybrid models here
19601959
switch (arch) {
19611960
case LLM_ARCH_FALCON_H1:
19621961
return true;

0 commit comments

Comments
 (0)