Skip to content

Commit 754b7a9

Browse files
committed
remove mistral arch
1 parent dcb3bc5 commit 754b7a9

File tree

4 files changed

+0
-41
lines changed

4 files changed

+0
-41
lines changed

gguf-py/gguf/constants.py

Lines changed: 0 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -686,7 +686,6 @@ class MODEL_TENSOR(IntEnum):
686686
MODEL_ARCH.SMOLLM3: "smollm3",
687687
MODEL_ARCH.LFM2: "lfm2",
688688
MODEL_ARCH.DREAM: "dream",
689-
MODEL_ARCH.MISTRAL: "mistral",
690689
}
691690

692691
VISION_PROJECTOR_TYPE_NAMES: dict[VISION_PROJECTOR_TYPE, str] = {
@@ -2436,21 +2435,6 @@ class MODEL_TENSOR(IntEnum):
24362435
MODEL_TENSOR.ATTN_V,
24372436
MODEL_TENSOR.ATTN_OUT,
24382437
],
2439-
MODEL_ARCH.MISTRAL: [
2440-
MODEL_TENSOR.TOKEN_EMBD,
2441-
MODEL_TENSOR.OUTPUT_NORM,
2442-
MODEL_TENSOR.OUTPUT,
2443-
MODEL_TENSOR.ATTN_NORM,
2444-
MODEL_TENSOR.ATTN_Q,
2445-
MODEL_TENSOR.ATTN_K,
2446-
MODEL_TENSOR.ATTN_V,
2447-
MODEL_TENSOR.ATTN_OUT,
2448-
MODEL_TENSOR.FFN_GATE_INP,
2449-
MODEL_TENSOR.FFN_NORM,
2450-
MODEL_TENSOR.FFN_GATE,
2451-
MODEL_TENSOR.FFN_DOWN,
2452-
MODEL_TENSOR.FFN_UP,
2453-
]
24542438
# TODO
24552439
}
24562440

src/llama-arch.cpp

Lines changed: 0 additions & 19 deletions
Original file line numberDiff line numberDiff line change
@@ -86,7 +86,6 @@ static const std::map<llm_arch, const char *> LLM_ARCH_NAMES = {
8686
{ LLM_ARCH_SMOLLM3, "smollm3" },
8787
{ LLM_ARCH_LFM2, "lfm2" },
8888
{ LLM_ARCH_DREAM, "dream" },
89-
{ LLM_ARCH_MISTRAL, "mistral" },
9089
{ LLM_ARCH_UNKNOWN, "(unknown)" },
9190
};
9291

@@ -1887,24 +1886,6 @@ static const std::map<llm_arch, std::map<llm_tensor, const char *>> LLM_TENSOR_N
18871886
{ LLM_TENSOR_TOKEN_EMBD_NORM, "token_embd_norm" },
18881887
}
18891888
},
1890-
{
1891-
LLM_ARCH_MISTRAL,
1892-
{
1893-
{ LLM_TENSOR_TOKEN_EMBD, "token_embd" },
1894-
{ LLM_TENSOR_OUTPUT_NORM, "output_norm" },
1895-
{ LLM_TENSOR_OUTPUT, "output" },
1896-
{ LLM_TENSOR_ATTN_NORM, "blk.%d.attn_norm" },
1897-
{ LLM_TENSOR_ATTN_Q, "blk.%d.attn_q" },
1898-
{ LLM_TENSOR_ATTN_K, "blk.%d.attn_k" },
1899-
{ LLM_TENSOR_ATTN_V, "blk.%d.attn_v" },
1900-
{ LLM_TENSOR_ATTN_OUT, "blk.%d.attn_output" },
1901-
{ LLM_TENSOR_FFN_GATE_INP, "blk.%d.ffn_gate_inp" },
1902-
{ LLM_TENSOR_FFN_NORM, "blk.%d.ffn_norm" },
1903-
{ LLM_TENSOR_FFN_GATE, "blk.%d.ffn_gate" },
1904-
{ LLM_TENSOR_FFN_DOWN, "blk.%d.ffn_down" },
1905-
{ LLM_TENSOR_FFN_UP, "blk.%d.ffn_up" },
1906-
},
1907-
},
19081889
{
19091890
LLM_ARCH_UNKNOWN,
19101891
{

src/llama-arch.h

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -90,7 +90,6 @@ enum llm_arch {
9090
LLM_ARCH_SMOLLM3,
9191
LLM_ARCH_LFM2,
9292
LLM_ARCH_DREAM,
93-
LLM_ARCH_MISTRAL,
9493
LLM_ARCH_UNKNOWN,
9594
};
9695

src/llama-model.cpp

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -570,8 +570,6 @@ void llama_model::load_hparams(llama_model_loader & ml) {
570570
// arch-specific KVs
571571
switch (arch) {
572572
case LLM_ARCH_LLAMA:
573-
case LLM_ARCH_MISTRAL:
574-
{
575573
ml.get_key(LLM_KV_ATTENTION_LAYERNORM_RMS_EPS, hparams.f_norm_rms_eps);
576574

577575
if (hparams.n_expert == 8) {
@@ -2019,7 +2017,6 @@ bool llama_model::load_tensors(llama_model_loader & ml) {
20192017
case LLM_ARCH_MINICPM:
20202018
case LLM_ARCH_GRANITE:
20212019
case LLM_ARCH_GRANITE_MOE:
2022-
case LLM_ARCH_MISTRAL:
20232020
{
20242021
tok_embd = create_tensor(tn(LLM_TENSOR_TOKEN_EMBD, "weight"), {n_embd, n_vocab}, 0);
20252022

@@ -16738,7 +16735,6 @@ llm_graph_result_ptr llama_model::build_graph(
1673816735

1673916736
switch (arch) {
1674016737
case LLM_ARCH_LLAMA:
16741-
case LLM_ARCH_MISTRAL:
1674216738
{
1674316739
llm = std::make_unique<llm_build_llama>(*this, params, gf);
1674416740
} break;
@@ -17209,7 +17205,6 @@ llama_rope_type llama_model_rope_type(const llama_model * model) {
1720917205
case LLM_ARCH_SMOLLM3:
1721017206
case LLM_ARCH_ARCEE:
1721117207
case LLM_ARCH_ERNIE4_5:
17212-
case LLM_ARCH_MISTRAL:
1721317208
return LLAMA_ROPE_TYPE_NORM;
1721417209

1721517210
// the pairs of head values are offset by n_rot/2

0 commit comments

Comments
 (0)