|
262 | 262 | # LLAMA_VOCAB_PRE_TYPE_SEED_CODER = 35, |
263 | 263 | # LLAMA_VOCAB_PRE_TYPE_HUNYUAN = 36, |
264 | 264 | # LLAMA_VOCAB_PRE_TYPE_KIMI_K2 = 37, |
| 265 | +# LLAMA_VOCAB_PRE_TYPE_HUNYUAN_DENSE = 38, |
265 | 266 | # }; |
266 | 267 | LLAMA_VOCAB_PRE_TYPE_DEFAULT = 0 |
267 | 268 | LLAMA_VOCAB_PRE_TYPE_LLAMA3 = 1 |
|
301 | 302 | LLAMA_VOCAB_PRE_TYPE_SEED_CODER = 35 |
302 | 303 | LLAMA_VOCAB_PRE_TYPE_HUNYUAN = 36 |
303 | 304 | LLAMA_VOCAB_PRE_TYPE_KIMI_K2 = 37 |
| 305 | +LLAMA_VOCAB_PRE_TYPE_HUNYUAN_DENSE = 38 |
304 | 306 |
|
305 | 307 |
|
306 | 308 | # // note: these values should be synchronized with ggml_rope |
|
403 | 405 | # //LLAMA_FTYPE_MOSTLY_Q4_0_8_8 = 35, // removed from gguf files, use Q4_0 and runtime repack |
404 | 406 | # LLAMA_FTYPE_MOSTLY_TQ1_0 = 36, // except 1d tensors |
405 | 407 | # LLAMA_FTYPE_MOSTLY_TQ2_0 = 37, // except 1d tensors |
| 408 | +# LLAMA_FTYPE_MOSTLY_MXFP4_MOE = 38, // except 1d tensors |
406 | 409 | # |
407 | 410 | # LLAMA_FTYPE_GUESSED = 1024, // not specified in the model file |
408 | 411 | # }; |
|
441 | 444 | # LLAMA_FTYPE_MOSTLY_Q4_0_8_8 = 35 |
442 | 445 | LLAMA_FTYPE_MOSTLY_TQ1_0 = 36 |
443 | 446 | LLAMA_FTYPE_MOSTLY_TQ2_0 = 37 |
| 447 | +LLAMA_FTYPE_MOSTLY_MXFP4_MOE = 38 |
444 | 448 | LLAMA_FTYPE_GUESSED = 1024 |
445 | 449 |
|
446 | 450 | # enum llama_rope_scaling_type { |
|
0 commit comments