Skip to content
Discussion options

You must be logged in to vote

The original model has fp8 weights and some extra tensors with dequantization scales. llama.cpp doesn't support this, you have to:

  1. Either Download bf16 DeepSeek R1, for example this: https://huggingface.co/unsloth/DeepSeek-R1-BF16. This will convert without any errors.
  2. Or convert the original fp8 model to bf16 by yourself. You can find instructions here: https://huggingface.co/huihui-ai/DeepSeek-R1-bf16. Note that the conversion script (it's bundled with DeepSeek V3 model) uses triton, so you need a GPU for this step.

Replies: 1 comment 6 replies

Comment options

You must be logged in to vote
6 replies
@fairydreaming
Comment options

Answer selected by CISC
@nickhuang99
Comment options

@fairydreaming
Comment options

@nickhuang99
Comment options

@fairydreaming
Comment options

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
5 participants