We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent f0518b3 commit ca8ea1bCopy full SHA for ca8ea1b
packages/gguf/src/quant-descriptions.ts
@@ -54,7 +54,7 @@ export const GGUF_QUANT_DESCRIPTIONS: Record<GGMLQuantizationType, { txt: string
54
src_url: "https://github.com/ggerganov/llama.cpp/pull/1684#issue-1739619305",
55
},
56
[GGMLQuantizationType.Q2_K]: {
57
- txt: `2-bit quantization (q). Super-blocks with 16 blocks, each block has 16 weight. Weight formula: w = q * block_scale(4-bit) + block_min(4-bit), resulting in 2.5625 bits-per-weight.`,
+ txt: `2-bit quantization (q). Super-blocks with 16 blocks, each block has 16 weight. Weight formula: w = q * block_scale(4-bit) + block_min(4-bit), resulting in 2.625 bits-per-weight.`,
58
59
60
[GGMLQuantizationType.IQ4_XS]: {
0 commit comments