We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 92f77a6 commit 1a31d0dCopy full SHA for 1a31d0d
examples/quantize/README.md
@@ -81,7 +81,7 @@ Several quantization methods are supported. They differ in the resulting model d
81
- [#4930 - imatrix for all k-quants](https://github.com/ggerganov/llama.cpp/pull/4930)
82
- [#4951 - imatrix on the GPU](https://github.com/ggerganov/llama.cpp/pull/4957)
83
- [#4969 - imatrix for legacy quants](https://github.com/ggerganov/llama.cpp/pull/4969)
84
- - [#4996 - k-qunats tuning](https://github.com/ggerganov/llama.cpp/pull/4996)
+ - [#4996 - k-quants tuning](https://github.com/ggerganov/llama.cpp/pull/4996)
85
- [#5060 - Q3_K_XS](https://github.com/ggerganov/llama.cpp/pull/5060)
86
- [#5196 - 3-bit i-quants](https://github.com/ggerganov/llama.cpp/pull/5196)
87
- [quantization tuning](https://github.com/ggerganov/llama.cpp/pull/5320), [another one](https://github.com/ggerganov/llama.cpp/pull/5334), and [another one](https://github.com/ggerganov/llama.cpp/pull/5361)
0 commit comments