We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 5b6c988 commit 942252eCopy full SHA for 942252e
src/diffusers/quantizers/gguf/utils.py
@@ -71,10 +71,9 @@
71
gguf.GGMLQuantizationType.IQ4_NL,
72
}
73
# TODO(Isotr0py): Currently, we don't have MMQ kernel for I-Matrix quantization.
74
-# Consolidate DEQUANT_TYPES, MMVQ_QUANT_TYPES and MMQ_QUANT_TYPES after we add
+# Consolidate DEQUANT_TYPES and MMQ_QUANT_TYPES after we add
75
# MMQ kernel for I-Matrix quantization.
76
DEQUANT_TYPES = STANDARD_QUANT_TYPES | KQUANT_TYPES | IMATRIX_QUANT_TYPES
77
-MMVQ_QUANT_TYPES = STANDARD_QUANT_TYPES | KQUANT_TYPES | IMATRIX_QUANT_TYPES
78
MMQ_QUANT_TYPES = STANDARD_QUANT_TYPES | KQUANT_TYPES
79
80
0 commit comments