We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent b9dfbfb commit c92e14aCopy full SHA for c92e14a
otherarch/sdcpp/ggml_extend.hpp
@@ -984,7 +984,8 @@ __STATIC_INLINE__ struct ggml_tensor* ggml_nn_attention_ext(struct ggml_context*
984
d_head == 112 ||
985
d_head == 128 ||
986
d_head == 256);
987
-#if 0
+// kcpp disable kv_pad (leejet/stable-diffusion.cpp#756)
988
+#if 1
989
can_use_flash_attn = can_use_flash_attn && L_k % 256 == 0;
990
#else
991
if (can_use_flash_attn && L_k % 256 != 0) {
0 commit comments