Skip to content

Commit d68f4a5

Browse files
committed
disable clip fa for now
1 parent 2b00292 commit d68f4a5

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

gpttype_adapter.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2470,7 +2470,7 @@ ModelLoadResult gpttype_load_model(const load_model_inputs inputs, FileFormat in
24702470
}
24712471
clip_context_params ctx_clip_params {
24722472
/* use_gpu */ true,
2473-
/* flash_attn_type */ (kcpp_data->flash_attn?CLIP_FLASH_ATTN_TYPE_ENABLED:CLIP_FLASH_ATTN_TYPE_DISABLED),
2473+
/* flash_attn_type */ CLIP_FLASH_ATTN_TYPE_DISABLED, //kcpp: disabled in 1.102.2 as some headsizes break on turing
24742474
/* image_min_tokens */ -1,
24752475
/* image_max_tokens */ -1,
24762476
};

koboldcpp.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -66,7 +66,7 @@
6666
extra_images_max = 4 # for kontext/qwen img
6767

6868
# global vars
69-
KcppVersion = "1.102.1"
69+
KcppVersion = "1.102.2"
7070
showdebug = True
7171
kcpp_instance = None #global running instance
7272
global_memory = {"tunnel_url": "", "restart_target":"", "input_to_exit":False, "load_complete":False, "restart_override_config_target":""}

0 commit comments

Comments
 (0)