Skip to content

Commit 566ec02

Browse files
committed
Merge branch 'concedo_experimental' into esocrok
2 parents 2e74492 + c00ae93 commit 566ec02

File tree

4 files changed

+164
-144
lines changed

4 files changed

+164
-144
lines changed

Makefile

Lines changed: 9 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -471,8 +471,13 @@ endif
471471

472472
ifdef NO_VULKAN_EXTENSIONS
473473
VKGEN_NOEXT_ADD = -DNO_VULKAN_EXTENSIONS
474+
VKGEN_SUFFIX = -noext
475+
else
476+
VKGEN_SUFFIX =
474477
endif
475478
VKGEN_NOEXT_FORCE = -DNO_VULKAN_EXTENSIONS
479+
VKGEN_HPP = ggml/src/ggml-vulkan-shaders$(VKGEN_SUFFIX).hpp
480+
VKGEN_CPP = ggml/src/ggml-vulkan-shaders$(VKGEN_SUFFIX).cpp
476481

477482
#
478483
# Print build information
@@ -685,9 +690,9 @@ ggml_v3-opencl.o: otherarch/ggml_v3-opencl.cpp otherarch/ggml_v3-opencl.h
685690
$(CXX) $(CXXFLAGS) $(CLBLAST_FLAGS) -c $< -o $@
686691

687692
#vulkan
688-
ggml-vulkan.o: ggml/src/ggml-vulkan/ggml-vulkan.cpp ggml/include/ggml-vulkan.h ggml/src/ggml-vulkan-shaders.cpp
693+
ggml-vulkan.o: ggml/src/ggml-vulkan/ggml-vulkan.cpp ggml/include/ggml-vulkan.h $(VKGEN_CPP)
689694
$(CXX) $(CXXFLAGS) $(VKGEN_NOEXT_ADD) $(VULKAN_FLAGS) -c $< -o $@
690-
ggml-vulkan-shaders.o: ggml/src/ggml-vulkan-shaders.cpp ggml/include/ggml-vulkan.h
695+
ggml-vulkan-shaders.o: $(VKGEN_CPP) ggml/include/ggml-vulkan.h
691696
$(CXX) $(CXXFLAGS) $(VKGEN_NOEXT_ADD) $(VULKAN_FLAGS) -c $< -o $@
692697
ggml-vulkan-noext.o: ggml/src/ggml-vulkan/ggml-vulkan.cpp ggml/include/ggml-vulkan.h ggml/src/ggml-vulkan-shaders-noext.cpp
693698
$(CXX) $(CXXFLAGS) $(VKGEN_NOEXT_FORCE) $(VULKAN_FLAGS) -c $< -o $@
@@ -785,7 +790,7 @@ vulkan-shaders-gen: ggml/src/ggml-vulkan/vulkan-shaders/vulkan-shaders-gen.cpp
785790
$(CXX) $(CXXFLAGS) $(VKGEN_NOEXT_ADD) $(filter-out %.h,$^) -o $@ $(LDFLAGS)
786791
ifeq ($(OS),Windows_NT)
787792
@echo 'Now rebuilding vulkan shaders for Windows...'
788-
$(shell) vulkan-shaders-gen --glslc glslc --input-dir ggml/src/ggml-vulkan/vulkan-shaders --target-hpp ggml/src/ggml-vulkan-shaders.hpp --target-cpp ggml/src/ggml-vulkan-shaders.cpp --output-dir vulkan-spv-tmp
793+
$(shell) vulkan-shaders-gen --glslc glslc --input-dir ggml/src/ggml-vulkan/vulkan-shaders --target-hpp $(VKGEN_HPP) --target-cpp $(VKGEN_CPP) --output-dir vulkan-spv-tmp
789794
@echo 'Vulkan Shaders Rebuilt for Windows...'
790795
else
791796
@echo 'Now rebuilding vulkan shaders for Linux...'
@@ -819,7 +824,7 @@ else
819824
echo "Error: No usable glslc found. Vulkan shaders cannot be compiled!"; \
820825
else \
821826
echo "Using GLSLC: $$GLSLC_BIN"; \
822-
./vulkan-shaders-gen --glslc "$$GLSLC_BIN" --input-dir ggml/src/ggml-vulkan/vulkan-shaders --target-hpp ggml/src/ggml-vulkan-shaders.hpp --target-cpp ggml/src/ggml-vulkan-shaders.cpp --output-dir vulkan-spv-tmp; \
827+
./vulkan-shaders-gen --glslc "$$GLSLC_BIN" --input-dir ggml/src/ggml-vulkan/vulkan-shaders --target-hpp $(VKGEN_HPP) --target-cpp $(VKGEN_CPP) --output-dir vulkan-spv-tmp; \
823828
fi
824829
@echo 'Vulkan Shaders Rebuilt for Linux...'
825830
endif

koboldcpp.py

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -69,7 +69,7 @@
6969
extra_images_max = 4
7070

7171
# global vars
72-
KcppVersion = "1.99.4"
72+
KcppVersion = "1.100"
7373
showdebug = True
7474
kcpp_instance = None #global running instance
7575
global_memory = {"tunnel_url": "", "restart_target":"", "input_to_exit":False, "load_complete":False, "restart_model": "", "currentConfig": None, "modelOverride": None, "currentModel": None}
@@ -1731,7 +1731,7 @@ def sd_load_model(model_filename,vae_filename,lora_filename,t5xxl_filename,clipl
17311731
inputs.flash_attention = args.sdflashattention
17321732
inputs.offload_cpu = args.sdoffloadcpu
17331733
inputs.vae_cpu = args.sdvaecpu
1734-
inputs.clip_cpu = args.sdclipcpu
1734+
inputs.clip_cpu = False if args.sdclipgpu else True
17351735
sdconvdirect = sd_convdirect_option(args.sdconvdirect)
17361736
inputs.diffusion_conv_direct = sdconvdirect == 'full'
17371737
inputs.vae_conv_direct = sdconvdirect in ['vaeonly', 'full']
@@ -5718,7 +5718,7 @@ def hide_tooltip(event):
57185718
sd_flash_attention_var = ctk.IntVar(value=0)
57195719
sd_offload_cpu_var = ctk.IntVar(value=0)
57205720
sd_vae_cpu_var = ctk.IntVar(value=0)
5721-
sd_clip_cpu_var = ctk.IntVar(value=0)
5721+
sd_clip_gpu_var = ctk.IntVar(value=0)
57225722
sd_vaeauto_var = ctk.IntVar(value=0)
57235723
sd_tiled_vae_var = ctk.StringVar(value=str(default_vae_tile_threshold))
57245724
sd_convdirect_var = ctk.StringVar(value=str(sd_convdirect_choices[0]))
@@ -6503,7 +6503,7 @@ def toggletaesd(a,b,c):
65036503
makecheckbox(images_tab, "SD Flash Attention", sd_flash_attention_var, 44,padx=230, tooltiptxt="Enable Flash Attention for image diffusion. May save memory or improve performance.")
65046504
makecheckbox(images_tab, "Model CPU Offload", sd_offload_cpu_var, 50,padx=8, tooltiptxt="Offload image weights in RAM to save VRAM, swap into VRAM when needed.")
65056505
makecheckbox(images_tab, "VAE on CPU", sd_vae_cpu_var, 50,padx=160, tooltiptxt="Force VAE to CPU only for image generation.")
6506-
makecheckbox(images_tab, "CLIP on CPU", sd_clip_cpu_var, 50,padx=280, tooltiptxt="Force CLIP to CPU only for image generation.")
6506+
makecheckbox(images_tab, "CLIP on GPU", sd_clip_gpu_var, 50,padx=280, tooltiptxt="Put CLIP and T5 to GPU for image generation. Otherwise, CLIP will use CPU.")
65076507

65086508
# audio tab
65096509
audio_tab = tabcontent["Audio"]
@@ -6751,8 +6751,8 @@ def export_vars():
67516751
args.sdoffloadcpu = True
67526752
if sd_vae_cpu_var.get()==1:
67536753
args.sdvaecpu = True
6754-
if sd_clip_cpu_var.get()==1:
6755-
args.sdclipcpu = True
6754+
if sd_clip_gpu_var.get()==1:
6755+
args.sdclipgpu = True
67566756
args.sdthreads = (0 if sd_threads_var.get()=="" else int(sd_threads_var.get()))
67576757
args.sdclamped = (0 if int(sd_clamped_var.get())<=0 else int(sd_clamped_var.get()))
67586758
args.sdclampedsoft = (0 if int(sd_clamped_soft_var.get())<=0 else int(sd_clamped_soft_var.get()))
@@ -6997,7 +6997,7 @@ def import_vars(dict):
69976997
sd_flash_attention_var.set(1 if ("sdflashattention" in dict and dict["sdflashattention"]) else 0)
69986998
sd_offload_cpu_var.set(1 if ("sdoffloadcpu" in dict and dict["sdoffloadcpu"]) else 0)
69996999
sd_vae_cpu_var.set(1 if ("sdvaecpu" in dict and dict["sdvaecpu"]) else 0)
7000-
sd_clip_cpu_var.set(1 if ("sdclipcpu" in dict and dict["sdclipcpu"]) else 0)
7000+
sd_clip_gpu_var.set(1 if ("sdclipgpu" in dict and dict["sdclipgpu"]) else 0)
70017001
sd_convdirect_var.set(sd_convdirect_option(dict.get("sdconvdirect")))
70027002
sd_vae_var.set(dict["sdvae"] if ("sdvae" in dict and dict["sdvae"]) else "")
70037003
sd_t5xxl_var.set(dict["sdt5xxl"] if ("sdt5xxl" in dict and dict["sdt5xxl"]) else "")
@@ -8868,7 +8868,7 @@ def range_checker(arg: str):
88688868
sdparsergroup.add_argument("--sdflashattention", help="Enables Flash Attention for image generation.", action='store_true')
88698869
sdparsergroup.add_argument("--sdoffloadcpu", help="Offload image weights in RAM to save VRAM, swap into VRAM when needed.", action='store_true')
88708870
sdparsergroup.add_argument("--sdvaecpu", help="Force VAE to CPU only for image generation.", action='store_true')
8871-
sdparsergroup.add_argument("--sdclipcpu", help="Force CLIP to CPU only for image generation.", action='store_true')
8871+
sdparsergroup.add_argument("--sdclipgpu", help="Put CLIP and T5 to GPU for image generation. Otherwise, CLIP will use CPU.", action='store_true')
88728872
sdparsergroup.add_argument("--sdconvdirect", help="Enables Conv2D Direct. May improve performance or reduce memory usage. Might crash if not supported by the backend. Can be 'off' (default) to disable, 'full' to turn it on for all operations, or 'vaeonly' to enable only for the VAE.", type=sd_convdirect_option, choices=sd_convdirect_choices, default=sd_convdirect_choices[0])
88738873
sdparsergroupvae = sdparsergroup.add_mutually_exclusive_group()
88748874
sdparsergroupvae.add_argument("--sdvae", metavar=('[filename]'), help="Specify an image generation safetensors VAE which replaces the one in the model.", default="")

0 commit comments

Comments
 (0)