Skip to content

Commit 499283c

Browse files
committed
rename define to match upstream
1 parent 22ef97d commit 499283c

File tree

6 files changed

+13
-11
lines changed

6 files changed

+13
-11
lines changed

CMakeLists.txt

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -96,7 +96,7 @@ if (LLAMA_CUBLAS)
9696

9797
add_compile_definitions(GGML_USE_LLAMAFILE)
9898
add_compile_definitions(GGML_USE_CUDA)
99-
add_compile_definitions(SD_USE_CUBLAS)
99+
add_compile_definitions(SD_USE_CUDA)
100100

101101
if (LLAMA_CUDA_F16 OR LLAMA_CUDA_DMMV_F16)
102102
add_compile_definitions(GGML_CUDA_F16)
@@ -177,7 +177,7 @@ if (LLAMA_HIPBLAS)
177177
list(APPEND GGML_SOURCES_ROCM ${SRCS})
178178
file(GLOB SRCS "ggml/src/ggml-cuda/template-instances/mmq*.cu")
179179
list(APPEND GGML_SOURCES_ROCM ${SRCS})
180-
add_compile_definitions(GGML_USE_HIP GGML_USE_CUDA SD_USE_CUBLAS)
180+
add_compile_definitions(GGML_USE_HIP GGML_USE_CUDA SD_USE_CUDA)
181181
add_library(ggml-rocm ${GGML_SOURCES_CUDA})
182182

183183
file(GLOB SRCS "ggml/src/ggml-cuda/template-instances/fattn-vec*q4_0-q4_0.cu")

Makefile

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -83,7 +83,7 @@ CLBLAST_FLAGS = -DGGML_USE_CLBLAST
8383
FAILSAFE_FLAGS = -DUSE_FAILSAFE
8484
VULKAN_FLAGS = -DGGML_USE_VULKAN -DSD_USE_VULKAN
8585
ifdef LLAMA_CUBLAS
86-
CUBLAS_FLAGS = -DGGML_USE_CUDA -DSD_USE_CUBLAS
86+
CUBLAS_FLAGS = -DGGML_USE_CUDA -DSD_USE_CUDA
8787
else
8888
CUBLAS_FLAGS =
8989
endif
@@ -177,7 +177,7 @@ OBJS_CUDA_TEMP_INST += $(patsubst %.cu,%.o,$(wildcard ggml/src/ggml-cuda/templat
177177
OBJS_CUDA_TEMP_INST += $(patsubst %.cu,%.o,$(wildcard ggml/src/ggml-cuda/template-instances/fattn-vec*f16-f16.cu))
178178

179179
ifdef LLAMA_CUBLAS
180-
CUBLAS_FLAGS = -DGGML_USE_CUDA -DSD_USE_CUBLAS -I/usr/local/cuda/include -I/opt/cuda/include -I$(CUDA_PATH)/targets/x86_64-linux/include
180+
CUBLAS_FLAGS = -DGGML_USE_CUDA -DSD_USE_CUDA -I/usr/local/cuda/include -I/opt/cuda/include -I$(CUDA_PATH)/targets/x86_64-linux/include
181181
CUBLASLD_FLAGS = -lcuda -lcublas -lcudart -lcublasLt -lpthread -ldl -lrt -L/usr/local/cuda/lib64 -L/opt/cuda/lib64 -L$(CUDA_PATH)/targets/x86_64-linux/lib -L$(CUDA_PATH)/lib64/stubs -L/usr/local/cuda/targets/aarch64-linux/lib -L/usr/local/cuda/targets/sbsa-linux/lib -L/usr/lib/wsl/lib
182182
CUBLAS_OBJS = ggml-cuda.o ggml_v3-cuda.o ggml_v2-cuda.o ggml_v2-cuda-legacy.o
183183
CUBLAS_OBJS += $(patsubst %.cu,%.o,$(filter-out ggml/src/ggml-cuda/ggml-cuda.cu, $(wildcard ggml/src/ggml-cuda/*.cu)))
@@ -256,7 +256,7 @@ ifdef DETECT_ROCWMMA
256256
HIPFLAGS += -DGGML_HIP_ROCWMMA_FATTN -I$(dir $(DETECT_ROCWMMA))
257257
endif
258258

259-
HIPFLAGS += -DGGML_USE_HIP -DGGML_HIP_NO_VMM -DGGML_USE_CUDA -DSD_USE_CUBLAS $(shell $(ROCM_PATH)/bin/hipconfig -C)
259+
HIPFLAGS += -DGGML_USE_HIP -DGGML_HIP_NO_VMM -DGGML_USE_CUDA -DSD_USE_CUDA $(shell $(ROCM_PATH)/bin/hipconfig -C)
260260
HIPLDFLAGS += -L$(ROCM_PATH)/lib -Wl,-rpath=$(ROCM_PATH)/lib
261261
HIPLDFLAGS += -L$(ROCM_PATH)/lib64 -Wl,-rpath=$(ROCM_PATH)/lib64
262262
HIPLDFLAGS += -lhipblas -lamdhip64 -lrocblas

koboldcpp.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3134,7 +3134,9 @@ def do_POST(self):
31343134
return
31353135
try:
31363136
genparams = json.loads(body)
3137-
schema = genparams.get('schema', {})
3137+
schema = genparams.get('schema', None)
3138+
if not schema:
3139+
schema = genparams
31383140
decoded = convert_json_to_gbnf(schema)
31393141
response_body = (json.dumps({"result": decoded,"success":(True if decoded else False)}).encode())
31403142
except Exception as e:

otherarch/sdcpp/ggml_extend.hpp

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@
2828

2929
#include "model.h"
3030

31-
#ifdef SD_USE_CUBLAS
31+
#ifdef SD_USE_CUDA
3232
#include "ggml-cuda.h"
3333
#endif
3434

@@ -782,7 +782,7 @@ __STATIC_INLINE__ struct ggml_tensor* ggml_nn_attention(struct ggml_context* ctx
782782
struct ggml_tensor* k,
783783
struct ggml_tensor* v,
784784
bool mask = false) {
785-
#if defined(SD_USE_FLASH_ATTENTION) && !defined(SD_USE_CUBLAS) && !defined(SD_USE_METAL) && !defined(SD_USE_VULKAN) && !defined(SD_USE_SYCL)
785+
#if defined(SD_USE_FLASH_ATTENTION) && !defined(SD_USE_CUDA) && !defined(SD_USE_METAL) && !defined(SD_USE_VULKAN) && !defined(SD_USE_SYCL)
786786
struct ggml_tensor* kqv = ggml_flash_attn(ctx, q, k, v, false); // [N * n_head, n_token, d_head]
787787
#else
788788
float d_head = (float)q->ne[0];
@@ -938,7 +938,7 @@ __STATIC_INLINE__ struct ggml_tensor* ggml_nn_group_norm(struct ggml_context* ct
938938
}
939939

940940
__STATIC_INLINE__ void ggml_backend_tensor_get_and_sync(ggml_backend_t backend, const struct ggml_tensor* tensor, void* data, size_t offset, size_t size) {
941-
#if defined(SD_USE_CUBLAS) || defined(SD_USE_SYCL)
941+
#if defined(SD_USE_CUDA) || defined(SD_USE_SYCL)
942942
if (!ggml_backend_is_cpu(backend)) {
943943
ggml_backend_tensor_get_async(backend, tensor, data, offset, size);
944944
ggml_backend_synchronize(backend);

otherarch/sdcpp/stable-diffusion.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -161,7 +161,7 @@ class StableDiffusionGGML {
161161
bool diffusion_flash_attn) {
162162
use_tiny_autoencoder = taesd_path.size() > 0;
163163
std::string taesd_path_fixed = taesd_path;
164-
#ifdef SD_USE_CUBLAS
164+
#ifdef SD_USE_CUDA
165165
LOG_DEBUG("Using CUDA backend");
166166
backend = ggml_backend_cuda_init(0);
167167
#endif

otherarch/sdcpp/upscaler.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ struct UpscalerGGML {
1515
}
1616

1717
bool load_from_file(const std::string& esrgan_path) {
18-
#ifdef SD_USE_CUBLAS
18+
#ifdef SD_USE_CUDA
1919
LOG_DEBUG("Using CUDA backend");
2020
backend = ggml_backend_cuda_init(0);
2121
#endif

0 commit comments

Comments
 (0)