Skip to content

Commit 16a0496

Browse files
Merge pull request #337 from janhq/update-dev-from-master-2025-11-24-00-38
Sync master with upstream release b7136
2 parents 3de2a4a + d5bc1ad commit 16a0496

File tree

15 files changed

+314
-766
lines changed

15 files changed

+314
-766
lines changed

.devops/vulkan.Dockerfile

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,7 @@
1-
ARG UBUNTU_VERSION=25.10
1+
ARG UBUNTU_VERSION=26.04
22

33
FROM ubuntu:$UBUNTU_VERSION AS build
44

5-
# Ref: https://vulkan.lunarg.com/doc/sdk/latest/linux/getting_started.html
6-
75
# Install build tools
86
RUN apt update && apt install -y git build-essential cmake wget xz-utils
97

.github/workflows/server.yml

Lines changed: 1 addition & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -351,16 +351,10 @@ jobs:
351351
fetch-depth: 0
352352
ref: ${{ github.event.inputs.sha || github.event.pull_request.head.sha || github.sha || github.head_ref || github.ref_name }}
353353

354-
- name: libCURL
355-
id: get_libcurl
356-
uses: ./.github/actions/windows-setup-curl
357-
358354
- name: Build
359355
id: cmake_build
360-
env:
361-
CURL_PATH: ${{ steps.get_libcurl.outputs.curl_path }}
362356
run: |
363-
cmake -B build -DCURL_LIBRARY="$env:CURL_PATH/lib/libcurl.dll.a" -DCURL_INCLUDE_DIR="$env:CURL_PATH/include"
357+
cmake -B build -DLLAMA_CURL=OFF -DLLAMA_BUILD_BORINGSSL=ON
364358
cmake --build build --config Release -j ${env:NUMBER_OF_PROCESSORS} --target llama-server
365359
366360
- name: Python setup
@@ -374,13 +368,6 @@ jobs:
374368
run: |
375369
pip install -r tools/server/tests/requirements.txt
376370
377-
- name: Copy Libcurl
378-
id: prepare_libcurl
379-
env:
380-
CURL_PATH: ${{ steps.get_libcurl.outputs.curl_path }}
381-
run: |
382-
cp $env:CURL_PATH/bin/libcurl-x64.dll ./build/bin/Release/libcurl-x64.dll
383-
384371
- name: Tests
385372
id: server_integration_tests
386373
if: ${{ !matrix.disabled_on_pr || !github.event.pull_request }}

ggml/src/ggml-cuda/cpy-utils.cuh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -212,6 +212,6 @@ static __device__ void cpy_blck_f32_iq4_nl(const char * cxi, char * cdsti) {
212212
}
213213

214214
template<typename src_t, typename dst_t>
215-
static __device__ void cpy_1_flt(const char * cxi, char * cdsti) {
215+
static __device__ void cpy_1_scalar(const char * cxi, char * cdsti) {
216216
*(dst_t *) cdsti = ggml_cuda_cast<dst_t>(*(const src_t *) cxi);
217217
}

ggml/src/ggml-cuda/cpy.cu

Lines changed: 85 additions & 47 deletions
Large diffs are not rendered by default.

ggml/src/ggml-cuda/ggml-cuda.cu

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -4115,6 +4115,9 @@ static bool ggml_backend_cuda_device_supports_op(ggml_backend_dev_t dev, const g
41154115
if (src0_type == GGML_TYPE_I32 && src1_type == GGML_TYPE_F32) {
41164116
return true;
41174117
}
4118+
if (src0_type == GGML_TYPE_I32 && src1_type == GGML_TYPE_I32) {
4119+
return true;
4120+
}
41184121
if (src0_type == src1_type && ggml_is_contiguous(op->src[0]) && ggml_is_contiguous(op->src[1])) {
41194122
return true;
41204123
}

0 commit comments

Comments
 (0)