Skip to content

Commit 6080f6e

Browse files
committed
cmake : enable curl by default
1 parent 7a84777 commit 6080f6e

File tree

10 files changed

+9
-16
lines changed

10 files changed

+9
-16
lines changed

.devops/cpu.Dockerfile

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -14,9 +14,9 @@ WORKDIR /app
1414
COPY . .
1515

1616
RUN if [ "$TARGETARCH" = "amd64" ]; then \
17-
cmake -S . -B build -DCMAKE_BUILD_TYPE=Release -DLLAMA_CURL=ON -DGGML_NATIVE=OFF -DGGML_BACKEND_DL=ON -DGGML_CPU_ALL_VARIANTS=ON; \
17+
cmake -S . -B build -DCMAKE_BUILD_TYPE=Release -DGGML_NATIVE=OFF -DGGML_BACKEND_DL=ON -DGGML_CPU_ALL_VARIANTS=ON; \
1818
elif [ "$TARGETARCH" = "arm64" ]; then \
19-
cmake -S . -B build -DCMAKE_BUILD_TYPE=Release -DLLAMA_CURL=ON -DGGML_NATIVE=OFF -DGGML_CPU_ARM_ARCH=${GGML_CPU_ARM_ARCH}; \
19+
cmake -S . -B build -DCMAKE_BUILD_TYPE=Release -DGGML_NATIVE=OFF -DGGML_CPU_ARM_ARCH=${GGML_CPU_ARM_ARCH}; \
2020
else \
2121
echo "Unsupported architecture"; \
2222
exit 1; \

.devops/cuda.Dockerfile

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@ COPY . .
2121
RUN if [ "${CUDA_DOCKER_ARCH}" != "default" ]; then \
2222
export CMAKE_ARGS="-DCMAKE_CUDA_ARCHITECTURES=${CUDA_DOCKER_ARCH}"; \
2323
fi && \
24-
cmake -B build -DGGML_NATIVE=OFF -DGGML_CUDA=ON -DLLAMA_CURL=ON ${CMAKE_ARGS} -DCMAKE_EXE_LINKER_FLAGS=-Wl,--allow-shlib-undefined . && \
24+
cmake -B build -DGGML_NATIVE=OFF -DGGML_CUDA=ON ${CMAKE_ARGS} -DCMAKE_EXE_LINKER_FLAGS=-Wl,--allow-shlib-undefined . && \
2525
cmake --build build --config Release -j$(nproc)
2626

2727
RUN mkdir -p /app/lib && \

.devops/intel.Dockerfile

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@ RUN if [ "${GGML_SYCL_F16}" = "ON" ]; then \
1717
&& export OPT_SYCL_F16="-DGGML_SYCL_F16=ON"; \
1818
fi && \
1919
echo "Building with dynamic libs" && \
20-
cmake -B build -DGGML_NATIVE=OFF -DGGML_SYCL=ON -DCMAKE_C_COMPILER=icx -DCMAKE_CXX_COMPILER=icpx -DLLAMA_CURL=ON ${OPT_SYCL_F16} && \
20+
cmake -B build -DGGML_NATIVE=OFF -DGGML_SYCL=ON -DCMAKE_C_COMPILER=icx -DCMAKE_CXX_COMPILER=icpx ${OPT_SYCL_F16} && \
2121
cmake --build build --config Release -j$(nproc)
2222

2323
RUN mkdir -p /app/lib && \

.devops/musa.Dockerfile

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -35,7 +35,7 @@ COPY . .
3535
RUN if [ "${MUSA_DOCKER_ARCH}" != "default" ]; then \
3636
export CMAKE_ARGS="-DMUSA_ARCHITECTURES=${MUSA_DOCKER_ARCH}"; \
3737
fi && \
38-
cmake -B build -DGGML_NATIVE=OFF -DGGML_MUSA=ON -DLLAMA_CURL=ON ${CMAKE_ARGS} -DCMAKE_EXE_LINKER_FLAGS=-Wl,--allow-shlib-undefined . && \
38+
cmake -B build -DGGML_NATIVE=OFF -DGGML_MUSA=ON ${CMAKE_ARGS} -DCMAKE_EXE_LINKER_FLAGS=-Wl,--allow-shlib-undefined . && \
3939
cmake --build build --config Release -j$(nproc)
4040

4141
RUN mkdir -p /app/lib && \

.devops/rocm.Dockerfile

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -40,7 +40,7 @@ WORKDIR /app
4040
COPY . .
4141

4242
RUN HIPCXX="$(hipconfig -l)/clang" HIP_PATH="$(hipconfig -R)" \
43-
cmake -S . -B build -DGGML_HIP=ON -DAMDGPU_TARGETS=$ROCM_DOCKER_ARCH -DCMAKE_BUILD_TYPE=Release -DLLAMA_CURL=ON \
43+
cmake -S . -B build -DGGML_HIP=ON -DAMDGPU_TARGETS=$ROCM_DOCKER_ARCH -DCMAKE_BUILD_TYPE=Release \
4444
&& cmake --build build --config Release -j$(nproc)
4545

4646
RUN mkdir -p /app/lib \

.github/workflows/bench.yml.disabled

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -104,7 +104,6 @@ jobs:
104104
cmake -B build \
105105
-DGGML_NATIVE=OFF \
106106
-DLLAMA_BUILD_SERVER=ON \
107-
-DLLAMA_CURL=ON \
108107
-DLLAMA_CUBLAS=ON \
109108
-DCUDAToolkit_ROOT=/usr/local/cuda \
110109
-DCMAKE_CUDA_COMPILER=/usr/local/cuda/bin/nvcc \

.github/workflows/build.yml

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -62,7 +62,6 @@ jobs:
6262
cmake -B build \
6363
-DCMAKE_BUILD_RPATH="@loader_path" \
6464
-DLLAMA_FATAL_WARNINGS=ON \
65-
-DLLAMA_CURL=ON \
6665
-DGGML_METAL_USE_BF16=ON \
6766
-DGGML_METAL_EMBED_LIBRARY=ON \
6867
-DGGML_RPC=ON
@@ -133,7 +132,6 @@ jobs:
133132
cmake -B build \
134133
-DCMAKE_BUILD_RPATH="@loader_path" \
135134
-DLLAMA_FATAL_WARNINGS=ON \
136-
-DLLAMA_CURL=ON \
137135
-DGGML_METAL=OFF \
138136
-DGGML_RPC=ON
139137
cmake --build build --config Release -j $(sysctl -n hw.logicalcpu)
@@ -207,7 +205,6 @@ jobs:
207205
run: |
208206
cmake -B build \
209207
-DLLAMA_FATAL_WARNINGS=ON \
210-
-DLLAMA_CURL=ON \
211208
-DGGML_RPC=ON
212209
cmake --build build --config Release -j $(nproc)
213210

.github/workflows/server.yml

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -129,7 +129,6 @@ jobs:
129129
cmake -B build \
130130
-DGGML_NATIVE=OFF \
131131
-DLLAMA_BUILD_SERVER=ON \
132-
-DLLAMA_CURL=ON \
133132
-DCMAKE_BUILD_TYPE=${{ matrix.build_type }} \
134133
-DLLAMA_SANITIZE_${{ matrix.sanitizer }}=ON \
135134
-DGGML_OPENMP=OFF ;
@@ -142,7 +141,6 @@ jobs:
142141
cmake -B build \
143142
-DGGML_NATIVE=OFF \
144143
-DLLAMA_BUILD_SERVER=ON \
145-
-DLLAMA_CURL=ON \
146144
-DCMAKE_BUILD_TYPE=${{ matrix.build_type }} \
147145
-DLLAMA_SANITIZE_${{ matrix.sanitizer }}=ON ;
148146
cmake --build build --config ${{ matrix.build_type }} -j $(nproc) --target llama-server
@@ -154,7 +152,6 @@ jobs:
154152
cmake -B build \
155153
-DGGML_NATIVE=OFF \
156154
-DLLAMA_BUILD_SERVER=ON \
157-
-DLLAMA_CURL=ON \
158155
-DCMAKE_BUILD_TYPE=${{ matrix.build_type }} ;
159156
cmake --build build --config ${{ matrix.build_type }} -j $(nproc) --target llama-server
160157
@@ -205,7 +202,7 @@ jobs:
205202
- name: Build
206203
id: cmake_build
207204
run: |
208-
cmake -B build -DLLAMA_CURL=ON -DCURL_LIBRARY="$env:RUNNER_TEMP/libcurl/lib/libcurl.dll.a" -DCURL_INCLUDE_DIR="$env:RUNNER_TEMP/libcurl/include"
205+
cmake -B build -DCURL_LIBRARY="$env:RUNNER_TEMP/libcurl/lib/libcurl.dll.a" -DCURL_INCLUDE_DIR="$env:RUNNER_TEMP/libcurl/include"
209206
cmake --build build --config Release -j ${env:NUMBER_OF_PROCESSORS} --target llama-server
210207
211208
- name: Python setup

CMakeLists.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -81,7 +81,7 @@ option(LLAMA_BUILD_EXAMPLES "llama: build examples" ${LLAMA_STANDALONE})
8181
option(LLAMA_BUILD_SERVER "llama: build server example" ${LLAMA_STANDALONE})
8282

8383
# 3rd party libs
84-
option(LLAMA_CURL "llama: use libcurl to download model from an URL" OFF)
84+
option(LLAMA_CURL "llama: use libcurl to download model from an URL" ON)
8585
option(LLAMA_LLGUIDANCE "llama-common: include LLGuidance library for structured output in common utils" OFF)
8686

8787
# Required for relocatable CMake package

examples/server/tests/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@ To mitigate it, you can increase values in `n_predict`, `kv_size`.
1717

1818
```shell
1919
cd ../../..
20-
cmake -B build -DLLAMA_CURL=ON
20+
cmake -B build
2121
cmake --build build --target llama-server
2222
```
2323

0 commit comments

Comments
 (0)