Skip to content

Commit 685bed3

Browse files
Use wget for model downloads instead of llama-cli
Since we disabled CURL support (-DLLAMA_CURL=OFF), llama-cli cannot download models from HuggingFace. Switch to using wget directly to download the TinyLlama model, following the same pattern used in build.yml workflow. Fixes: Model download failures in performance-cpu and performance-metal jobs Co-Authored-By: Alex Peng <[email protected]>
1 parent 8c28faf commit 685bed3

File tree

1 file changed

+6
-13
lines changed

1 file changed

+6
-13
lines changed

.github/workflows/performance-regression.yml

Lines changed: 6 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -83,10 +83,8 @@ jobs:
8383
mkdir -p models
8484
# Download TinyLlama test model if not present
8585
if [ ! -f models/tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf ]; then
86-
./build/bin/llama-cli --hf-repo TinyLlama/TinyLlama-1.1B-Chat-v1.0 \
87-
--hf-file tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf \
88-
--model-download-only
89-
mv ~/.cache/llama.cpp/*/tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf models/
86+
wget -q --show-progress -O models/tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf \
87+
https://huggingface.co/TheBloke/TinyLlama-1.1B-Chat-v1.0-GGUF/resolve/main/tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf
9088
fi
9189
9290
- name: Restore baseline database
@@ -219,10 +217,8 @@ jobs:
219217
run: |
220218
mkdir -p models
221219
if [ ! -f models/tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf ]; then
222-
./build/bin/llama-cli --hf-repo TinyLlama/TinyLlama-1.1B-Chat-v1.0 \
223-
--hf-file tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf \
224-
--model-download-only
225-
mv ~/.cache/llama.cpp/*/tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf models/
220+
wget -q --show-progress -O models/tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf \
221+
https://huggingface.co/TheBloke/TinyLlama-1.1B-Chat-v1.0-GGUF/resolve/main/tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf
226222
fi
227223
228224
- name: Restore baseline database
@@ -353,11 +349,8 @@ jobs:
353349
run: |
354350
mkdir -p models
355351
if [ ! -f models/tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf ]; then
356-
./build/bin/llama-cli --hf-repo TinyLlama/TinyLlama-1.1B-Chat-v1.0 \
357-
--hf-file tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf \
358-
--model-download-only
359-
mv ~/.cache/llama.cpp/*/tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf models/ || \
360-
mv ~/Library/Caches/llama.cpp/*/tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf models/
352+
wget -q --show-progress -O models/tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf \
353+
https://huggingface.co/TheBloke/TinyLlama-1.1B-Chat-v1.0-GGUF/resolve/main/tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf
361354
fi
362355
363356
- name: Restore baseline database

0 commit comments

Comments
 (0)