File tree Expand file tree Collapse file tree 1 file changed +16
-0
lines changed Expand file tree Collapse file tree 1 file changed +16
-0
lines changed Original file line number Diff line number Diff line change @@ -244,6 +244,13 @@ Download `FP16` quantized `Qwen3` .gguf files from:
244
244
- https://huggingface.co/ggml-org/Qwen3-4B-GGUF
245
245
- https://huggingface.co/ggml-org/Qwen3-8B-GGUF
246
246
247
+ Download ` FP16 ` quantized ` Qwen2.5 ` .gguf files from:
248
+ - https://huggingface.co/bartowski/Qwen2.5-0.5B-Instruct-GGUF
249
+ - https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct-GGUF
250
+
251
+ Download ` FP16 ` quantized ` DeepSeek-R1-Distill-Qwen ` .gguf files from:
252
+ - https://huggingface.co/hdnh2006/DeepSeek-R1-Distill-Qwen-1.5B-GGUF
253
+
247
254
Please be gentle with [ huggingface.co] ( https://huggingface.co ) servers:
248
255
249
256
** Note** FP16 models are first-class citizens for the current version.
@@ -274,6 +281,15 @@ wget https://huggingface.co/ggml-org/Qwen3-0.6B-GGUF/resolve/main/Qwen3-8B-f16.g
274
281
275
282
# Phi-3-mini-4k - FP16
276
283
wget https://huggingface.co/microsoft/Phi-3-mini-4k-instruct-gguf/resolve/main/Phi-3-mini-4k-instruct-fp16.gguf
284
+
285
+ # Qwen2.5 (0.5B)
286
+ wget https://huggingface.co/bartowski/Qwen2.5-0.5B-Instruct-GGUF/resolve/main/Qwen2.5-0.5B-Instruct-f16.gguf
287
+
288
+ # Qwen2.5 (1.5B)
289
+ wget https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct-GGUF/resolve/main/qwen2.5-1.5b-instruct-fp16.gguf
290
+
291
+ # DeepSeek-R1-Distill-Qwen (1.5B)
292
+ wget https://huggingface.co/hdnh2006/DeepSeek-R1-Distill-Qwen-1.5B-GGUF/resolve/main/DeepSeek-R1-Distill-Qwen-1.5B-F16.gguf
277
293
```
278
294
279
295
** [ Experimental] ** you can download the Q8 and Q4 used in the original implementation of Llama3.java, but for now are going to be dequanted to FP16 for TornadoVM support:
You can’t perform that action at this time.
0 commit comments