Skip to content

Commit 1403b4f

Browse files
Update readme with model links
1 parent eead727 commit 1403b4f

File tree

1 file changed

+16
-0
lines changed

1 file changed

+16
-0
lines changed

README.md

Lines changed: 16 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -244,6 +244,13 @@ Download `FP16` quantized `Qwen3` .gguf files from:
244244
- https://huggingface.co/ggml-org/Qwen3-4B-GGUF
245245
- https://huggingface.co/ggml-org/Qwen3-8B-GGUF
246246

247+
Download `FP16` quantized `Qwen2.5` .gguf files from:
248+
- https://huggingface.co/bartowski/Qwen2.5-0.5B-Instruct-GGUF
249+
- https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct-GGUF
250+
251+
Download `FP16` quantized `DeepSeek-R1-Distill-Qwen` .gguf files from:
252+
- https://huggingface.co/hdnh2006/DeepSeek-R1-Distill-Qwen-1.5B-GGUF
253+
247254
Please be gentle with [huggingface.co](https://huggingface.co) servers:
248255

249256
**Note** FP16 models are first-class citizens for the current version.
@@ -274,6 +281,15 @@ wget https://huggingface.co/ggml-org/Qwen3-0.6B-GGUF/resolve/main/Qwen3-8B-f16.g
274281
275282
# Phi-3-mini-4k - FP16
276283
wget https://huggingface.co/microsoft/Phi-3-mini-4k-instruct-gguf/resolve/main/Phi-3-mini-4k-instruct-fp16.gguf
284+
285+
# Qwen2.5 (0.5B)
286+
wget https://huggingface.co/bartowski/Qwen2.5-0.5B-Instruct-GGUF/resolve/main/Qwen2.5-0.5B-Instruct-f16.gguf
287+
288+
# Qwen2.5 (1.5B)
289+
wget https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct-GGUF/resolve/main/qwen2.5-1.5b-instruct-fp16.gguf
290+
291+
# DeepSeek-R1-Distill-Qwen (1.5B)
292+
wget https://huggingface.co/hdnh2006/DeepSeek-R1-Distill-Qwen-1.5B-GGUF/resolve/main/DeepSeek-R1-Distill-Qwen-1.5B-F16.gguf
277293
```
278294

279295
**[Experimental]** you can download the Q8 and Q4 used in the original implementation of Llama3.java, but for now are going to be dequanted to FP16 for TornadoVM support:

0 commit comments

Comments
 (0)