Skip to content

Commit 04488ee

Browse files
authored
add DeepSeek-R1-Distill-Qwen-14B-int4-ov (#956)
* Improve interface compatibility Add Dockerfile for Linux platform * del unimportant comments * Update README.md * add DeepSeek-R1-Distill-Qwen-14B-int4-ov
1 parent fad59bb commit 04488ee

File tree

1 file changed

+1
-0
lines changed

1 file changed

+1
-0
lines changed

modules/ollama_openvino/README.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -615,6 +615,7 @@ The native Ollama only supports models in the GGUF format, the Ollama-OV invoke
615615
| DeepSeek-R1-Distill-Qwen-1.5B-int4-ov-npu | 1.5B | 1.1GB | INT4_SYM_CW | [ModelScope](https://modelscope.cn/models/zhaohb/DeepSeek-R1-Distill-Qwen-1.5B-int4-ov-npu/summary) | NPU(best) |
616616
| DeepSeek-R1-Distill-Qwen-7B-int4-ov | 7B | 4.3GB | INT4_SYM_128 | [ModelScope](https://modelscope.cn/models/zhaohb/DeepSeek-R1-Distill-Qwen-7B-int4-ov) | CPU, GPU, NPU(base) |
617617
| DeepSeek-R1-Distill-Qwen-7B-int4-ov-npu | 7B | 4.1GB | INT4_SYM_CW | [ModelScope](https://modelscope.cn/models/zhaohb/DeepSeek-R1-Distill-Qwen-7B-int4-ov-npu) | NPU(best) |
618+
| DeepSeek-R1-Distill-Qwen-14B-int4-ov | 14B | 8.0GB | INT4_SYM_128 | [ModelScope](https://modelscope.cn/models/zhaohb/DeepSeek-R1-Distill-Qwen-14B-int4-ov) | CPU, GPU, NPU(base) |
618619
| DeepSeek-R1-Distill-llama-8B-int4-ov | 8B | 4.5GB | INT4_SYM_128 | [ModelScope](https://modelscope.cn/models/zhaohb/DeepSeek-R1-Distill-Llama-8B-int4-ov) | CPU, GPU, NPU(base) |
619620
| DeepSeek-R1-Distill-llama-8B-int4-ov-npu | 8B | 4.2GB | INT4_SYM_CW | [ModelScope](https://modelscope.cn/models/zhaohb/DeepSeek-R1-Distill-Llama-8B-int4-ov-npu) | NPU(best) |
620621
| llama-3.2-1b-instruct-int4-ov | 1B | 0.8GB | INT4_SYM_128 | [ModelScope](https://modelscope.cn/models/FionaZhao/llama-3.2-1b-instruct-int4-ov/files) | CPU, GPU, NPU(base) |

0 commit comments

Comments
 (0)