diff --git a/modules/ollama_openvino/README.md b/modules/ollama_openvino/README.md index 8f0853c55..01bafc2d0 100644 --- a/modules/ollama_openvino/README.md +++ b/modules/ollama_openvino/README.md @@ -611,7 +611,7 @@ docker run --rm -it ollama_openvino:v1 The native Ollama only supports models in the GGUF format, the Ollama-OV invoke OpenVINO GenAI which requires models in the OpenVINO format. Therefore, we have enabled support for OpenVINO model files in Ollama. For public LLMs, you can access and download OpenVINO IR model from HuggingFace or ModelScope: | Model | Parameters | Size |Compression | Download | Device | | ------------------ | ---------- | ----- | -----------|-------------------- |----------| -| DeepSeek-R1-Distill-Qwen-1.5B-int4-ov | 1.5B | 1.1GB | INT4_SYM_128 | [ModelScope](https://modelscope.cn/models/zhaohb/DeepSeek-R1-Distill-Qwen-1.5B-int4-ov/summary) | CPU, GPU, NPU(base) | +| DeepSeek-R1-Distill-Qwen-1.5B-int4-ov | 1.5B | 1.4GB | INT4_ASYM_32 | [ModelScope](https://modelscope.cn/models/zhaohb/DeepSeek-R1-Distill-Qwen-1.5B-int4-gs-32-ov) | CPU, GPU, NPU(base) | | DeepSeek-R1-Distill-Qwen-1.5B-int4-ov-npu | 1.5B | 1.1GB | INT4_SYM_CW | [ModelScope](https://modelscope.cn/models/zhaohb/DeepSeek-R1-Distill-Qwen-1.5B-int4-ov-npu/summary) | NPU(best) | | DeepSeek-R1-Distill-Qwen-7B-int4-ov | 7B | 4.3GB | INT4_SYM_128 | [ModelScope](https://modelscope.cn/models/zhaohb/DeepSeek-R1-Distill-Qwen-7B-int4-ov) | CPU, GPU, NPU(base) | | DeepSeek-R1-Distill-Qwen-7B-int4-ov-npu | 7B | 4.1GB | INT4_SYM_CW | [ModelScope](https://modelscope.cn/models/zhaohb/DeepSeek-R1-Distill-Qwen-7B-int4-ov-npu) | NPU(best) |