Skip to content

Commit 5e7947b

Browse files
AuditAIHcrazywoolagemini-code-assist[bot]
authored
Ollama rerank readme (#2335)
* Bump version from 0.1.0 to 0.1.1 * Add files via upload * Revise README for Ollama model integration details Updated the README to reflect changes in model integration and added instructions for integrating Ollama Rerank. * Fix typo in Ollama README regarding rerank models * Update README.md * update ollama readme * Apply suggestions from code review Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com> --------- Co-authored-by: crazywoola <100913391+crazywoola@users.noreply.github.com> Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com>
1 parent 45a0341 commit 5e7947b

File tree

3 files changed

+19
-2
lines changed

3 files changed

+19
-2
lines changed

models/ollama/README.md

Lines changed: 18 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -47,4 +47,21 @@ Click "Save" to use the model in the application after verifying that there are
4747

4848
The integration method for Embedding models is similar to LLM, just change the model type to Text Embedding.
4949

50-
For more detail, please check [Dify's official document](https://docs.dify.ai/development/models-integration/ollama).
50+
For more detail, please check [Dify's official document](https://docs.dify.ai/en/use-dify/workspace/model-providers#local-ollama).
51+
52+
#### 5. Integrate Ollama Rerank in Dify
53+
Hint: ollama officially does not support rerank models, please try locally deploying tools like vllm, llama.cpp, tei, xinference, etc., and fill in the complete URL ending with "rerank". Deployment reference [llama.cpp deployment tutorial for Qwen3-Reranker](https://github.com/AuditAIH/rerank_for_dify)
54+
55+
In `Settings > Model Providers > Ollama`, fill in:
56+
57+
![](./_assets/ollama_rerank.png)
58+
59+
- Model Name: `Qwen3-Reranker`
60+
- Base URL: `http://<your-ollama-endpoint-domain>:11434`
61+
- The plugin appends `/api/rerank` if the URL doesn't end with `/rerank`. For other services like `llama.cpp`, provide the full endpoint URL, e.g., `http://host.docker.internal:11435/v1/rerank`.
62+
- If Dify is deployed using Docker, consider using a local network IP address, e.g., `http://192.168.1.100:11434` or `http://172.17.0.1:11434` or `http://host.docker.internal:11434` to access the service.
63+
-
64+
- Model Type: `Rerank`
65+
- Model Context Length: `4096`
66+
67+
Click "Add" to use the model in the application after verifying that there are no errors.
199 KB
Loading

models/ollama/manifest.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -32,4 +32,4 @@ resource:
3232
tool:
3333
enabled: true
3434
type: plugin
35-
version: 0.1.1
35+
version: 0.1.2

0 commit comments

Comments
 (0)