You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/hub/ollama.md
+21-3Lines changed: 21 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -6,6 +6,16 @@ Ollama is an application based on llama.cpp to interact with LLMs directly throu
6
6
7
7
Getting started is as simple as:
8
8
9
+
1. Enable `ollama` under your [Local Apps settings](https://huggingface.co/settings/local-apps).
10
+
2. On a model page, choose `ollama` from `Use this model` dropdown. For example: [bartowski/Llama-3.2-1B-Instruct-GGUF](https://huggingface.co/bartowski/Llama-3.2-1B-Instruct-GGUF).
@@ -25,14 +35,22 @@ ollama run hf.co/bartowski/Humanish-LLama3-8B-Instruct-GGUF
25
35
26
36
By default, the `Q4_K_M` quantization scheme is used, when it's present inside the model repo. If not, we default to picking one reasonable quant type present inside the repo.
27
37
28
-
To select a different scheme, simply add a tag:
38
+
To select a different scheme, simply:
39
+
40
+
1. From `Files and versions` tab on a model page, open GGUF viewer on a particular GGUF file.
41
+
2. Choose `ollama` from `Use this model` dropdown.
0 commit comments