You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/hub/ollama.md
+7-5Lines changed: 7 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -12,7 +12,7 @@ ollama run hf.co/{username}/{repository}
12
12
13
13
Please note that you can use both `hf.co` and `huggingface.co` as the domain name.
14
14
15
-
Here are some other models that you can try:
15
+
Here are some models you can try:
16
16
17
17
```sh
18
18
ollama run hf.co/bartowski/Llama-3.2-1B-Instruct-GGUF
@@ -23,7 +23,9 @@ ollama run hf.co/bartowski/Humanish-LLama3-8B-Instruct-GGUF
23
23
24
24
## Custom Quantization
25
25
26
-
By default, the `Q4_K_M` quantization scheme is used. To select a different scheme, simply add a tag:
26
+
By default, the `Q4_K_M` quantization scheme is used, when it's present inside the model repo. If not, we default to picking one reasonable quant type present inside the repo.
27
+
28
+
To select a different scheme, simply add a tag:
27
29
28
30
```sh
29
31
ollama run hf.co/{username}/{repository}:{quantization}
@@ -40,15 +42,15 @@ ollama run hf.co/bartowski/Llama-3.2-3B-Instruct-GGUF:Q8_0
40
42
# the quantization name is case-insensitive, this will also work
41
43
ollama run hf.co/bartowski/Llama-3.2-3B-Instruct-GGUF:iq3_m
42
44
43
-
# you can also select a specific file
45
+
# you can also directly use the full filename as a tag
44
46
ollama run hf.co/bartowski/Llama-3.2-3B-Instruct-GGUF:Llama-3.2-3B-Instruct-IQ3_M.gguf
45
47
```
46
48
47
49
## Custom Chat Template and Parameters
48
50
49
51
By default, a template will be selected automatically from a list of commonly used templates. It will be selected based on the built-in `tokenizer.chat_template` metadata stored inside the GGUF file.
50
52
51
-
If your GGUF file doesn't have a built-in template or uses a custom chat template, you can create a new file called `template` in the repository. The template must be a Go template, not a Jinja template. Here's an example:
53
+
If your GGUF file doesn't have a built-in template or if you want to customize your chat template, you can create a new file called `template` in the repository. The template must be a Go template, not a Jinja template. Here's an example:
52
54
53
55
```
54
56
{{ if .System }}<|system|>
@@ -59,7 +61,7 @@ If your GGUF file doesn't have a built-in template or uses a custom chat templat
59
61
{{ .Response }}<|end|>
60
62
```
61
63
62
-
To know more about Go template format, please refer to [this documentation](https://github.com/ollama/ollama/blob/main/docs/template.md)
64
+
To know more about the Go template format, please refer to [this documentation](https://github.com/ollama/ollama/blob/main/docs/template.md)
63
65
64
66
You can optionally configure a system prompt by putting it into a new file named `system` in the repository.
0 commit comments