Skip to content

Commit 6a102ac

Browse files
julien-cngxson
andauthored
ollama tweaks (#1448)
* ollama tweaks * suggested by @pcuenca * move TOC --------- Co-authored-by: Xuan Son Nguyen <[email protected]>
1 parent 481c63a commit 6a102ac

File tree

2 files changed

+9
-7
lines changed

2 files changed

+9
-7
lines changed

docs/hub/_toctree.yml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -144,8 +144,8 @@
144144
title: GGUF usage with llama.cpp
145145
- local: gguf-gpt4all
146146
title: GGUF usage with GPT4All
147-
- local: ollama
148-
title: Use Ollama with GGUF Model
147+
- local: ollama
148+
title: GGUF usage with Ollama
149149
- title: Datasets
150150
local: datasets
151151
isExpanded: true

docs/hub/ollama.md

Lines changed: 7 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ ollama run hf.co/{username}/{repository}
1212

1313
Please note that you can use both `hf.co` and `huggingface.co` as the domain name.
1414

15-
Here are some other models that you can try:
15+
Here are some models you can try:
1616

1717
```sh
1818
ollama run hf.co/bartowski/Llama-3.2-1B-Instruct-GGUF
@@ -23,7 +23,9 @@ ollama run hf.co/bartowski/Humanish-LLama3-8B-Instruct-GGUF
2323

2424
## Custom Quantization
2525

26-
By default, the `Q4_K_M` quantization scheme is used. To select a different scheme, simply add a tag:
26+
By default, the `Q4_K_M` quantization scheme is used, when it's present inside the model repo. If not, we default to picking one reasonable quant type present inside the repo.
27+
28+
To select a different scheme, simply add a tag:
2729

2830
```sh
2931
ollama run hf.co/{username}/{repository}:{quantization}
@@ -40,15 +42,15 @@ ollama run hf.co/bartowski/Llama-3.2-3B-Instruct-GGUF:Q8_0
4042
# the quantization name is case-insensitive, this will also work
4143
ollama run hf.co/bartowski/Llama-3.2-3B-Instruct-GGUF:iq3_m
4244

43-
# you can also select a specific file
45+
# you can also directly use the full filename as a tag
4446
ollama run hf.co/bartowski/Llama-3.2-3B-Instruct-GGUF:Llama-3.2-3B-Instruct-IQ3_M.gguf
4547
```
4648

4749
## Custom Chat Template and Parameters
4850

4951
By default, a template will be selected automatically from a list of commonly used templates. It will be selected based on the built-in `tokenizer.chat_template` metadata stored inside the GGUF file.
5052

51-
If your GGUF file doesn't have a built-in template or uses a custom chat template, you can create a new file called `template` in the repository. The template must be a Go template, not a Jinja template. Here's an example:
53+
If your GGUF file doesn't have a built-in template or if you want to customize your chat template, you can create a new file called `template` in the repository. The template must be a Go template, not a Jinja template. Here's an example:
5254

5355
```
5456
{{ if .System }}<|system|>
@@ -59,7 +61,7 @@ If your GGUF file doesn't have a built-in template or uses a custom chat templat
5961
{{ .Response }}<|end|>
6062
```
6163

62-
To know more about Go template format, please refer to [this documentation](https://github.com/ollama/ollama/blob/main/docs/template.md)
64+
To know more about the Go template format, please refer to [this documentation](https://github.com/ollama/ollama/blob/main/docs/template.md)
6365

6466
You can optionally configure a system prompt by putting it into a new file named `system` in the repository.
6567

0 commit comments

Comments
 (0)