Skip to content

Commit e7dfdfc

Browse files
committed
upd llama.cpp docs
1 parent cffc8ee commit e7dfdfc

File tree

1 file changed

+3
-5
lines changed

1 file changed

+3
-5
lines changed

docs/hub/gguf-llamacpp.md

Lines changed: 3 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -31,8 +31,7 @@ Once installed, you can use the `llama-cli` or `llama-server` as follows:
3131

3232
```bash
3333
llama-cli
34-
--hf-repo lmstudio-community/Meta-Llama-3-8B-Instruct-GGUF \
35-
--hf-file Meta-Llama-3-8B-Instruct-Q8_0.gguf \
34+
-hf bartowski/Llama-3.2-3B-Instruct-GGUF:Q8_0 \
3635
-p "You are a helpful assistant" -cnv
3736
```
3837

@@ -42,8 +41,7 @@ Additionally, you can invoke an OpenAI spec chat completions endpoint directly u
4241

4342
```bash
4443
llama-server \
45-
--hf-repo lmstudio-community/Meta-Llama-3-8B-Instruct-GGUF \
46-
--hf-file Meta-Llama-3-8B-Instruct-Q8_0.gguf
44+
-hf bartowski/Llama-3.2-3B-Instruct-GGUF:Q8_0
4745
```
4846

4947
After running the server you can simply utilise the endpoint as below:
@@ -66,6 +64,6 @@ curl http://localhost:8080/v1/chat/completions \
6664
}'
6765
```
6866

69-
Replace `--hf-repo` with any valid Hugging Face hub repo name and `--hf-file` with the GGUF file name in the hub repo - off you go! 🦙
67+
Replace `-hf` with any valid Hugging Face hub repo name - off you go! 🦙
7068

7169
Note: Remember to `build` llama.cpp with `LLAMA_CURL=1` :)

0 commit comments

Comments
 (0)