Skip to content

Commit 82ab6bf

Browse files
authored
Add documentation how to use with llama.cpp (#318)
## Summary Add documentation on how to run guidellm with llama.cpp server ## Details guidellm can run against a llama.cpp server when the model metadata is prefetched and the server is started with the right arguments. ## Test Plan Verified that guidellm runs successfully when following the documented steps ## Related Issues n/a --- - [x] "I certify that all code in this PR is my own, except as noted below." ## Use of AI - [ ] Includes AI-assisted code completion - [ ] Includes code generated by an AI application - [ ] Includes AI-generated tests (NOTE: AI written tests should have a docstring that includes `## WRITTEN BY AI ##`) Signed-off-by: Radoslav Gerganov <[email protected]>
1 parent 9573703 commit 82ab6bf

File tree

1 file changed

+18
-0
lines changed

1 file changed

+18
-0
lines changed

docs/backends.md

Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -40,6 +40,24 @@ docker run --gpus 1 -ti --shm-size 1g --ipc=host --rm -p 8080:80 \
4040

4141
For more information on starting a TGI server, see the [TGI Documentation](https://huggingface.co/docs/text-generation-inference/index).
4242

43+
### 3. llama.cpp
44+
45+
[llama.cpp](https://github.com/ggml-org/llama.cpp) provides lightweight, OpenAI-compatible server through its [llama-server](https://github.com/ggml-org/llama.cpp/blob/master/tools/server) tool.
46+
47+
To start a llama.cpp server with the gpt-oss-20b model, you can use the following command:
48+
49+
```bash
50+
llama-server -hf ggml-org/gpt-oss-20b-GGUF --alias gpt-oss-20b --ctx-size 0 --jinja -ub 2048 -b 2048
51+
```
52+
53+
Note that we are providing an alias `gpt-oss-20b` for the model name because `guidellm` is using it to retrieve model metadata in JSON format and such metadata is not included in GGUF model repositories. A simple workaround is to download the metadata files from safetensors repository and place them in a local directory named after the alias:
54+
55+
```bash
56+
huggingface-cli download openai/gpt-oss-20b --include "*.json" --local-dir gpt-oss-20b/
57+
```
58+
59+
Now you can run `guidellm` as usual and it will be able to fetch the model metadata from the local directory.
60+
4361
## Expanding Backend Support
4462

4563
GuideLLM is an open platform, and we encourage contributions to extend its backend support. Whether it's adding new server implementations, integrating with Python-based backends, or enhancing existing capabilities, your contributions are welcome. For more details on how to contribute, see the [CONTRIBUTING.md](https://github.com/vllm-project/guidellm/blob/main/CONTRIBUTING.md) file.

0 commit comments

Comments
 (0)