Skip to content

Commit b46e8ec

Browse files
committed
readme: add llama-swap to Infrastructure
1 parent 61037d7 commit b46e8ec

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -192,7 +192,6 @@ Instructions for adding support for new models: [HOWTO-add-model.md](docs/develo
192192
- [crashr/gppm](https://github.com/crashr/gppm) – launch llama.cpp instances utilizing NVIDIA Tesla P40 or P100 GPUs with reduced idle power consumption
193193
- [gpustack/gguf-parser](https://github.com/gpustack/gguf-parser-go/tree/main/cmd/gguf-parser) - review/check the GGUF file and estimate the memory usage
194194
- [Styled Lines](https://marketplace.unity.com/packages/tools/generative-ai/styled-lines-llama-cpp-model-292902) (proprietary licensed, async wrapper of inference part for game development in Unity3d with pre-built Mobile and Web platform wrappers and a model example)
195-
- [llama-swap](https://github.com/mostlygeek/llama-swap) - transparent proxy for automatic model switching with llama-server
196195

197196
</details>
198197

@@ -202,6 +201,7 @@ Instructions for adding support for new models: [HOWTO-add-model.md](docs/develo
202201
- [Paddler](https://github.com/distantmagic/paddler) - Stateful load balancer custom-tailored for llama.cpp
203202
- [GPUStack](https://github.com/gpustack/gpustack) - Manage GPU clusters for running LLMs
204203
- [llama_cpp_canister](https://github.com/onicai/llama_cpp_canister) - llama.cpp as a smart contract on the Internet Computer, using WebAssembly
204+
- [llama-swap](https://github.com/mostlygeek/llama-swap) - transparent proxy that adds automatic model switching with llama-server
205205

206206
</details>
207207

0 commit comments

Comments
 (0)