You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -192,7 +192,6 @@ Instructions for adding support for new models: [HOWTO-add-model.md](docs/develo
192
192
-[crashr/gppm](https://github.com/crashr/gppm) – launch llama.cpp instances utilizing NVIDIA Tesla P40 or P100 GPUs with reduced idle power consumption
193
193
-[gpustack/gguf-parser](https://github.com/gpustack/gguf-parser-go/tree/main/cmd/gguf-parser) - review/check the GGUF file and estimate the memory usage
194
194
-[Styled Lines](https://marketplace.unity.com/packages/tools/generative-ai/styled-lines-llama-cpp-model-292902) (proprietary licensed, async wrapper of inference part for game development in Unity3d with pre-built Mobile and Web platform wrappers and a model example)
195
-
-[llama-swap](https://github.com/mostlygeek/llama-swap) - transparent proxy for automatic model switching with llama-server
196
195
197
196
</details>
198
197
@@ -202,6 +201,7 @@ Instructions for adding support for new models: [HOWTO-add-model.md](docs/develo
202
201
-[Paddler](https://github.com/distantmagic/paddler) - Stateful load balancer custom-tailored for llama.cpp
203
202
-[GPUStack](https://github.com/gpustack/gpustack) - Manage GPU clusters for running LLMs
204
203
-[llama_cpp_canister](https://github.com/onicai/llama_cpp_canister) - llama.cpp as a smart contract on the Internet Computer, using WebAssembly
204
+
-[llama-swap](https://github.com/mostlygeek/llama-swap) - transparent proxy that adds automatic model switching with llama-server
0 commit comments