You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
@@ -45,10 +44,12 @@ In the above example, we configure [Jetbrains Junie](https://www.jetbrains.com/j
45
44
-**🛡️ Production Ready**: Rate limiting, request size limits, graceful shutdown
46
45
-**⚡ High Performance**: Sub-millisecond endpoint selection with lock-free atomic stats
47
46
-**🎯 LLM-Optimised**: Streaming-first design with optimised timeouts for long inference
48
-
-**⚙️ High Performance**: Designed to be very lightweight & efficient, runs on less than 50Mb RAM. See [best practices](https://thushan.github.io/olla/configuration/practices/performance/)
47
+
-**⚙️ High Performance**: Designed to be very [lightweight & efficient](https://thushan.github.io/olla/configuration/practices/performance/), runs on less than 50Mb RAM.
49
48
50
49
### Supported Backends
51
50
51
+
Olla natively supports the following backend providers. Learn more about [Olla Integrations](https://thushan.github.io/olla/integrations/overview/).
52
+
52
53
*[Ollama](https://github.com/ollama/ollama) - native support for Ollama, including model unification. \
53
54
Use: `/olla/ollama/`
54
55
*[LM Studio](https://lmstudio.ai/) - native support for LMStudio, including model unification. \
@@ -61,13 +62,15 @@ In the above example, we configure [Jetbrains Junie](https://www.jetbrains.com/j
61
62
62
63
### Supported Frontends
63
64
64
-
*[OpenWebUI](https://github.com/open-webui/open-webui) - You can route you OpenWebUI Ollama endpoint to Olla and serve unified models easily. [See Examples](https://github.com/thushan/olla/tree/main/examples/ollama-openwebui)
65
+
*[OpenWebUI](https://github.com/open-webui/open-webui) - You can route you [OpenWebUI Ollama endpoint to Olla](https://thushan.github.io/olla/integrations/frontend/openwebui/) and serve unified models easily.
65
66
66
67
Coming soon - but you can use the OpenAI compatibility in the interim:
0 commit comments