Skip to content

Commit 2d8cb97

Browse files
committed
readme update
1 parent 411cf30 commit 2d8cb97

File tree

1 file changed

+9
-4
lines changed

1 file changed

+9
-4
lines changed

readme.md

Lines changed: 9 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,6 @@
1111
<a href="https://github.com/lemonade-sdk/lemonade"><img src="https://img.shields.io/badge/Lemonade-openai-lightblue.svg" alt="Lemonade AI: OpenAI Compatible"></a>
1212
<a href="https://github.com/InternLM/lmdeploy"><img src="https://img.shields.io/badge/LM Deploy-openai-lightblue.svg" alt="Lemonade AI: OpenAI Compatible"></a>
1313
</P>
14-
<br />
1514
<p>
1615
<a href="https://thushan.github.io/olla/"><img src="https://img.shields.io/badge/📖_Documentation-0078D4?style=for-the-badge&logoColor=white" height="32" alt="Documentation"></a>
1716
&nbsp;
@@ -45,10 +44,12 @@ In the above example, we configure [Jetbrains Junie](https://www.jetbrains.com/j
4544
- **🛡️ Production Ready**: Rate limiting, request size limits, graceful shutdown
4645
- **⚡ High Performance**: Sub-millisecond endpoint selection with lock-free atomic stats
4746
- **🎯 LLM-Optimised**: Streaming-first design with optimised timeouts for long inference
48-
- **⚙️ High Performance**: Designed to be very lightweight & efficient, runs on less than 50Mb RAM. See [best practices](https://thushan.github.io/olla/configuration/practices/performance/)
47+
- **⚙️ High Performance**: Designed to be very [lightweight & efficient](https://thushan.github.io/olla/configuration/practices/performance/), runs on less than 50Mb RAM.
4948

5049
### Supported Backends
5150

51+
Olla natively supports the following backend providers. Learn more about [Olla Integrations](https://thushan.github.io/olla/integrations/overview/).
52+
5253
* [Ollama](https://github.com/ollama/ollama) - native support for Ollama, including model unification. \
5354
Use: `/olla/ollama/`
5455
* [LM Studio](https://lmstudio.ai/) - native support for LMStudio, including model unification. \
@@ -61,13 +62,15 @@ In the above example, we configure [Jetbrains Junie](https://www.jetbrains.com/j
6162

6263
### Supported Frontends
6364

64-
* [OpenWebUI](https://github.com/open-webui/open-webui) - You can route you OpenWebUI Ollama endpoint to Olla and serve unified models easily. [See Examples](https://github.com/thushan/olla/tree/main/examples/ollama-openwebui)
65+
* [OpenWebUI](https://github.com/open-webui/open-webui) - You can route you [OpenWebUI Ollama endpoint to Olla](https://thushan.github.io/olla/integrations/frontend/openwebui/) and serve unified models easily.
6566

6667
Coming soon - but you can use the OpenAI compatibility in the interim:
6768

6869
* [LMDeploy](https://github.com/InternLM/lmdeploy)
6970
* [Lemonade](https://github.com/lemonade-sdk/lemonade)
7071

72+
Learn more about [Olla's Profile System](https://thushan.github.io/olla/concepts/profile-system/).
73+
7174
## Platform Support
7275

7376
Olla runs on multiple platforms and architectures:
@@ -77,7 +80,7 @@ Olla runs on multiple platforms and architectures:
7780
| Linux ||| Full support including Raspberry Pi 4+ |
7881
| macOS ||| Intel and Apple Silicon (M1/M2/M3/M4) |
7982
| Windows ||| Windows 10/11 and Windows on ARM |
80-
| Docker ||| Multi-architecture images available |
83+
| Docker ||| Multi-architecture images (amd64/arm64) |
8184

8285
## Quick Start
8386

@@ -145,6 +148,8 @@ Complete setup with [OpenWebUI](https://github.com/open-webui/open-webui) + Olla
145148
# Access OpenWebUI at http://localhost:3000
146149
```
147150

151+
You can learn more about [OpenWebUI Ollama with Olla](https://thushan.github.io/olla/integrations/frontend/openwebui/).
152+
148153
More examples coming soon:
149154
- **Multi-Provider Setup**: Ollama + LM Studio + OpenAI-compatible endpoints
150155
- **High-Availability**: Production deployment with failover

0 commit comments

Comments
 (0)