Skip to content

Commit dc329fe

Browse files
image update
1 parent 8c49b64 commit dc329fe

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

cloud-infrastructure/ai-infra-gpu/ai-infrastructure/litellm/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,8 +1,8 @@
11
# Calling multiple vLLM inference servers using LiteLLM
22

3-
In this tutorial we explain how to use a LiteLLM Proxy Server to call multiple LLM inference endpoints from a single interface. LiteLLM interacts will 100+ LLMs such as OpenAI, Coheren, NVIDIA Triton and NIM, etc. Here we will use two vLLM inference servers.
3+
In this tutorial we explain how to use a LiteLLM Proxy Server to call multiple LLM inference endpoints from a single interface. LiteLLM interacts will 100+ LLMs such as OpenAI, Cohere, NVIDIA Triton and NIM, etc. Here we will use two vLLM inference servers.
44

5-
![Hybrid shards](assets/images/litellm.avif "LiteLLM")
5+
![Hybrid shards](assets/images/litellm.png "LiteLLM")
66

77
## Introduction
88

51.4 KB
Loading

0 commit comments

Comments
 (0)