Using a remote hosted vLLM #1510
Unanswered
anupkashyap
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I'm trying to use LiteLLM with a vllm hosted on a GPU instance. From the docs, I see examples of vLLM running on the same instance as the literLLM installation and that it requires vllm dependency (which requires cuda). However since I am just making a call to a remote cluster that hosts the vLLM, I doubt that there is any GPU requirement on the liteLLM installation. I was able to get around this issue by modifying the "custom" provider to make calls to the server and process the response format. Please help me if there is a better way to do this.
Beta Was this translation helpful? Give feedback.
All reactions