How to use Vertex AI Llama 405B API (whilst it is still free) #3835
Replies: 1 comment
-
i was wondering the same |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Firstly, become a massive fan of LibreChat - to the point I've connected our Azure GPT4o, Google Gemini 1.5 Flash & Pro plus numerous Ollama offline open source models (Phi, Llama3, Gemma2) seamlessly.
However, I'm stuck with how I can connect (or maybe I can't?) the Llama 3.1 405b API (currently free) from Google / Vertex AI.
Note: I have a valid, billable Google Cloud / Vertex AI account already.
I have an existing PROJECT_ID.
The instructions, which work with Postman etc are as follows:
Try Llama 3.1 API Service
To use Llama 3.1 API service with the command line interface (CLI), do the following:
Is there anyway to integrate this into Libre? (Even if it does mean I have to get a new bearer token every 60 mins)
Many thanks for any advice or guidance.
Beta Was this translation helpful? Give feedback.
All reactions