-
Beta Was this translation helpful? Give feedback.
Replies: 3 comments 2 replies
-
|
Solution: to get the "/api/chat/completions" endpoint working with ollama-hosted models, one need to configure the config.toml as if it was an openai-model-provider: [models.providers.openai]
models = ["llama3.3", "gpt-oss", "mistral-large"]
api_key_env = "OPENAI_API_KEY"
base_url = "https://123.45.67.89/api" |
Beta Was this translation helpful? Give feedback.
-
|
good to see you found a workaround kapsner -- can you let me know more about how your ollama instance is set up/hosted? |
Beta Was this translation helpful? Give feedback.
-
|
Sure, actually I am querying openwebui (https://openwebui.com/), which uses ollama for model handling. Those are different pods in a kubernetes stack. However, API key and base-URL refer to our openwebui instance (I myself was using 'ollama' and 'openwebui' interchangeably, which was totally wrong and I just realized it by finding the solution to this issue) - thus, using the "openai" style config here for the local setup makes sense. (Probably it could be helpful to add openwebui as a separate model provider to make it even more clear how to configure it) |
Beta Was this translation helpful? Give feedback.


Solution: to get the "/api/chat/completions" endpoint working with ollama-hosted models, one need to configure the config.toml as if it was an openai-model-provider: