ollama timeout value #21941
Replies: 4 comments 8 replies
-
|
The Frigate configuration doesn't currently expose a timeout setting for Ollama requests(1)(2). The default timeout appears to be 60 seconds, which can cause issues during model pulls or inference on CPU-only systems(3). Based on the GitHub discussion, there are a few recommendations:
The "aborting completion request due to client closing the connection" message suggests Frigate is timing out before Ollama completes the inference. Without GPU acceleration, this is expected behavior with the current implementation(3). If you need to use CPU-only Ollama, you may want to consider using one of the cloud-based GenAI providers (Gemini, OpenAI, Azure OpenAI) instead, or stick with Frigate's built-in Jina CLIP model for semantic search(1)(3). 📚 Sources:
Have another question? Just tag @inkeep. |
Beta Was this translation helpful? Give feedback.
-
|
You can control this in 0.17 using provider_options |
Beta Was this translation helpful? Give feedback.
-
|
I am running llama on a large windows image. My problem may be I did not install the CUDA dev kit to take advantage of the GPUs on the graphics card. |
Beta Was this translation helpful? Give feedback.
-
|
I have switched to the new qwen3.5:35b and now the problem appears her too from time to time now. Modell is preloaded completly in vram. Problem is, if a person walks tru our yard more than one cameras detects this and the llm has a hard time to generate all the stuff for alerts and descriptions. It would be helpfull if frigates waits a little longer for the response of the llm. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
getting client timeouts using call to local ollama. Is there a timeout value in config to increase
log from ollama:
level=INFO source=runner.go:682 msg="aborting completion request due to client closing the connection"
Beta Was this translation helpful? Give feedback.
All reactions