[Beta Support]: llamacpp /props fallback for llama-swap proxy #22747
daptify14
started this conversation in
Beta Support
Replies: 1 comment 5 replies
-
|
Thanks, that is odd that it does not follow the same scheme, will see what it looks like to try and support that |
Beta Was this translation helpful? Give feedback.
5 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Describe the problem you are having
After #22737, the llamacpp provider queries
GET /props?model=<name>to auto-detect context size, modalities, and tool support. This works with direct llama-server and llama.cpp router mode, but returns 404 behind llama-swap.When
/propsfails,_supports_visionand_supports_toolsdefault toFalse. Doesn't look like these flags are used yet, but could matter if they are in the future.llama-swap exposes per-model passthrough at
/upstream/<name>/props, tested it and the response structure seems identical to/props?model=<name>. I know the docs only mention llama.cpp / llama-server directly, but could it make sense to try that path as a fallback in_init_provider()before falling back to defaults?Beta Version
0.18.0-68dfb15
Issue Category
Other
Frigate config file
Relevant Frigate log output
Relevant go2rtc log output (if applicable)
No response
Install method
Proxmox via Docker
docker-compose file or Docker CLI command
Operating system
Debian
CPU / GPU / Hardware
No response
Screenshots
No response
Steps to reproduce
No response
Any other information that may be helpful
No response
Beta Was this translation helpful? Give feedback.
All reactions