Skip to content

Commit f3d5544

Browse files
committed
Remove top_k from vision model UI options
1 parent 35a1438 commit f3d5544

File tree

4 files changed

+7
-10
lines changed

4 files changed

+7
-10
lines changed

charts/azimuth-image-analysis/azimuth-ui.schema.yaml

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -27,5 +27,7 @@ sortOrder:
2727
- /azimuth-llm/ui/appSettings/llm_params/frequency_penalty
2828
- /azimuth-llm/ui/appSettings/llm_params/presence_penalty
2929
- /azimuth-llm/ui/appSettings/llm_params/top_p
30-
- /azimuth-llm/ui/appSettings/llm_params/top_k
30+
# vLLM responds with HTTP 400 BadRequest when top_k is
31+
# passed to a vision model (but ollama accepts it)
32+
# - /azimuth-llm/ui/appSettings/llm_params/top_k
3133
- /azimuth-llm/api/modelMaxContextLength

charts/azimuth-image-analysis/ci/ui-only-values.yaml

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -11,6 +11,5 @@ azimuth-llm:
1111
max_tokens: 101
1212
temperature: 0.1
1313
top_p: 0.15
14-
top_k: 1
1514
presence_penalty: 0.9
1615
frequency_penalty: 1

charts/azimuth-image-analysis/values.schema.json

Lines changed: 1 addition & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -58,6 +58,7 @@
5858
"description": "Model name supplied to the OpenAI client in frontend web app. Should match huggingface.model above."
5959
},
6060
"llm_params": {
61+
"$comment": "top_k parameter causes vLLM to error for most (all?) vision models so is excluded here",
6162
"type": "object",
6263
"properties": {
6364
"max_tokens": {
@@ -82,13 +83,6 @@
8283
"exclusiveMinimum": 0,
8384
"maximum": 1
8485
},
85-
"top_k": {
86-
"type": "integer",
87-
"title": "LLM Top K",
88-
"description": "The [top k](https://docs.vllm.ai/en/stable/dev/sampling_params.html) value to use when generating LLM responses (must be an integer).",
89-
"default": -1,
90-
"minimum": -1
91-
},
9286
"presence_penalty": {
9387
"type": "number",
9488
"title": "LLM Presence Penalty",

web-apps/image-analysis/defaults.yml

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,9 @@ llm_params:
1818
max_tokens:
1919
temperature:
2020
top_p:
21-
top_k:
21+
# vLLM rejects requests with top_k parameter for
22+
# most (all?) vision models so can't use it here
23+
# top_k:
2224
frequency_penalty:
2325
presence_penalty:
2426

0 commit comments

Comments
 (0)