Skip to content

Commit 5926519

Browse files
authored
docs: add note around model name consistency (#205)
Signed-off-by: bitliu <[email protected]>
1 parent e75fc0f commit 5926519

File tree

4 files changed

+62
-3
lines changed

4 files changed

+62
-3
lines changed

website/docs/getting-started/configuration.md

Lines changed: 19 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -141,7 +141,7 @@ vllm_endpoints:
141141
address: "127.0.0.1" # Your server IP - MUST be IP address format
142142
port: 8000 # Your server port
143143
models:
144-
- "llama2-7b" # Model name
144+
- "llama2-7b" # Model name - must match vLLM --served-model-name
145145
weight: 1 # Load balancing weight
146146
```
147147
@@ -176,13 +176,30 @@ address: "127.0.0.1/api" # ❌ Remove path, use IP only
176176
address: "127.0.0.1:8080" # ❌ Use separate 'port' field
177177
```
178178

179+
#### Model Name Consistency
180+
181+
The model names in the `models` array must **exactly match** the `--served-model-name` parameter used when starting your vLLM server:
182+
183+
```bash
184+
# vLLM server command:
185+
vllm serve meta-llama/Llama-2-7b-hf --served-model-name llama2-7b
186+
187+
# config.yaml must use the same name:
188+
vllm_endpoints:
189+
- models: ["llama2-7b"] # ✅ Matches --served-model-name
190+
191+
model_config:
192+
"llama2-7b": # ✅ Matches --served-model-name
193+
# ... configuration
194+
```
195+
179196
### Model Settings
180197

181198
Configure model-specific settings:
182199

183200
```yaml
184201
model_config:
185-
"llama2-7b":
202+
"llama2-7b": # Must match the model name in vllm_endpoints
186203
pii_policy:
187204
allow_by_default: true # Allow PII by default
188205
pii_types_allowed: ["EMAIL_ADDRESS", "PERSON"]

website/docs/getting-started/installation.md

Lines changed: 19 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -130,6 +130,25 @@ The `address` field **must** contain a valid IP address (IPv4 or IPv6). Domain n
130130
- `"http://127.0.0.1"` → Remove protocol prefix
131131
- `"127.0.0.1:8080"` → Use separate `port` field
132132

133+
**⚠️ Important: Model Name Consistency**
134+
135+
The model name in your configuration **must exactly match** the `--served-model-name` parameter used when starting your vLLM server:
136+
137+
```bash
138+
# When starting vLLM server:
139+
vllm serve microsoft/phi-4 --port 11434 --served-model-name your-model-name
140+
141+
# The config.yaml must use the same name:
142+
vllm_endpoints:
143+
- models: ["your-model-name"] # ✅ Must match --served-model-name
144+
145+
model_config:
146+
"your-model-name": # ✅ Must match --served-model-name
147+
# ... configuration
148+
```
149+
150+
If these names don't match, the router won't be able to route requests to your model.
151+
133152
The default configuration includes example endpoints that you should update for your setup.
134153

135154
## Running the Router

website/docs/getting-started/reasoning.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,7 @@ vllm_endpoints:
3434
- name: "endpoint1"
3535
address: "127.0.0.1"
3636
port: 8000
37-
models: ["deepseek-v31", "qwen3-30b", "openai/gpt-oss-20b"]
37+
models: ["deepseek-v31", "qwen3-30b", "openai/gpt-oss-20b"] # Must match --served-model-name
3838
weight: 1
3939

4040
# Reasoning family configurations (how to express reasoning for a family)

website/docs/training/model-performance-eval.md

Lines changed: 23 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -59,6 +59,29 @@ see code in [/src/training/model_eval](https://github.com/vllm-project/semantic-
5959
pip install -r requirements.txt
6060
```
6161

62+
**⚠️ Critical Configuration Requirement:**
63+
64+
The `--served-model-name` parameter in your vLLM command **must exactly match** the model names in your `config/config.yaml`:
65+
66+
```yaml
67+
# config/config.yaml must match the --served-model-name values above
68+
vllm_endpoints:
69+
- name: "endpoint1"
70+
address: "127.0.0.1"
71+
port: 11434
72+
models: ["phi4"] # ✅ Matches --served_model_name phi4
73+
- name: "endpoint2"
74+
address: "127.0.0.1"
75+
port: 11435
76+
models: ["qwen3-0.6B"] # ✅ Matches --served_model_name qwen3-0.6B
77+
78+
model_config:
79+
"phi4": # ✅ Matches --served_model_name phi4
80+
# ... configuration
81+
"qwen3-0.6B": # ✅ Matches --served_model_name qwen3-0.6B
82+
# ... configuration
83+
```
84+
6285
**Optional tip:**
6386

6487
- Ensure your `config/config.yaml` includes your deployed model names under `vllm_endpoints[].models` and any pricing/policy under `model_config` if you plan to use the generated config directly.

0 commit comments

Comments
 (0)