We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 60bc25e commit f509a20Copy full SHA for f509a20
docs/deployment/integrations/production-stack.md
@@ -55,7 +55,7 @@ sudo kubectl port-forward svc/vllm-router-service 30080:80
55
And then you can send out a query to the OpenAI-compatible API to check the available models:
56
57
```bash
58
-curl -o- http://localhost:30080/models
+curl -o- http://localhost:30080/v1/models
59
```
60
61
??? console "Output"
@@ -78,7 +78,7 @@ curl -o- http://localhost:30080/models
78
To send an actual chatting request, you can issue a curl request to the OpenAI `/completion` endpoint:
79
80
81
-curl -X POST http://localhost:30080/completions \
+curl -X POST http://localhost:30080/v1/completions \
82
-H "Content-Type: application/json" \
83
-d '{
84
"model": "facebook/opt-125m",
0 commit comments