Skip to content
Open
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion examples/04_GPTOSS120B_Example/Readme.md
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@ export MODEL_NAME=openai/gpt-oss-120b
`vLLM` can be launched via:

```
docker run --runtime nvidia --gpus all -v ${HF_HOME}:/root/.cache/huggingface --env "HUGGING_FACE_HUB_TOKEN=$HF_TOKEN" -p 8000:8000 --ipc=host vllm/vllm-openai:latest --model ${MODEL_NAME} --gpu_memory_utilization 0.95
docker run --runtime nvidia --gpus all -v ${HF_HOME}:/root/.cache/huggingface --env "HUGGING_FACE_HUB_TOKEN=$HF_TOKEN" -p 3000:3000 --ipc=host vllm/vllm-openai:latest --model ${MODEL_NAME} --gpu_memory_utilization 0.95
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

This change correctly aligns the vLLM port with the configuration file. However, the 'Debugging' section in this file is now outdated as it still refers to port 8000.

To maintain consistency, please update the mitmproxy command and the related explanation in the 'Debugging' section to use port 3000.

For example, line 70 should be:

mitmproxy -p 8001  --mode reverse:http://localhost:3000/

The surrounding text should also be updated to refer to port 3000.

Additionally, there's a typo on line 73: mitmproxy is launched on port 8001, not 8081.

Copy link
Copy Markdown
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

+1

```

`SGLang` can be launched via:
Expand Down
Loading