Skip to content

Commit 067c34a

Browse files
author
Roger Wang
authored
docs: remove deprecated disable-log-requests flag (#22113)
Signed-off-by: Roger Wang <[email protected]>
1 parent c64861d commit 067c34a

File tree

19 files changed

+14
-41
lines changed

19 files changed

+14
-41
lines changed

.buildkite/scripts/tpu/run_bm.sh

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -44,7 +44,6 @@ echo
4444

4545
VLLM_USE_V1=1 vllm serve $MODEL \
4646
--seed 42 \
47-
--disable-log-requests \
4847
--max-num-seqs $MAX_NUM_SEQS \
4948
--max-num-batched-tokens $MAX_NUM_BATCHED_TOKENS \
5049
--tensor-parallel-size $TENSOR_PARALLEL_SIZE \

benchmarks/README.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -91,7 +91,7 @@ become available.
9191
First start serving your model
9292

9393
```bash
94-
vllm serve NousResearch/Hermes-3-Llama-3.1-8B --disable-log-requests
94+
vllm serve NousResearch/Hermes-3-Llama-3.1-8B
9595
```
9696

9797
Then run the benchmarking script
@@ -146,7 +146,7 @@ If the dataset you want to benchmark is not supported yet in vLLM, even then you
146146

147147
```bash
148148
# start server
149-
VLLM_USE_V1=1 vllm serve meta-llama/Llama-3.1-8B-Instruct --disable-log-requests
149+
VLLM_USE_V1=1 vllm serve meta-llama/Llama-3.1-8B-Instruct
150150
```
151151

152152
```bash
@@ -171,7 +171,7 @@ You can skip applying chat template if your data already has it by using `--cust
171171

172172
```bash
173173
# need a model with vision capability here
174-
vllm serve Qwen/Qwen2-VL-7B-Instruct --disable-log-requests
174+
vllm serve Qwen/Qwen2-VL-7B-Instruct
175175
```
176176

177177
```bash
@@ -205,7 +205,7 @@ vllm bench serve \
205205
### Other HuggingFaceDataset Examples
206206

207207
```bash
208-
vllm serve Qwen/Qwen2-VL-7B-Instruct --disable-log-requests
208+
vllm serve Qwen/Qwen2-VL-7B-Instruct
209209
```
210210

211211
`lmms-lab/LLaVA-OneVision-Data`:
@@ -430,7 +430,7 @@ Benchmark the performance of structured output generation (JSON, grammar, regex)
430430
### Server Setup
431431

432432
```bash
433-
vllm serve NousResearch/Hermes-3-Llama-3.1-8B --disable-log-requests
433+
vllm serve NousResearch/Hermes-3-Llama-3.1-8B
434434
```
435435

436436
### JSON Schema Benchmark

benchmarks/auto_tune/auto_tune.sh

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -60,7 +60,6 @@ start_server() {
6060
pkill -f vllm
6161

6262
VLLM_USE_V1=1 VLLM_SERVER_DEV_MODE=1 VLLM_TORCH_PROFILER_DIR=$profile_dir vllm serve $MODEL \
63-
--disable-log-requests \
6463
--port 8004 \
6564
--gpu-memory-utilization $gpu_memory_utilization \
6665
--max-num-seqs $max_num_seqs \

benchmarks/benchmark_serving.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -5,8 +5,7 @@
55
On the server side, run one of the following commands:
66
vLLM OpenAI API server
77
vllm serve <your_model> \
8-
--swap-space 16 \
9-
--disable-log-requests
8+
--swap-space 16
109
1110
On the client side, run:
1211
python benchmarks/benchmark_serving.py \

benchmarks/benchmark_serving_structured_output.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@
44
55
On the server side, run one of the following commands:
66
(vLLM OpenAI API server)
7-
vllm serve <your_model> --disable-log-requests
7+
vllm serve <your_model>
88
99
On the client side, run:
1010
python benchmarks/benchmark_serving_structured_output.py \

docs/design/p2p_nccl_connector.md

Lines changed: 0 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -109,7 +109,6 @@ python3 disagg_proxy_p2p_nccl_xpyd.py &
109109
--max-num-seqs 256 \
110110
--trust-remote-code \
111111
--gpu-memory-utilization 0.9 \
112-
--disable-log-request \
113112
--kv-transfer-config \
114113
'{"kv_connector":"P2pNcclConnector","kv_role":"kv_producer","kv_buffer_size":"1e1","kv_port":"21001","kv_connector_extra_config":{"proxy_ip":"10.0.1.1","proxy_port":"30001","http_port":"20001"}}' > /var/vllm.log 2>&1 &
115114
```
@@ -131,7 +130,6 @@ python3 disagg_proxy_p2p_nccl_xpyd.py &
131130
--max-num-seqs 256 \
132131
--trust-remote-code \
133132
--gpu-memory-utilization 0.7 \
134-
--disable-log-request \
135133
--kv-transfer-config \
136134
'{"kv_connector":"P2pNcclConnector","kv_role":"kv_consumer","kv_buffer_size":"8e9","kv_port":"22001","kv_connector_extra_config":{"proxy_ip":"10.0.1.1","proxy_port":"30001","http_port":"20002"}}' > /var/vllm.log 2>&1 &
137135
```
@@ -153,7 +151,6 @@ python3 disagg_proxy_p2p_nccl_xpyd.py &
153151
--max-num-seqs 256 \
154152
--trust-remote-code \
155153
--gpu-memory-utilization 0.7 \
156-
--disable-log-request \
157154
--kv-transfer-config \
158155
'{"kv_connector":"P2pNcclConnector","kv_role":"kv_consumer","kv_buffer_size":"8e9","kv_port":"23001","kv_connector_extra_config":{"proxy_ip":"10.0.1.1","proxy_port":"30001","http_port":"20003"}}' > /var/vllm.log 2>&1 &
159156
```
@@ -175,7 +172,6 @@ python3 disagg_proxy_p2p_nccl_xpyd.py &
175172
--max-num-seqs 256 \
176173
--trust-remote-code \
177174
--gpu-memory-utilization 0.7 \
178-
--disable-log-request \
179175
--kv-transfer-config \
180176
'{"kv_connector":"P2pNcclConnector","kv_role":"kv_consumer","kv_buffer_size":"8e9","kv_port":"24001","kv_connector_extra_config":{"proxy_ip":"10.0.1.1","proxy_port":"30001","http_port":"20004"}}' > /var/vllm.log 2>&1 &
181177
```
@@ -206,7 +202,6 @@ python3 disagg_proxy_p2p_nccl_xpyd.py &
206202
--max-num-seqs 256 \
207203
--trust-remote-code \
208204
--gpu-memory-utilization 0.9 \
209-
--disable-log-request \
210205
--kv-transfer-config \
211206
'{"kv_connector":"P2pNcclConnector","kv_role":"kv_producer","kv_buffer_size":"1e1","kv_port":"21001","kv_connector_extra_config":{"proxy_ip":"10.0.1.1","proxy_port":"30001","http_port":"20001"}}' > /var/vllm.log 2>&1 &
212207
```
@@ -228,7 +223,6 @@ python3 disagg_proxy_p2p_nccl_xpyd.py &
228223
--max-num-seqs 256 \
229224
--trust-remote-code \
230225
--gpu-memory-utilization 0.9 \
231-
--disable-log-request \
232226
--kv-transfer-config \
233227
'{"kv_connector":"P2pNcclConnector","kv_role":"kv_producer","kv_buffer_size":"1e1","kv_port":"22001","kv_connector_extra_config":{"proxy_ip":"10.0.1.1","proxy_port":"30001","http_port":"20002"}}' > /var/vllm.log 2>&1 &
234228
```
@@ -250,7 +244,6 @@ python3 disagg_proxy_p2p_nccl_xpyd.py &
250244
--max-num-seqs 256 \
251245
--trust-remote-code \
252246
--gpu-memory-utilization 0.9 \
253-
--disable-log-request \
254247
--kv-transfer-config \
255248
'{"kv_connector":"P2pNcclConnector","kv_role":"kv_producer","kv_buffer_size":"1e1","kv_port":"23001","kv_connector_extra_config":{"proxy_ip":"10.0.1.1","proxy_port":"30001","http_port":"20003"}}' > /var/vllm.log 2>&1 &
256249
```
@@ -272,7 +265,6 @@ python3 disagg_proxy_p2p_nccl_xpyd.py &
272265
--max-num-seqs 256 \
273266
--trust-remote-code \
274267
--gpu-memory-utilization 0.7 \
275-
--disable-log-request \
276268
--kv-transfer-config \
277269
'{"kv_connector":"P2pNcclConnector","kv_role":"kv_consumer","kv_buffer_size":"8e9","kv_port":"24001","kv_connector_extra_config":{"proxy_ip":"10.0.1.1","proxy_port":"30001","http_port":"20004"}}' > /var/vllm.log 2>&1 &
278270
```

docs/models/supported_models.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -255,7 +255,7 @@ export https_proxy=http://your.proxy.server:port
255255
https_proxy=http://your.proxy.server:port huggingface-cli download <model_name>
256256

257257
# or use vllm cmd directly
258-
https_proxy=http://your.proxy.server:port vllm serve <model_name> --disable-log-requests
258+
https_proxy=http://your.proxy.server:port vllm serve <model_name>
259259
```
260260

261261
- Set the proxy in Python interpreter:

examples/online_serving/disaggregated_serving_p2p_nccl_xpyd/disagg_example_p2p_nccl_xpyd.sh

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -178,7 +178,6 @@ main() {
178178
--max-num-seqs 256 \
179179
--trust-remote-code \
180180
--gpu-memory-utilization 0.9 \
181-
--disable-log-request \
182181
--kv-transfer-config \
183182
"{\"kv_connector\":\"P2pNcclConnector\",\"kv_role\":\"kv_producer\",\"kv_buffer_size\":\"1e1\",\"kv_port\":\"$kv_port\",\"kv_connector_extra_config\":{\"proxy_ip\":\"0.0.0.0\",\"proxy_port\":\"$PROXY_PORT\",\"http_port\":\"$port\",\"send_type\":\"PUT_ASYNC\",\"nccl_num_channels\":\"16\"}}" > prefill$((i+1)).log 2>&1 &
184183
PIDS+=($!)
@@ -207,7 +206,6 @@ main() {
207206
--max-num-seqs 256 \
208207
--trust-remote-code \
209208
--gpu-memory-utilization 0.7 \
210-
--disable-log-request \
211209
--kv-transfer-config \
212210
"{\"kv_connector\":\"P2pNcclConnector\",\"kv_role\":\"kv_consumer\",\"kv_buffer_size\":\"8e9\",\"kv_port\":\"$kv_port\",\"kv_connector_extra_config\":{\"proxy_ip\":\"0.0.0.0\",\"proxy_port\":\"$PROXY_PORT\",\"http_port\":\"$port\",\"send_type\":\"PUT_ASYNC\",\"nccl_num_channels\":\"16\"}}" > decode$((i+1)).log 2>&1 &
213211
PIDS+=($!)

examples/online_serving/prometheus_grafana/README.md

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -13,8 +13,7 @@ Prometheus metric logging is enabled by default in the OpenAI-compatible server.
1313

1414
```bash
1515
vllm serve mistralai/Mistral-7B-v0.1 \
16-
--max-model-len 2048 \
17-
--disable-log-requests
16+
--max-model-len 2048
1817
```
1918

2019
Launch Prometheus and Grafana servers with `docker compose`:

examples/others/lmcache/disagg_prefill_lmcache_v1/disagg_vllm_launcher.sh

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,6 @@ if [[ $1 == "prefiller" ]]; then
2828
CUDA_VISIBLE_DEVICES=0 \
2929
vllm serve $MODEL \
3030
--port 8100 \
31-
--disable-log-requests \
3231
--enforce-eager \
3332
--kv-transfer-config \
3433
'{"kv_connector":"LMCacheConnectorV1","kv_role":"kv_producer","kv_connector_extra_config": {"discard_partial_chunks": false, "lmcache_rpc_port": "producer1"}}'
@@ -46,7 +45,6 @@ elif [[ $1 == "decoder" ]]; then
4645
CUDA_VISIBLE_DEVICES=1 \
4746
vllm serve $MODEL \
4847
--port 8200 \
49-
--disable-log-requests \
5048
--enforce-eager \
5149
--kv-transfer-config \
5250
'{"kv_connector":"LMCacheConnectorV1","kv_role":"kv_consumer","kv_connector_extra_config": {"discard_partial_chunks": false, "lmcache_rpc_port": "consumer1"}}'

0 commit comments

Comments
 (0)