Skip to content

Commit 95862f7

Browse files
[Benchmark][Doc] Update throughput benchmark and README (#15998)
Signed-off-by: StevenShi-23 <[email protected]> Signed-off-by: Roger Wang <[email protected]> Co-authored-by: Roger Wang <[email protected]>
1 parent 230b131 commit 95862f7

File tree

2 files changed

+44
-11
lines changed

2 files changed

+44
-11
lines changed

benchmarks/README.md

Lines changed: 29 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -51,6 +51,12 @@ become available.
5151
<td style="text-align: center;">✅</td>
5252
<td style="text-align: center;">✅</td>
5353
<td><code>likaixin/InstructCoder</code></td>
54+
</tr>
55+
<tr>
56+
<td><strong>HuggingFace-AIMO</strong></td>
57+
<td style="text-align: center;">✅</td>
58+
<td style="text-align: center;">✅</td>
59+
<td><code>AI-MO/aimo-validation-aime</code> , <code>AI-MO/NuminaMath-1.5</code>, <code>AI-MO/NuminaMath-CoT</code></td>
5460
</tr>
5561
<tr>
5662
<td><strong>HuggingFace-Other</strong></td>
@@ -187,6 +193,17 @@ python3 vllm/benchmarks/benchmark_serving.py \
187193
--num-prompts 10
188194
```
189195

196+
**`AI-MO/aimo-validation-aime`**
197+
198+
``` bash
199+
python3 vllm/benchmarks/benchmark_serving.py \
200+
--model Qwen/QwQ-32B \
201+
--dataset-name hf \
202+
--dataset-path AI-MO/aimo-validation-aime \
203+
--num-prompts 10 \
204+
--seed 42
205+
```
206+
190207
---
191208
## Example - Offline Throughput Benchmark
192209

@@ -278,6 +295,18 @@ python3 vllm/benchmarks/benchmark_throughput.py \
278295
--num-prompts 10
279296
```
280297

298+
**`AI-MO/aimo-validation-aime`**
299+
300+
```bash
301+
python3 benchmarks/benchmark_throughput.py \
302+
--model Qwen/QwQ-32B \
303+
--backend vllm \
304+
--dataset-name hf \
305+
--dataset-path AI-MO/aimo-validation-aime \
306+
--hf-split train \
307+
--num-prompts 10
308+
```
309+
281310
### Benchmark with LoRA Adapters
282311

283312
``` bash

benchmarks/benchmark_throughput.py

Lines changed: 15 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -11,10 +11,10 @@
1111

1212
import torch
1313
import uvloop
14-
from benchmark_dataset import (BurstGPTDataset, ConversationDataset,
15-
InstructCoderDataset, RandomDataset,
16-
SampleRequest, ShareGPTDataset, SonnetDataset,
17-
VisionArenaDataset)
14+
from benchmark_dataset import (AIMODataset, BurstGPTDataset,
15+
ConversationDataset, InstructCoderDataset,
16+
RandomDataset, SampleRequest, ShareGPTDataset,
17+
SonnetDataset, VisionArenaDataset)
1818
from benchmark_utils import convert_to_pytorch_benchmark_format, write_to_json
1919
from tqdm import tqdm
2020
from transformers import (AutoModelForCausalLM, AutoTokenizer,
@@ -332,7 +332,10 @@ def get_requests(args, tokenizer):
332332
common_kwargs['dataset_subset'] = args.hf_subset
333333
common_kwargs['dataset_split'] = args.hf_split
334334
sample_kwargs["enable_multimodal_chat"] = True
335-
335+
elif args.dataset_path in AIMODataset.SUPPORTED_DATASET_PATHS:
336+
dataset_cls = AIMODataset
337+
common_kwargs['dataset_subset'] = None
338+
common_kwargs['dataset_split'] = "train"
336339
else:
337340
raise ValueError(f"Unknown dataset name: {args.dataset_name}")
338341
# Remove None values
@@ -467,12 +470,13 @@ def validate_args(args):
467470
since --dataset-name is not 'hf'.",
468471
stacklevel=2)
469472
elif args.dataset_name == "hf":
470-
if args.dataset_path in VisionArenaDataset.SUPPORTED_DATASET_PATHS:
471-
assert args.backend == "vllm-chat", "VisionArenaDataset needs to use vllm-chat as the backend." #noqa: E501
472-
elif args.dataset_path in InstructCoderDataset.SUPPORTED_DATASET_PATHS:
473-
assert args.backend == "vllm", "InstructCoder dataset needs to use vllm as the backend." #noqa: E501
474-
elif args.dataset_path in ConversationDataset.SUPPORTED_DATASET_PATHS:
475-
assert args.backend == "vllm-chat", "ConversationDataset needs to use vllm-chat as the backend." #noqa: E501
473+
if args.dataset_path in (
474+
VisionArenaDataset.SUPPORTED_DATASET_PATHS.keys()
475+
| ConversationDataset.SUPPORTED_DATASET_PATHS):
476+
assert args.backend == "vllm-chat", f"{args.dataset_path} needs to use vllm-chat as the backend." #noqa: E501
477+
elif args.dataset_path in (InstructCoderDataset.SUPPORTED_DATASET_PATHS
478+
| AIMODataset.SUPPORTED_DATASET_PATHS):
479+
assert args.backend == "vllm", f"{args.dataset_path} needs to use vllm as the backend." #noqa: E501
476480
else:
477481
raise ValueError(
478482
f"{args.dataset_path} is not supported by hf dataset.")

0 commit comments

Comments
 (0)