Skip to content

Commit 9140e7f

Browse files
authored
Merge branch 'main' into etiotto.remove_masks
2 parents 7fb33e9 + 37bc4ad commit 9140e7f

File tree

6 files changed

+637
-15
lines changed

6 files changed

+637
-15
lines changed

.github/workflows/third-party-benchmarks.yml

Lines changed: 23 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -82,6 +82,25 @@ jobs:
8282
cd benchmarks
8383
pip install .
8484
85+
- name: Run sglang benchmark int8
86+
if: ${{ steps.install-benchmarks.outcome == 'success' && !cancelled() && (inputs.benchmarks == '' || contains(fromJson(inputs.benchmarks || '[]'), 'sglang')) }}
87+
run: |
88+
source ./scripts/capture-hw-details.sh
89+
90+
./scripts/test-triton.sh --install-sglang --skip-pip-install --skip-pytorch-install
91+
cd benchmarks/third_party/sglang
92+
python scaled_mm_benchmark.py --reports $REPORTS
93+
python ../vllm/transform_results.py $REPORTS/scaled_mm_benchmark.csv $REPORTS/scaled-mm-int8-report.csv --tag $TAG --benchmark scaled-mm-int8 --param_cols="M,N,K" --bgroup sglang
94+
95+
- name: Run sglang benchmark with fp8
96+
if: ${{ steps.install-benchmarks.outcome == 'success' && !cancelled() && (inputs.benchmarks == '' || contains(fromJson(inputs.benchmarks || '[]'), 'sglang')) }}
97+
run: |
98+
source ./scripts/capture-hw-details.sh
99+
100+
cd benchmarks/third_party/sglang
101+
FP8="1" python scaled_mm_benchmark.py --reports $REPORTS
102+
python ../vllm/transform_results.py $REPORTS/scaled_mm_benchmark.csv $REPORTS/scaled-mm-fp8-report.csv --tag $TAG --benchmark scaled-mm-fp8 --param_cols="M,N,K" --bgroup sglang
103+
85104
- name: Run vllm benchmarks bf16
86105
if: ${{ steps.install-benchmarks.outcome == 'success' && !cancelled() && (inputs.benchmarks == '' || contains(fromJson(inputs.benchmarks || '[]'), 'vllm')) }}
87106
run: |
@@ -92,7 +111,8 @@ jobs:
92111
93112
cd benchmarks/third_party/vllm
94113
python batched_moe_benchmark.py --reports $REPORTS
95-
python transform_results.py $REPORTS/moe-gemm-performance.csv $REPORTS/moe-gemm-report.csv --tag $TAG --benchmark moe-bf16-benchmark
114+
python transform_results.py $REPORTS/moe-gemm-performance.csv $REPORTS/moe-gemm-report.csv --tag $TAG --benchmark moe-bf16-benchmark --param_cols="num_experts,max_tokens_per_expert,K,N" --bgroup vllm
115+
96116
97117
- name: Run vllm benchmarks fp8
98118
if: ${{ steps.install-benchmarks.outcome == 'success' && !cancelled() && (inputs.benchmarks == '' || contains(fromJson(inputs.benchmarks || '[]'), 'vllm')) }}
@@ -101,7 +121,8 @@ jobs:
101121
102122
cd benchmarks/third_party/vllm
103123
FP8="1" python batched_moe_benchmark.py --reports $REPORTS
104-
python transform_results.py $REPORTS/moe-gemm-performance.csv $REPORTS/moe-gemm-fp8-report.csv --tag $TAG --benchmark moe-fp8-benchmark
124+
python transform_results.py $REPORTS/moe-gemm-performance.csv $REPORTS/moe-gemm-fp8-report.csv --tag $TAG --benchmark moe-fp8-benchmark --param_cols="num_experts,max_tokens_per_expert,K,N" --bgroup vllm
125+
105126
106127
- name: Run Liger-Kernel benchmarks
107128
if: ${{ steps.install.outcome == 'success' && !cancelled() && (inputs.benchmarks == '' || contains(fromJson(inputs.benchmarks || '[]'), 'liger')) }}

0 commit comments

Comments
 (0)