8282 cd benchmarks
8383 pip install .
8484
85+ - name : Run sglang benchmark int8
86+ if : ${{ steps.install-benchmarks.outcome == 'success' && !cancelled() && (inputs.benchmarks == '' || contains(fromJson(inputs.benchmarks || '[]'), 'sglang')) }}
87+ run : |
88+ source ./scripts/capture-hw-details.sh
89+
90+ ./scripts/test-triton.sh --install-sglang --skip-pip-install --skip-pytorch-install
91+ cd benchmarks/third_party/sglang
92+ python scaled_mm_benchmark.py --reports $REPORTS
93+ python ../vllm/transform_results.py $REPORTS/scaled_mm_benchmark.csv $REPORTS/scaled-mm-int8-report.csv --tag $TAG --benchmark scaled-mm-int8 --param_cols="M,N,K" --bgroup sglang
94+
95+ - name : Run sglang benchmark with fp8
96+ if : ${{ steps.install-benchmarks.outcome == 'success' && !cancelled() && (inputs.benchmarks == '' || contains(fromJson(inputs.benchmarks || '[]'), 'sglang')) }}
97+ run : |
98+ source ./scripts/capture-hw-details.sh
99+
100+ cd benchmarks/third_party/sglang
101+ FP8="1" python scaled_mm_benchmark.py --reports $REPORTS
102+ python ../vllm/transform_results.py $REPORTS/scaled_mm_benchmark.csv $REPORTS/scaled-mm-fp8-report.csv --tag $TAG --benchmark scaled-mm-fp8 --param_cols="M,N,K" --bgroup sglang
103+
85104 - name : Run vllm benchmarks bf16
86105 if : ${{ steps.install-benchmarks.outcome == 'success' && !cancelled() && (inputs.benchmarks == '' || contains(fromJson(inputs.benchmarks || '[]'), 'vllm')) }}
87106 run : |
92111
93112 cd benchmarks/third_party/vllm
94113 python batched_moe_benchmark.py --reports $REPORTS
95- python transform_results.py $REPORTS/moe-gemm-performance.csv $REPORTS/moe-gemm-report.csv --tag $TAG --benchmark moe-bf16-benchmark
114+ python transform_results.py $REPORTS/moe-gemm-performance.csv $REPORTS/moe-gemm-report.csv --tag $TAG --benchmark moe-bf16-benchmark --param_cols="num_experts,max_tokens_per_expert,K,N" --bgroup vllm
115+
96116
97117 - name : Run vllm benchmarks fp8
98118 if : ${{ steps.install-benchmarks.outcome == 'success' && !cancelled() && (inputs.benchmarks == '' || contains(fromJson(inputs.benchmarks || '[]'), 'vllm')) }}
@@ -101,7 +121,8 @@ jobs:
101121
102122 cd benchmarks/third_party/vllm
103123 FP8="1" python batched_moe_benchmark.py --reports $REPORTS
104- python transform_results.py $REPORTS/moe-gemm-performance.csv $REPORTS/moe-gemm-fp8-report.csv --tag $TAG --benchmark moe-fp8-benchmark
124+ python transform_results.py $REPORTS/moe-gemm-performance.csv $REPORTS/moe-gemm-fp8-report.csv --tag $TAG --benchmark moe-fp8-benchmark --param_cols="num_experts,max_tokens_per_expert,K,N" --bgroup vllm
125+
105126
106127 - name : Run Liger-Kernel benchmarks
107128 if : ${{ steps.install.outcome == 'success' && !cancelled() && (inputs.benchmarks == '' || contains(fromJson(inputs.benchmarks || '[]'), 'liger')) }}
0 commit comments