8282          cd benchmarks 
8383          pip install . 
8484
85+        - name : Run sglang benchmark int8 
86+         if : ${{ steps.install-benchmarks.outcome == 'success' && !cancelled() && (inputs.benchmarks == '' || contains(fromJson(inputs.benchmarks || '[]'), 'sglang')) }} 
87+         run : | 
88+           source ./scripts/capture-hw-details.sh 
89+ 
90+           ./scripts/test-triton.sh --install-sglang --skip-pip-install --skip-pytorch-install 
91+           cd benchmarks/third_party/sglang 
92+           python scaled_mm_benchmark.py --reports $REPORTS 
93+           python ../vllm/transform_results.py $REPORTS/scaled_mm_benchmark.csv $REPORTS/scaled-mm-int8-report.csv --tag $TAG --benchmark scaled-mm-int8 --param_cols="M,N,K" --bgroup sglang 
94+ 
95+        - name : Run sglang benchmark with fp8 
96+         if : ${{ steps.install-benchmarks.outcome == 'success' && !cancelled() && (inputs.benchmarks == '' || contains(fromJson(inputs.benchmarks || '[]'), 'sglang')) }} 
97+         run : | 
98+           source ./scripts/capture-hw-details.sh 
99+ 
100+           cd benchmarks/third_party/sglang 
101+           FP8="1" python scaled_mm_benchmark.py --reports $REPORTS 
102+           python ../vllm/transform_results.py $REPORTS/scaled_mm_benchmark.csv $REPORTS/scaled-mm-fp8-report.csv --tag $TAG --benchmark scaled-mm-fp8 --param_cols="M,N,K" --bgroup sglang 
103+ 
85104       - name : Run vllm benchmarks bf16 
86105        if : ${{ steps.install-benchmarks.outcome == 'success' && !cancelled() && (inputs.benchmarks == '' || contains(fromJson(inputs.benchmarks || '[]'), 'vllm')) }} 
87106        run : | 
92111
93112          cd benchmarks/third_party/vllm 
94113          python batched_moe_benchmark.py --reports $REPORTS 
95-           python transform_results.py $REPORTS/moe-gemm-performance.csv $REPORTS/moe-gemm-report.csv --tag $TAG --benchmark moe-bf16-benchmark 
114+           python transform_results.py $REPORTS/moe-gemm-performance.csv $REPORTS/moe-gemm-report.csv --tag $TAG --benchmark moe-bf16-benchmark --param_cols="num_experts,max_tokens_per_expert,K,N" --bgroup vllm 
115+ 
96116
97117       - name : Run vllm benchmarks fp8 
98118        if : ${{ steps.install-benchmarks.outcome == 'success' && !cancelled() && (inputs.benchmarks == '' || contains(fromJson(inputs.benchmarks || '[]'), 'vllm')) }} 
@@ -101,7 +121,8 @@ jobs:
101121
102122          cd benchmarks/third_party/vllm 
103123          FP8="1" python batched_moe_benchmark.py --reports $REPORTS 
104-           python transform_results.py $REPORTS/moe-gemm-performance.csv $REPORTS/moe-gemm-fp8-report.csv --tag $TAG --benchmark moe-fp8-benchmark 
124+           python transform_results.py $REPORTS/moe-gemm-performance.csv $REPORTS/moe-gemm-fp8-report.csv --tag $TAG --benchmark moe-fp8-benchmark --param_cols="num_experts,max_tokens_per_expert,K,N" --bgroup vllm 
125+ 
105126
106127       - name : Run Liger-Kernel benchmarks 
107128        if : ${{ steps.install.outcome == 'success' && !cancelled() && (inputs.benchmarks == '' || contains(fromJson(inputs.benchmarks || '[]'), 'liger')) }} 
0 commit comments