|
61 | 61 | # "examples.jagged_mean",
|
62 | 62 | # "jagged_mean_tritonbench",
|
63 | 63 | # ),
|
| 64 | + # Multiple kernel mappings: (<tritonbench_module_path>, [(<helion_module>, <helion_func>), ...]) |
| 65 | + "gemm": ( |
| 66 | + "tritonbench.operators.gemm.operator", |
| 67 | + [ |
| 68 | + ("examples.matmul", "matmul"), |
| 69 | + ("examples.matmul_split_k", "matmul_split_k"), |
| 70 | + ], |
| 71 | + ), |
64 | 72 | "flash_attention": (
|
65 | 73 | "tritonbench.operators.flash_attention.operator",
|
66 | 74 | "examples.attention",
|
67 | 75 | "attention",
|
68 | 76 | ),
|
69 |
| - "fp8_attention": ( |
70 |
| - "tritonbench.operators.fp8_attention.operator", |
71 |
| - "examples.fp8_attention", |
72 |
| - "fp8_attention_tritonbench", |
73 |
| - ), |
74 | 77 | "fp8_gemm": (
|
75 | 78 | "tritonbench.operators.fp8_gemm.fp8_gemm",
|
76 | 79 | "examples.fp8_gemm",
|
77 | 80 | "fp8_gemm_tritonbench",
|
78 | 81 | ),
|
79 |
| - # Multiple kernel mappings: (<tritonbench_module_path>, [(<helion_module>, <helion_func>), ...]) |
80 |
| - "gemm": ( |
81 |
| - "tritonbench.operators.gemm.operator", |
82 |
| - [ |
83 |
| - ("examples.matmul", "matmul"), |
84 |
| - ("examples.matmul_split_k", "matmul_split_k"), |
85 |
| - ], |
| 82 | + "fp8_attention": ( |
| 83 | + "tritonbench.operators.fp8_attention.operator", |
| 84 | + "examples.fp8_attention", |
| 85 | + "fp8_attention_tritonbench", |
86 | 86 | ),
|
87 | 87 | }
|
88 | 88 |
|
|
0 commit comments