Skip to content

Commit e427f29

Browse files
authored
Do not run benchmark for batch_size >= 16 on b580 (#5423)
Signed-off-by: Ettore Tiotto <[email protected]>
1 parent da48f1a commit e427f29

File tree

1 file changed

+6
-0
lines changed

1 file changed

+6
-0
lines changed

benchmarks/triton_kernels_benchmark/flex_attention_benchmark_causal_mask.py

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -74,6 +74,12 @@ def causal_mask(_, __, q_idx, kv_idx):
7474
batch_sizes = [16, 32, 64] if throughput_test else [batch_size]
7575
fa_kernel_mode = os.getenv('FA_KERNEL_MODE', 'fwd')
7676

77+
if torch.xpu.get_device_name() == '580':
78+
old_count = len(batch_sizes)
79+
batch_sizes = [size for size in batch_sizes if size < 16]
80+
if len(batch_sizes) != old_count:
81+
print('Skipping running batch_sizes >= 16 on b580')
82+
7783

7884
# Kernel profiling for Backward mode is not working as expected:
7985
# For details: https://github.com/pytorch/pytorch/issues/144778

0 commit comments

Comments
 (0)