|
| 1 | +diff --git a/test/inductor/test_flex_attention.py b/test/inductor/test_flex_attention.py |
| 2 | +index 816d3b93ec..27175cf539 100644 |
| 3 | +--- a/test/inductor/test_flex_attention.py |
| 4 | ++++ b/test/inductor/test_flex_attention.py |
| 5 | +@@ -1706,6 +1706,7 @@ class TestFlexAttention(InductorTestCase): |
| 6 | + @dtypes(*device_configs["cpu"].dtypes_fast) |
| 7 | + @dtypesIfCUDA(*device_configs["cuda"].dtypes_fast) |
| 8 | + @dtypesIfXPU(*device_configs["xpu"].dtypes_fast) |
| 9 | ++ @skip_on_xpu |
| 10 | + def test_load_from_bias_seq_only(self, device, dtype): |
| 11 | + bias = torch.randn(S, S, device=device, dtype=dtype) |
| 12 | + |
| 13 | +@@ -1719,6 +1720,7 @@ class TestFlexAttention(InductorTestCase): |
| 14 | + @dtypes(*device_configs["cpu"].dtypes_fast) |
| 15 | + @dtypesIfCUDA(*device_configs["cuda"].dtypes_fast) |
| 16 | + @dtypesIfXPU(*device_configs["xpu"].dtypes_fast) |
| 17 | ++ @skip_on_xpu |
| 18 | + def test_load_from_bias_seq_batch(self, device, dtype): |
| 19 | + bias = torch.randn(B, S, S, device=device, dtype=dtype) |
| 20 | + |
| 21 | +@@ -1779,6 +1781,7 @@ class TestFlexAttention(InductorTestCase): |
| 22 | + @dtypes(*device_configs["cpu"].dtypes_fast) |
| 23 | + @dtypesIfCUDA(*device_configs["cuda"].dtypes_fast) |
| 24 | + @dtypesIfXPU(*device_configs["xpu"].dtypes_fast) |
| 25 | ++ @skip_on_xpu |
| 26 | + def test_load_from_bias_head_seq_batch(self, device, dtype): |
| 27 | + bias = torch.randn(B, H, S, S, device=device, dtype=dtype) |
| 28 | + |
| 29 | +@@ -1896,6 +1899,7 @@ def forward(self, arg0_1, arg1_1, arg2_1, arg3_1, arg4_1): |
| 30 | + @dtypes(*device_configs["cpu"].dtypes_fast) |
| 31 | + @dtypesIfCUDA(*device_configs["cuda"].dtypes_fast) |
| 32 | + @dtypesIfXPU(*device_configs["xpu"].dtypes_fast) |
| 33 | ++ @skip_on_xpu |
| 34 | + def test_silu_on_score(self, device, dtype): |
| 35 | + def silu_score(score, b, h, q, kv): |
| 36 | + return torch.nn.functional.silu(score) |
| 37 | +@@ -2755,6 +2759,7 @@ def forward(self, arg0_1, arg1_1, arg2_1, arg3_1, arg4_1): |
| 38 | + @dtypesIfCUDA(*device_configs["cuda"].dtypes) |
| 39 | + @dtypesIfXPU(*device_configs["xpu"].dtypes) |
| 40 | + @common_utils.parametrize("head_dims", [(D, D // 2), (D // 2, D)]) |
| 41 | ++ @skip_on_xpu |
| 42 | + def test_non_equal_head_dims(self, device, dtype, score_mod, head_dims): |
| 43 | + qk_d, v_d = head_dims |
| 44 | + self.run_test(score_mod, dtype, device, B, H, S, qk_d, B, H, S, V_D=v_d) |
| 45 | +@@ -2848,6 +2853,7 @@ def forward(self, arg0_1, arg1_1, arg2_1, arg3_1, arg4_1): |
| 46 | + @dtypes(*device_configs["cpu"].dtypes_fast) |
| 47 | + @dtypesIfCUDA(*device_configs["cuda"].dtypes_fast) |
| 48 | + @dtypesIfXPU(*device_configs["xpu"].dtypes_fast) |
| 49 | ++ @skip_on_xpu |
| 50 | + def test_non_pow_2_headdim(self, device, dtype, head_dim): |
| 51 | + self.run_test(_rel_bias, dtype, device, B, H, S, head_dim, B, H, S, head_dim) |
| 52 | + |
0 commit comments