@@ -1353,8 +1353,6 @@ def test_atomic_rmw(op, dtype_x_str, mode, sem, device):
13531353            pytest .xfail ("Only test atomic bfloat16/float16 ops on GPU" )
13541354    if  "uint"  in  dtype_x_str  and  mode  in  ["min_neg" , "all_neg" ]:
13551355        pytest .xfail ("uint cannot be negative" )
1356-     if  is_xpu () and  dtype_x_str  ==  'bfloat16' :
1357-         pytest .skip ("bfloat16 not yet supported for xpu" )
13581356
13591357    n_programs  =  5 
13601358
@@ -1443,8 +1441,6 @@ def kernel(X):
14431441                          for  check_return_val  in  ([True , False ] if  is_hip () else  [True ])]) 
14441442def  test_tensor_atomic_rmw (shape , axis , num_ctas , dtype_x_str , check_return_val , device ):
14451443    check_type_supported (dtype_x_str , device )
1446-     if  is_xpu () and  dtype_x_str  ==  'bfloat16' :
1447-         pytest .skip ("bfloat16 not yet supported for xpu" )
14481444    shape0 , shape1  =  shape 
14491445    # triton kernel 
14501446
@@ -1524,8 +1520,6 @@ def torch_to_triton_dtype(t):
15241520                                                         for  dtype_x_str  in  ['bfloat16' , 'float16' , 'float32' ]]) 
15251521def  test_tensor_atomic_add_non_exclusive_offset (size , num_ctas , dtype_x_str , device ):
15261522    check_type_supported (dtype_x_str , device )
1527-     if  is_xpu () and  dtype_x_str  ==  'bfloat16' :
1528-         pytest .skip ("bfloat16 not yet supported for xpu" )
15291523
15301524    @triton .jit  
15311525    def  kernel (X , val , NUM : tl .constexpr ):
@@ -1550,8 +1544,6 @@ def kernel(X, val, NUM: tl.constexpr):
15501544                                                         for  dtype_x_str  in  ['bfloat16' , 'float16' , 'float32' ]]) 
15511545def  test_tensor_atomic_add_shift_1 (size , num_ctas , dtype_x_str , device ):
15521546    check_type_supported (dtype_x_str , device )
1553-     if  is_xpu () and  dtype_x_str  ==  'bfloat16' :
1554-         pytest .skip ("bfloat16 not yet supported for xpu" )
15551547
15561548    @triton .jit  
15571549    def  kernel (X , val , NUM : tl .constexpr ):
@@ -1588,9 +1580,6 @@ def test_tensor_atomic_add_access_patterns(shape, idx_order, mask_step, num_ctas
15881580    if  is_interpreter ():
15891581        pytest .xfail ("not supported in the interpreter" )
15901582
1591-     if  is_xpu () and  dtype_x_str  ==  'bfloat16' :
1592-         pytest .skip ("bfloat16 not yet supported for xpu" )
1593- 
15941583    @triton .jit  
15951584    def  kernel (in_ptr , idx_ptr , out_ptr , shape0 , shape1 , mask_step , XBLOCK : tl .constexpr ):
15961585        xoffset  =  tl .program_id (0 ) *  XBLOCK 
0 commit comments