@@ -9,7 +9,7 @@ gpu.module @test {
99 // CHECK-SAME: !xegpu.tensor_desc<8x16xf32>, !xegpu.tensor_desc<8x16xf32>,
1010 // CHECK-SAME: !xegpu.tensor_desc<8x16xf32>, !xegpu.tensor_desc<8x16xf32>,
1111 // CHECK-SAME: !xegpu.tensor_desc<8x16xf32>, !xegpu.tensor_desc<8x16xf32>
12- // CHECK-SAME: to !xegpu.tensor_desc<24x32xf32, #xegpu.layout<inst_data = [8, 16]>> {__xetile_blocking_inner_block__ = array<i64: 8, 16>, __xetile_blocking_unpack__ }
12+ // CHECK-SAME: to !xegpu.tensor_desc<24x32xf32, #xegpu.layout<inst_data = [8, 16]>> {__xegpu_blocking_tile_shape__ = array<i64: 8, 16>, __xegpu_blocking_unpack__ }
1313 gpu.func @test_create_nd_tdesc (%src: memref <24 x32 xf32 >) -> !xegpu.tensor_desc <24 x32 xf32 , #xegpu.layout <inst_data = [8 , 16 ]>> {
1414 %tdesc = xegpu.create_nd_tdesc %src [0 , 0 ] : memref <24 x32 xf32 > -> !xegpu.tensor_desc <24 x32 xf32 , #xegpu.layout <inst_data = [8 , 16 ]>>
1515 gpu.return %tdesc : !xegpu.tensor_desc <24 x32 xf32 , #xegpu.layout <inst_data = [8 , 16 ]>>
@@ -22,7 +22,7 @@ gpu.module @test {
2222 // CHECK-COUNT-2: [[tdesc:%.+]] = xegpu.create_nd_tdesc [[arg0]][{{.*}}] : memref<64xf32> -> !xegpu.tensor_desc<16xf32>
2323 // CHECK: [[cast:%.+]] = builtin.unrealized_conversion_cast
2424 // CHECK-SAME: !xegpu.tensor_desc<16xf32>, !xegpu.tensor_desc<16xf32>
25- // CHECK-SAME: to !xegpu.tensor_desc<32xf32, #xegpu.layout<inst_data = [16]>> {__xetile_blocking_inner_block__ = array<i64: 16>, __xetile_blocking_unpack__ }
25+ // CHECK-SAME: to !xegpu.tensor_desc<32xf32, #xegpu.layout<inst_data = [16]>> {__xegpu_blocking_tile_shape__ = array<i64: 16>, __xegpu_blocking_unpack__ }
2626 gpu.func @test_create_nd_tdesc_1d (%src: memref <64 xf32 >) -> !xegpu.tensor_desc <32 xf32 , #xegpu.layout <inst_data = [16 ]>> {
2727 %tdesc = xegpu.create_nd_tdesc %src [0 ] : memref <64 xf32 > -> !xegpu.tensor_desc <32 xf32 , #xegpu.layout <inst_data = [16 ]>>
2828 gpu.return %tdesc : !xegpu.tensor_desc <32 xf32 , #xegpu.layout <inst_data = [16 ]>>
0 commit comments