@@ -37,18 +37,18 @@ module attributes {transform.with_named_sequence} {
3737// -----
3838
3939///----------------------------------------------------------------------------------------
40- /// [Pattern: PadOpVectorizationWithTransferReadPattern
40+ /// [Pattern: PadOpVectorizationWithTransferWritePattern]
4141///----------------------------------------------------------------------------------------
4242func.func private @make_vector () -> vector <7 x9 xf32 >
4343
44- // CHECK-LABEL: func @pad_and_transfer_write_static
44+ // CHECK-LABEL: func @pad_and_transfer_write_static_low_and_high
4545// CHECK-SAME: %[[ARG0:.*]]: tensor<5x6xf32>
4646// CHECK-NOT: tensor.pad
4747// CHECK: %[[C0:.*]] = arith.constant 0 : index
4848// CHECK: %[[VEC0:.*]] = call @make_vector() : () -> vector<7x9xf32>
4949// CHECK: %[[RESULT:.*]] = vector.transfer_write %[[VEC0]], %[[ARG0]][%[[C0]], %[[C0]]] : vector<7x9xf32>, tensor<5x6xf32>
5050// CHECK: return %[[RESULT]]
51- func.func @pad_and_transfer_write_static (
51+ func.func @pad_and_transfer_write_static_low_and_high (
5252 %arg0: tensor <5 x6 xf32 >) -> tensor <5 x6 xf32 > {
5353 %c0 = arith.constant 0 : index
5454 %c5 = arith.constant 5.0 : f32
@@ -78,15 +78,15 @@ module attributes {transform.with_named_sequence} {
7878
7979func.func private @make_vector () -> vector <7 x9 xf32 >
8080
81- // CHECK-LABEL: func @pad_and_transfer_write_dynamic_static
81+ // CHECK-LABEL: func @pad_and_transfer_write_static_low_dynamic_high
8282// CHECK-SAME: %[[ARG0:.*]]: tensor<?x?xf32>, %[[SIZE:.*]]: index, %[[PADDING:.*]]: index
8383// CHECK-NOT: tensor.pad
8484// CHECK: %[[C0:.*]] = arith.constant 0 : index
8585// CHECK: %[[SUB:.*]] = tensor.extract_slice %[[ARG0]][0, 0] [%[[SIZE]], 6] [1, 1] : tensor<?x?xf32> to tensor<?x6xf32>
8686// CHECK: %[[VEC0:.*]] = call @make_vector() : () -> vector<7x9xf32>
8787// CHECK: %[[RESULT:.*]] = vector.transfer_write %[[VEC0]], %[[SUB]][%[[C0]], %[[C0]]] : vector<7x9xf32>, tensor<?x6xf32>
8888// CHECK: return %[[RESULT]]
89- func.func @pad_and_transfer_write_dynamic_static (
89+ func.func @pad_and_transfer_write_static_low_dynamic_high (
9090 %arg0: tensor <?x?xf32 >, %size: index , %padding: index ) -> tensor <?x6 xf32 > {
9191 %c0 = arith.constant 0 : index
9292 %c5 = arith.constant 5.0 : f32
@@ -166,7 +166,9 @@ module attributes {transform.with_named_sequence} {
166166
167167func.func private @make_vector () -> tensor <12 x13 xf32 >
168168
169- // Same as @pad_and_insert_slice_dest in vectorization-wit-patterns.mlir, but
169+ // Same as @pad_and_insert_slice_dest in vectorization-with-patterns.mlir, but
170+ // over here linalg::fill is not vectorized (patterns for linalg.fill are not
171+ // included here)
170172// CHECK-LABEL: func.func @pad_and_insert_slice_dest(
171173// CHECK-SAME: %[[ARG_0:.*]]: tensor<1x5x6xf32>) -> tensor<1x12x13xf32> {
172174// CHECK-NOT: tensor.pad
0 commit comments