@@ -1679,12 +1679,12 @@ func.func @load_0d(%memref : memref<200x100xf32>, %i : index, %j : index) -> vec
1679
1679
1680
1680
// -----
1681
1681
1682
- func.func @load_alignment (%memref : memref <200 x100 xf32 >, %i : index , %j : index ) -> vector <8 xf32 > {
1682
+ func.func @load_with_alignment (%memref : memref <200 x100 xf32 >, %i : index , %j : index ) -> vector <8 xf32 > {
1683
1683
%0 = vector.load %memref [%i , %j ] { alignment = 8 } : memref <200 x100 xf32 >, vector <8 xf32 >
1684
1684
return %0 : vector <8 xf32 >
1685
1685
}
1686
1686
1687
- // CHECK-LABEL: func @load_alignment
1687
+ // CHECK-LABEL: func @load_with_alignment
1688
1688
// CHECK: llvm.load {{.*}} {alignment = 8 : i64} : !llvm.ptr -> vector<8xf32>
1689
1689
1690
1690
// -----
@@ -1795,12 +1795,12 @@ func.func @store_0d(%memref : memref<200x100xf32>, %i : index, %j : index) {
1795
1795
1796
1796
// -----
1797
1797
1798
- func.func @store_alignment (%memref : memref <200 x100 xf32 >, %i : index , %j : index , %val : vector <4 xf32 >) {
1798
+ func.func @store_with_alignment (%memref : memref <200 x100 xf32 >, %i : index , %j : index , %val : vector <4 xf32 >) {
1799
1799
vector.store %val , %memref [%i , %j ] {alignment = 8 } : memref <200 x100 xf32 >, vector <4 xf32 >
1800
1800
return
1801
1801
}
1802
1802
1803
- // CHECK-LABEL: func @store_alignment
1803
+ // CHECK-LABEL: func @store_with_alignment
1804
1804
// CHECK: llvm.store %{{.*}} {alignment = 8 : i64} : vector<4xf32>, !llvm.ptr
1805
1805
1806
1806
// -----
@@ -1859,12 +1859,12 @@ func.func @masked_load_index_scalable(%arg0: memref<?xindex>, %arg1: vector<[16]
1859
1859
1860
1860
// -----
1861
1861
1862
- func.func @masked_load_alignment (%arg0: memref <?xf32 >, %arg1: vector <16 xi1 >, %arg2: vector <16 xf32 >, %arg3: index ) -> vector <16 xf32 > {
1862
+ func.func @masked_load_with_alignment (%arg0: memref <?xf32 >, %arg1: vector <16 xi1 >, %arg2: vector <16 xf32 >, %arg3: index ) -> vector <16 xf32 > {
1863
1863
%0 = vector.maskedload %arg0 [%arg3 ], %arg1 , %arg2 { alignment = 2 } : memref <?xf32 >, vector <16 xi1 >, vector <16 xf32 > into vector <16 xf32 >
1864
1864
return %0 : vector <16 xf32 >
1865
1865
}
1866
1866
1867
- // CHECK-LABEL: func @masked_load_alignment
1867
+ // CHECK-LABEL: func @masked_load_with_alignment
1868
1868
// CHECK: llvm.intr.masked.load %{{.*}} {alignment = 2 : i32} : (!llvm.ptr, vector<16xi1>, vector<16xf32>) -> vector<16xf32>
1869
1869
1870
1870
// -----
@@ -1921,12 +1921,12 @@ func.func @masked_store_index_scalable(%arg0: memref<?xindex>, %arg1: vector<[16
1921
1921
1922
1922
// -----
1923
1923
1924
- func.func @masked_store_alignment (%arg0: memref <?xf32 >, %arg1: vector <16 xi1 >, %arg2: vector <16 xf32 >, %arg3: index ) {
1924
+ func.func @masked_store_with_alignment (%arg0: memref <?xf32 >, %arg1: vector <16 xi1 >, %arg2: vector <16 xf32 >, %arg3: index ) {
1925
1925
vector.maskedstore %arg0 [%arg3 ], %arg1 , %arg2 { alignment = 2 } : memref <?xf32 >, vector <16 xi1 >, vector <16 xf32 >
1926
1926
return
1927
1927
}
1928
1928
1929
- // CHECK-LABEL: func @masked_store_alignment
1929
+ // CHECK-LABEL: func @masked_store_with_alignment
1930
1930
// CHECK: llvm.intr.masked.store %{{.*}} {alignment = 2 : i32} : vector<16xf32>, vector<16xi1> into !llvm.ptr
1931
1931
1932
1932
// -----
0 commit comments