@@ -1685,11 +1685,7 @@ func.func @load_alignment(%memref : memref<200x100xf32>, %i : index, %j : index)
1685
1685
}
1686
1686
1687
1687
// CHECK-LABEL: func @load_alignment
1688
- // CHECK: %[[C100:.*]] = llvm.mlir.constant(100 : index) : i64
1689
- // CHECK: %[[MUL:.*]] = llvm.mul %{{.*}}, %[[C100]] : i64
1690
- // CHECK: %[[ADD:.*]] = llvm.add %[[MUL]], %{{.*}} : i64
1691
- // CHECK: %[[GEP:.*]] = llvm.getelementptr %{{.*}}[%[[ADD]]] : (!llvm.ptr, i64) -> !llvm.ptr, f32
1692
- // CHECK: llvm.load %[[GEP]] {alignment = 8 : i64} : !llvm.ptr -> vector<8xf32>
1688
+ // CHECK: llvm.load {{.*}} {alignment = 8 : i64} : !llvm.ptr -> vector<8xf32>
1693
1689
1694
1690
// -----
1695
1691
@@ -1806,11 +1802,7 @@ func.func @store_alignment(%memref : memref<200x100xf32>, %i : index, %j : index
1806
1802
}
1807
1803
1808
1804
// CHECK-LABEL: func @store_alignment
1809
- // CHECK: %[[C100:.*]] = llvm.mlir.constant(100 : index) : i64
1810
- // CHECK: %[[MUL:.*]] = llvm.mul %{{.*}}, %[[C100]] : i64
1811
- // CHECK: %[[ADD:.*]] = llvm.add %[[MUL]], %{{.*}} : i64
1812
- // CHECK: %[[GEP:.*]] = llvm.getelementptr %{{.*}}[%[[ADD]]] : (!llvm.ptr, i64) -> !llvm.ptr, f32
1813
- // CHECK: llvm.store %{{.*}}, %[[GEP]] {alignment = 8 : i64} : vector<4xf32>, !llvm.ptr
1805
+ // CHECK: llvm.store %{{.*}} {alignment = 8 : i64} : vector<4xf32>, !llvm.ptr
1814
1806
1815
1807
// -----
1816
1808
@@ -1870,16 +1862,12 @@ func.func @masked_load_index_scalable(%arg0: memref<?xindex>, %arg1: vector<[16]
1870
1862
1871
1863
func.func @masked_load_alignment (%arg0: memref <?xf32 >, %arg1: vector <16 xi1 >, %arg2: vector <16 xf32 >) -> vector <16 xf32 > {
1872
1864
%c0 = arith.constant 0 : index
1873
- %0 = vector.maskedload %arg0 [%c0 ], %arg1 , %arg2 { alignment = 8 } : memref <?xf32 >, vector <16 xi1 >, vector <16 xf32 > into vector <16 xf32 >
1865
+ %0 = vector.maskedload %arg0 [%c0 ], %arg1 , %arg2 { alignment = 2 } : memref <?xf32 >, vector <16 xi1 >, vector <16 xf32 > into vector <16 xf32 >
1874
1866
return %0 : vector <16 xf32 >
1875
1867
}
1876
1868
1877
- // CHECK-LABEL: func @masked_load
1878
- // CHECK: %[[CO:.*]] = arith.constant 0 : index
1879
- // CHECK: %[[C:.*]] = builtin.unrealized_conversion_cast %[[CO]] : index to i64
1880
- // CHECK: %[[P:.*]] = llvm.getelementptr %{{.*}}[%[[C]]] : (!llvm.ptr, i64) -> !llvm.ptr, f32
1881
- // CHECK: %[[L:.*]] = llvm.intr.masked.load %[[P]], %{{.*}}, %{{.*}} {alignment = 8 : i32} : (!llvm.ptr, vector<16xi1>, vector<16xf32>) -> vector<16xf32>
1882
- // CHECK: return %[[L]] : vector<16xf32>
1869
+ // CHECK-LABEL: func @masked_load_alignment
1870
+ // CHECK: llvm.intr.masked.load %{{.*}} {alignment = 2 : i32} : (!llvm.ptr, vector<16xi1>, vector<16xf32>) -> vector<16xf32>
1883
1871
1884
1872
// -----
1885
1873
@@ -1937,15 +1925,12 @@ func.func @masked_store_index_scalable(%arg0: memref<?xindex>, %arg1: vector<[16
1937
1925
1938
1926
func.func @masked_store_alignment (%arg0: memref <?xf32 >, %arg1: vector <16 xi1 >, %arg2: vector <16 xf32 >) {
1939
1927
%c0 = arith.constant 0 : index
1940
- vector.maskedstore %arg0 [%c0 ], %arg1 , %arg2 { alignment = 8 } : memref <?xf32 >, vector <16 xi1 >, vector <16 xf32 >
1928
+ vector.maskedstore %arg0 [%c0 ], %arg1 , %arg2 { alignment = 2 } : memref <?xf32 >, vector <16 xi1 >, vector <16 xf32 >
1941
1929
return
1942
1930
}
1943
1931
1944
- // CHECK-LABEL: func @masked_store
1945
- // CHECK: %[[CO:.*]] = arith.constant 0 : index
1946
- // CHECK: %[[C:.*]] = builtin.unrealized_conversion_cast %[[CO]] : index to i64
1947
- // CHECK: %[[P:.*]] = llvm.getelementptr %{{.*}}[%[[C]]] : (!llvm.ptr, i64) -> !llvm.ptr, f32
1948
- // CHECK: llvm.intr.masked.store %{{.*}}, %[[P]], %{{.*}} {alignment = 8 : i32} : vector<16xf32>, vector<16xi1> into !llvm.ptr
1932
+ // CHECK-LABEL: func @masked_store_alignment
1933
+ // CHECK: llvm.intr.masked.store %{{.*}} {alignment = 2 : i32} : vector<16xf32>, vector<16xi1> into !llvm.ptr
1949
1934
1950
1935
// -----
1951
1936
0 commit comments