|
| 1 | +// RUN: mlir-translate -mlir-to-llvmir %s | FileCheck %s |
| 2 | + |
| 3 | +llvm.func @_QPfoo(%arg0: !llvm.ptr {fir.bindc_name = "array", llvm.nocapture}, %arg1: !llvm.ptr {fir.bindc_name = "t", llvm.nocapture}) { |
| 4 | + %0 = llvm.mlir.constant(0 : i64) : i32 |
| 5 | + %1 = llvm.mlir.constant(1 : i32) : i32 |
| 6 | + %2 = llvm.mlir.constant(10 : i64) : i64 |
| 7 | + %3 = llvm.mlir.constant(1 : i64) : i64 |
| 8 | + %4 = llvm.alloca %3 x i32 {bindc_name = "i", pinned} : (i64) -> !llvm.ptr |
| 9 | + %5 = llvm.load %arg1 : !llvm.ptr -> i32 |
| 10 | + %6 = llvm.icmp "ne" %5, %0 : i32 |
| 11 | + %7 = llvm.trunc %2 : i64 to i32 |
| 12 | + omp.wsloop { |
| 13 | + omp.simd if(%6) { |
| 14 | + omp.loop_nest (%arg2) : i32 = (%1) to (%7) inclusive step (%1) { |
| 15 | + llvm.store %arg2, %4 : i32, !llvm.ptr |
| 16 | + %8 = llvm.load %4 : !llvm.ptr -> i32 |
| 17 | + %9 = llvm.sext %8 : i32 to i64 |
| 18 | + %10 = llvm.getelementptr %arg0[%9] : (!llvm.ptr, i64) -> !llvm.ptr, i32 |
| 19 | + llvm.store %8, %10 : i32, !llvm.ptr |
| 20 | + omp.yield |
| 21 | + } |
| 22 | + } {omp.composite} |
| 23 | + } {omp.composite} |
| 24 | + llvm.return |
| 25 | +} |
| 26 | + |
| 27 | +// CHECK-LABEL: @_QPfoo |
| 28 | +// ... |
| 29 | +// CHECK: omp_loop.preheader: ; preds = |
| 30 | +// CHECK: store i32 0, ptr %[[LB_ADDR:.*]], align 4 |
| 31 | +// CHECK: store i32 9, ptr %[[UB_ADDR:.*]], align 4 |
| 32 | +// CHECK: store i32 1, ptr %[[STEP_ADDR:.*]], align 4 |
| 33 | +// CHECK: %[[VAL_15:.*]] = call i32 @__kmpc_global_thread_num(ptr @1) |
| 34 | +// CHECK: call void @__kmpc_for_static_init_4u(ptr @1, i32 %[[VAL_15]], i32 34, ptr %{{.*}}, ptr %[[LB_ADDR]], ptr %[[UB_ADDR]], ptr %[[STEP_ADDR]], i32 1, i32 0) |
| 35 | +// CHECK: %[[LB:.*]] = load i32, ptr %[[LB_ADDR]], align 4 |
| 36 | +// CHECK: %[[UB:.*]] = load i32, ptr %[[UB_ADDR]], align 4 |
| 37 | +// CHECK: %[[VAL_18:.*]] = sub i32 %[[UB]], %[[LB]] |
| 38 | +// CHECK: %[[COUNT:.*]] = add i32 %[[VAL_18]], 1 |
| 39 | +// CHECK: br label %[[OMP_LOOP_HEADER:.*]] |
| 40 | +// CHECK: omp_loop.header: ; preds = %[[OMP_LOOP_INC:.*]], %[[OMP_LOOP_PREHEADER:.*]] |
| 41 | +// CHECK: %[[IV:.*]] = phi i32 [ 0, %[[OMP_LOOP_PREHEADER]] ], [ %[[NEW_IV:.*]], %[[OMP_LOOP_INC]] ] |
| 42 | +// CHECK: br label %[[OMP_LOOP_COND:.*]] |
| 43 | +// CHECK: omp_loop.cond: ; preds = %[[OMP_LOOP_HEADER]] |
| 44 | +// CHECK: %[[VAL_25:.*]] = icmp ult i32 %[[IV]], %[[COUNT]] |
| 45 | +// CHECK: br i1 %[[VAL_25]], label %[[OMP_LOOP_BODY:.*]], label %[[OMP_LOOP_EXIT:.*]] |
| 46 | +// CHECK: omp_loop.body: ; preds = %[[OMP_LOOP_COND]] |
| 47 | +// CHECK: %[[VAL_28:.*]] = add i32 %[[IV]], %[[LB]] |
| 48 | +// CHECK: %[[VAL_29:.*]] = mul i32 %[[VAL_28]], 1 |
| 49 | +// CHECK: %[[VAL_30:.*]] = add i32 %[[VAL_29]], 1 |
| 50 | +// This is the IF clause: |
| 51 | +// CHECK: br i1 %{{.*}}, label %[[SIMD_IF_THEN:.*]], label %[[SIMD_IF_ELSE:.*]] |
| 52 | + |
| 53 | +// CHECK: simd.if.then: ; preds = %[[OMP_LOOP_BODY]] |
| 54 | +// CHECK: br label %[[VAL_33:.*]] |
| 55 | +// CHECK: omp.loop_nest.region: ; preds = %[[SIMD_IF_THEN]] |
| 56 | +// This version contains !llvm.access.group metadata for SIMD |
| 57 | +// CHECK: store i32 %[[VAL_30]], ptr %{{.*}}, align 4, !llvm.access.group !1 |
| 58 | +// CHECK: %[[VAL_34:.*]] = load i32, ptr %{{.*}}, align 4, !llvm.access.group !1 |
| 59 | +// CHECK: %[[VAL_35:.*]] = sext i32 %[[VAL_34]] to i64 |
| 60 | +// CHECK: %[[VAL_36:.*]] = getelementptr i32, ptr %[[VAL_37:.*]], i64 %[[VAL_35]] |
| 61 | +// CHECK: store i32 %[[VAL_34]], ptr %[[VAL_36]], align 4, !llvm.access.group !1 |
| 62 | +// CHECK: br label %[[OMP_REGION_CONT3:.*]] |
| 63 | +// CHECK: omp.region.cont3: ; preds = %[[VAL_33]] |
| 64 | +// CHECK: br label %[[SIMD_PRE_LATCH:.*]] |
| 65 | + |
| 66 | +// CHECK: simd.pre_latch: ; preds = %[[OMP_REGION_CONT3]], %[[OMP_REGION_CONT35:.*]] |
| 67 | +// CHECK: br label %[[OMP_LOOP_INC]] |
| 68 | +// CHECK: omp_loop.inc: ; preds = %[[SIMD_PRE_LATCH]] |
| 69 | +// CHECK: %[[NEW_IV]] = add nuw i32 %[[IV]], 1 |
| 70 | +// CHECK: br label %[[OMP_LOOP_HEADER]], !llvm.loop !2 |
| 71 | + |
| 72 | +// CHECK: simd.if.else: ; preds = %[[OMP_LOOP_BODY]] |
| 73 | +// CHECK: br label %[[VAL_41:.*]] |
| 74 | +// CHECK: omp.loop_nest.region4: ; preds = %[[SIMD_IF_ELSE]] |
| 75 | +// No llvm.access.group metadata for else clause |
| 76 | +// CHECK: store i32 %[[VAL_30]], ptr %{{.*}}, align 4 |
| 77 | +// CHECK: %[[VAL_42:.*]] = load i32, ptr %{{.*}}, align 4 |
| 78 | +// CHECK: %[[VAL_43:.*]] = sext i32 %[[VAL_42]] to i64 |
| 79 | +// CHECK: %[[VAL_44:.*]] = getelementptr i32, ptr %[[VAL_37]], i64 %[[VAL_43]] |
| 80 | +// CHECK: store i32 %[[VAL_42]], ptr %[[VAL_44]], align 4 |
| 81 | +// CHECK: br label %[[OMP_REGION_CONT35]] |
| 82 | +// CHECK: omp.region.cont35: ; preds = %[[VAL_41]] |
| 83 | +// CHECK: br label %[[SIMD_PRE_LATCH]] |
| 84 | + |
| 85 | +// CHECK: omp_loop.exit: ; preds = %[[OMP_LOOP_COND]] |
| 86 | +// CHECK: call void @__kmpc_for_static_fini(ptr @1, i32 %[[VAL_15]]) |
| 87 | +// CHECK: %[[VAL_45:.*]] = call i32 @__kmpc_global_thread_num(ptr @1) |
| 88 | +// CHECK: call void @__kmpc_barrier(ptr @2, i32 %[[VAL_45]]) |
| 89 | + |
| 90 | +// CHECK: !1 = distinct !{} |
| 91 | +// CHECK: !2 = distinct !{!2, !3} |
| 92 | +// CHECK: !3 = !{!"llvm.loop.parallel_accesses", !1} |
| 93 | +// CHECK-NOT: llvm.loop.vectorize |
0 commit comments