@@ -3438,9 +3438,8 @@ define <vscale x 4 x i32> @vbrev_v(<vscale x 4 x i32> %a, iXLen %vl) {
3438
3438
define <vscale x 4 x i32 > @vbrev8_v (<vscale x 4 x i32 > %a , iXLen %vl ) {
3439
3439
; CHECK-LABEL: vbrev8_v:
3440
3440
; CHECK: # %bb.0:
3441
- ; CHECK-NEXT: vsetvli a1, zero, e32, m2, ta, ma
3442
- ; CHECK-NEXT: vbrev8.v v10, v8
3443
3441
; CHECK-NEXT: vsetvli zero, a0, e32, m2, ta, ma
3442
+ ; CHECK-NEXT: vbrev8.v v10, v8
3444
3443
; CHECK-NEXT: vadd.vv v8, v10, v8
3445
3444
; CHECK-NEXT: ret
3446
3445
%1 = call <vscale x 4 x i32 > @llvm.riscv.vbrev8.nxv4i32 (<vscale x 4 x i32 > undef , <vscale x 4 x i32 > %a , iXLen -1 )
@@ -3451,9 +3450,8 @@ define <vscale x 4 x i32> @vbrev8_v(<vscale x 4 x i32> %a, iXLen %vl) {
3451
3450
define <vscale x 4 x i32 > @vrev8_v (<vscale x 4 x i32 > %a , iXLen %vl ) {
3452
3451
; CHECK-LABEL: vrev8_v:
3453
3452
; CHECK: # %bb.0:
3454
- ; CHECK-NEXT: vsetvli a1, zero, e32, m2, ta, ma
3455
- ; CHECK-NEXT: vrev8.v v10, v8
3456
3453
; CHECK-NEXT: vsetvli zero, a0, e32, m2, ta, ma
3454
+ ; CHECK-NEXT: vrev8.v v10, v8
3457
3455
; CHECK-NEXT: vadd.vv v8, v10, v8
3458
3456
; CHECK-NEXT: ret
3459
3457
%1 = call <vscale x 4 x i32 > @llvm.riscv.vrev8.nxv4i32 (<vscale x 4 x i32 > undef , <vscale x 4 x i32 > %a , iXLen -1 )
@@ -3560,9 +3558,8 @@ define <vscale x 4 x i32> @vrol_vx(<vscale x 4 x i32> %a, iXLen %b, iXLen %vl) {
3560
3558
define <vscale x 2 x i64 > @vclmul_vv (<vscale x 2 x i64 > %a , <vscale x 2 x i64 > %b , iXLen %vl ) {
3561
3559
; CHECK-LABEL: vclmul_vv:
3562
3560
; CHECK: # %bb.0:
3563
- ; CHECK-NEXT: vsetvli a1, zero, e64, m2, ta, ma
3564
- ; CHECK-NEXT: vclmul.vv v10, v8, v10
3565
3561
; CHECK-NEXT: vsetvli zero, a0, e64, m2, ta, ma
3562
+ ; CHECK-NEXT: vclmul.vv v10, v8, v10
3566
3563
; CHECK-NEXT: vadd.vv v8, v10, v8
3567
3564
; CHECK-NEXT: ret
3568
3565
%1 = call <vscale x 2 x i64 > @llvm.riscv.vclmul.nxv2i64.nxv2i64 (<vscale x 2 x i64 > undef , <vscale x 2 x i64 > %a , <vscale x 2 x i64 > %b , iXLen -1 )
@@ -3573,9 +3570,8 @@ define <vscale x 2 x i64> @vclmul_vv(<vscale x 2 x i64> %a, <vscale x 2 x i64> %
3573
3570
define <vscale x 2 x i64 > @vclmul_vx (<vscale x 2 x i64 > %a , i32 %b , iXLen %vl ) {
3574
3571
; CHECK-LABEL: vclmul_vx:
3575
3572
; CHECK: # %bb.0:
3576
- ; CHECK-NEXT: vsetvli a2, zero, e64, m2, ta, ma
3577
- ; CHECK-NEXT: vclmul.vx v10, v8, a0
3578
3573
; CHECK-NEXT: vsetvli zero, a1, e64, m2, ta, ma
3574
+ ; CHECK-NEXT: vclmul.vx v10, v8, a0
3579
3575
; CHECK-NEXT: vadd.vv v8, v10, v8
3580
3576
; CHECK-NEXT: ret
3581
3577
%1 = call <vscale x 2 x i64 > @llvm.riscv.vclmul.nxv2i64.i32 (<vscale x 2 x i64 > undef , <vscale x 2 x i64 > %a , i32 %b , iXLen -1 )
@@ -3586,9 +3582,8 @@ define <vscale x 2 x i64> @vclmul_vx(<vscale x 2 x i64> %a, i32 %b, iXLen %vl) {
3586
3582
define <vscale x 2 x i64 > @vclmulh_vv (<vscale x 2 x i64 > %a , <vscale x 2 x i64 > %b , iXLen %vl ) {
3587
3583
; CHECK-LABEL: vclmulh_vv:
3588
3584
; CHECK: # %bb.0:
3589
- ; CHECK-NEXT: vsetvli a1, zero, e64, m2, ta, ma
3590
- ; CHECK-NEXT: vclmulh.vv v10, v8, v10
3591
3585
; CHECK-NEXT: vsetvli zero, a0, e64, m2, ta, ma
3586
+ ; CHECK-NEXT: vclmulh.vv v10, v8, v10
3592
3587
; CHECK-NEXT: vadd.vv v8, v10, v8
3593
3588
; CHECK-NEXT: ret
3594
3589
%1 = call <vscale x 2 x i64 > @llvm.riscv.vclmulh.nxv2i64.nxv2i64 (<vscale x 2 x i64 > undef , <vscale x 2 x i64 > %a , <vscale x 2 x i64 > %b , iXLen -1 )
@@ -3599,9 +3594,8 @@ define <vscale x 2 x i64> @vclmulh_vv(<vscale x 2 x i64> %a, <vscale x 2 x i64>
3599
3594
define <vscale x 2 x i64 > @vclmulh_vx (<vscale x 2 x i64 > %a , i32 %b , iXLen %vl ) {
3600
3595
; CHECK-LABEL: vclmulh_vx:
3601
3596
; CHECK: # %bb.0:
3602
- ; CHECK-NEXT: vsetvli a2, zero, e64, m2, ta, ma
3603
- ; CHECK-NEXT: vclmulh.vx v10, v8, a0
3604
3597
; CHECK-NEXT: vsetvli zero, a1, e64, m2, ta, ma
3598
+ ; CHECK-NEXT: vclmulh.vx v10, v8, a0
3605
3599
; CHECK-NEXT: vadd.vv v8, v10, v8
3606
3600
; CHECK-NEXT: ret
3607
3601
%1 = call <vscale x 2 x i64 > @llvm.riscv.vclmulh.nxv2i64.i32 (<vscale x 2 x i64 > undef , <vscale x 2 x i64 > %a , i32 %b , iXLen -1 )
0 commit comments