diff --git a/results/MultiSource/Applications/ClamAV/CMakeFiles/clamscan.dir/libclamav_autoit.s b/results/MultiSource/Applications/ClamAV/CMakeFiles/clamscan.dir/libclamav_autoit.s index 3aa2843..20c8945 100644 --- a/results/MultiSource/Applications/ClamAV/CMakeFiles/clamscan.dir/libclamav_autoit.s +++ b/results/MultiSource/Applications/ClamAV/CMakeFiles/clamscan.dir/libclamav_autoit.s @@ -4978,53 +4978,53 @@ LAME_decrypt: # @LAME_decrypt st.d $fp, $sp, 104 # 8-byte Folded Spill st.d $s0, $sp, 96 # 8-byte Folded Spill lu12i.w $a3, -342684 - ori $a6, $a3, 2821 - mul.d $a2, $a2, $a6 + ori $a4, $a3, 2821 + mul.d $a2, $a2, $a4 addi.d $a2, $a2, 1 - mul.d $a3, $a2, $a6 + mul.d $a3, $a2, $a4 addi.d $a3, $a3, 1 st.w $a3, $sp, 32 - mul.d $a3, $a3, $a6 - rotri.w $a7, $a2, 23 + mul.d $a3, $a3, $a4 + rotri.w $a5, $a2, 23 addi.d $a2, $a3, 1 st.w $a2, $sp, 36 - mul.d $a3, $a2, $a6 + mul.d $a3, $a2, $a4 addi.d $a3, $a3, 1 st.w $a3, $sp, 40 - mul.d $a4, $a3, $a6 - addi.d $a4, $a4, 1 - st.w $a4, $sp, 44 - mul.d $a5, $a4, $a6 - addi.d $a5, $a5, 1 - st.w $a5, $sp, 48 - mul.d $t0, $a5, $a6 + mul.d $a6, $a3, $a4 + addi.d $a6, $a6, 1 + st.w $a6, $sp, 44 + mul.d $a7, $a6, $a4 + addi.d $a7, $a7, 1 + st.w $a7, $sp, 48 + mul.d $t0, $a7, $a4 addi.d $t0, $t0, 1 st.w $t0, $sp, 52 - mul.d $t1, $t0, $a6 + mul.d $t1, $t0, $a4 addi.d $t1, $t1, 1 st.w $t1, $sp, 56 - mul.d $t2, $t1, $a6 + mul.d $t2, $t1, $a4 addi.d $t2, $t2, 1 st.w $t2, $sp, 60 - mul.d $t3, $t2, $a6 + mul.d $t3, $t2, $a4 addi.d $t3, $t3, 1 - mul.d $t4, $t3, $a6 + mul.d $t4, $t3, $a4 addi.d $t4, $t4, 1 - mul.d $t5, $t4, $a6 + mul.d $t5, $t4, $a4 addi.d $t5, $t5, 1 - mul.d $t6, $t5, $a6 + mul.d $t6, $t5, $a4 addi.d $t6, $t6, 1 - mul.d $t7, $t6, $a6 + mul.d $t7, $t6, $a4 addi.d $t7, $t7, 1 - mul.d $t8, $t7, $a6 + mul.d $t8, $t7, $a4 addi.d $t8, $t8, 1 - mul.d $fp, $t8, $a6 + mul.d $fp, $t8, $a4 addi.d $fp, $fp, 1 - mul.d $a6, $fp, $a6 - addi.d $a6, $a6, 1 + mul.d $a4, $fp, $a4 + addi.d $a4, $a4, 1 rotri.w $s0, $t4, 19 - add.d $a7, $s0, $a7 - st.w $a7, $sp, 28 + add.d $a5, $s0, $a5 + st.w $a5, $sp, 28 vinsgr2vr.w $vr0, $t3, 0 vinsgr2vr.w $vr0, $t4, 1 vinsgr2vr.w $vr0, $t5, 2 @@ -5032,23 +5032,19 @@ LAME_decrypt: # @LAME_decrypt vinsgr2vr.w $vr1, $t7, 0 vinsgr2vr.w $vr1, $t8, 1 vinsgr2vr.w $vr1, $fp, 2 - vinsgr2vr.w $vr1, $a6, 3 + vinsgr2vr.w $vr1, $a4, 3 xvpermi.q $xr0, $xr1, 2 - xvsrli.w $xr1, $xr0, 23 - xvslli.w $xr0, $xr0, 9 - xvor.v $xr0, $xr0, $xr1 + xvrotri.w $xr0, $xr0, 23 vinsgr2vr.w $vr1, $a2, 0 vinsgr2vr.w $vr1, $a3, 1 - vinsgr2vr.w $vr1, $a4, 2 - vinsgr2vr.w $vr1, $a5, 3 + vinsgr2vr.w $vr1, $a6, 2 + vinsgr2vr.w $vr1, $a7, 3 vinsgr2vr.w $vr2, $t0, 0 vinsgr2vr.w $vr2, $t1, 1 vinsgr2vr.w $vr2, $t2, 2 vinsgr2vr.w $vr2, $t3, 3 xvpermi.q $xr1, $xr2, 2 - xvsrli.w $xr2, $xr1, 19 - xvslli.w $xr1, $xr1, 13 - xvor.v $xr1, $xr1, $xr2 + xvrotri.w $xr1, $xr1, 19 xvadd.w $xr0, $xr0, $xr1 xvst $xr0, $sp, 64 beqz $a1, .LBB2_3 diff --git a/results/MultiSource/Applications/ClamAV/CMakeFiles/clamscan.dir/shared_sha256.s b/results/MultiSource/Applications/ClamAV/CMakeFiles/clamscan.dir/shared_sha256.s index 2d92430..1158fef 100644 --- a/results/MultiSource/Applications/ClamAV/CMakeFiles/clamscan.dir/shared_sha256.s +++ b/results/MultiSource/Applications/ClamAV/CMakeFiles/clamscan.dir/shared_sha256.s @@ -304,23 +304,13 @@ sha256_final: # @sha256_final .section .rodata.cst16,"aM",@progbits,16 .p2align 4, 0x0 # -- Begin function sha256_transform .LCPI4_0: - .word 17 # 0x11 - .word 7 # 0x7 + .word 4294967281 # 0xfffffff1 + .word 4294967271 # 0xffffffe7 .word 0 # 0x0 .word 0 # 0x0 .LCPI4_1: - .word 15 # 0xf - .word 25 # 0x19 - .word 0 # 0x0 - .word 0 # 0x0 -.LCPI4_2: - .word 19 # 0x13 - .word 18 # 0x12 - .word 0 # 0x0 - .word 0 # 0x0 -.LCPI4_3: - .word 13 # 0xd - .word 14 # 0xe + .word 4294967283 # 0xfffffff3 + .word 4294967282 # 0xfffffff2 .word 0 # 0x0 .word 0 # 0x0 .text @@ -328,18 +318,18 @@ sha256_final: # @sha256_final .type sha256_transform,@function sha256_transform: # @sha256_transform # %bb.0: - addi.d $sp, $sp, -176 - st.d $ra, $sp, 168 # 8-byte Folded Spill - st.d $fp, $sp, 160 # 8-byte Folded Spill - st.d $s0, $sp, 152 # 8-byte Folded Spill - st.d $s1, $sp, 144 # 8-byte Folded Spill - st.d $s2, $sp, 136 # 8-byte Folded Spill - st.d $s3, $sp, 128 # 8-byte Folded Spill - st.d $s4, $sp, 120 # 8-byte Folded Spill - st.d $s5, $sp, 112 # 8-byte Folded Spill - st.d $s6, $sp, 104 # 8-byte Folded Spill - st.d $s7, $sp, 96 # 8-byte Folded Spill - st.d $s8, $sp, 88 # 8-byte Folded Spill + addi.d $sp, $sp, -192 + st.d $ra, $sp, 184 # 8-byte Folded Spill + st.d $fp, $sp, 176 # 8-byte Folded Spill + st.d $s0, $sp, 168 # 8-byte Folded Spill + st.d $s1, $sp, 160 # 8-byte Folded Spill + st.d $s2, $sp, 152 # 8-byte Folded Spill + st.d $s3, $sp, 144 # 8-byte Folded Spill + st.d $s4, $sp, 136 # 8-byte Folded Spill + st.d $s5, $sp, 128 # 8-byte Folded Spill + st.d $s6, $sp, 120 # 8-byte Folded Spill + st.d $s7, $sp, 112 # 8-byte Folded Spill + st.d $s8, $sp, 104 # 8-byte Folded Spill ld.w $s3, $a0, 0 ld.w $s4, $a0, 4 ld.w $s6, $a0, 8 @@ -347,7 +337,7 @@ sha256_transform: # @sha256_transform ld.w $s5, $a0, 16 ld.w $s8, $a0, 20 ld.w $ra, $a0, 24 - st.d $a0, $sp, 16 # 8-byte Folded Spill + st.d $a0, $sp, 8 # 8-byte Folded Spill ld.w $s2, $a0, 28 ori $a3, $zero, 1 pcalau12i $a0, %pc_hi20(K) @@ -570,35 +560,38 @@ sha256_transform: # @sha256_transform move $a3, $zero bnez $a4, .LBB4_1 # %bb.2: # %.lr.ph - ld.w $a4, $a1, 56 + ld.w $a2, $a1, 56 ld.w $a3, $a1, 36 - ld.w $t1, $a1, 4 + ld.w $a4, $a1, 4 + st.d $a4, $sp, 88 # 8-byte Folded Spill ld.w $a5, $a1, 0 - ld.w $t0, $a1, 60 + ld.w $s0, $a1, 60 ld.w $a6, $a1, 40 - ld.w $a2, $a1, 8 - st.d $a2, $sp, 56 # 8-byte Folded Spill - move $a2, $a4 + ld.w $t7, $a1, 8 ld.w $a7, $a1, 44 ld.w $t4, $a1, 12 ld.w $t2, $a1, 48 ld.w $a4, $a1, 16 - st.d $a4, $sp, 48 # 8-byte Folded Spill + st.d $a4, $sp, 80 # 8-byte Folded Spill ld.w $t3, $a1, 52 ld.w $a4, $a1, 20 - st.d $a4, $sp, 40 # 8-byte Folded Spill - ld.w $s0, $a1, 24 - ld.w $fp, $a1, 28 + st.d $a4, $sp, 96 # 8-byte Folded Spill + ld.w $a4, $a1, 24 + st.d $a4, $sp, 56 # 8-byte Folded Spill + ld.w $a4, $a1, 28 + st.d $a4, $sp, 64 # 8-byte Folded Spill addi.d $t6, $a0, 64 - ld.w $t7, $a1, 32 + ld.w $a0, $a1, 32 + st.d $a0, $sp, 72 # 8-byte Folded Spill ori $a4, $zero, 16 ori $a0, $zero, 10 lu32i.d $a0, 3 vreplgr2vr.d $vr0, $a0 .p2align 4, , 16 .LBB4_3: # =>This Inner Loop Header: Depth=1 - st.d $a2, $sp, 64 # 8-byte Folded Spill - st.d $a4, $sp, 32 # 8-byte Folded Spill + st.d $t3, $sp, 16 # 8-byte Folded Spill + st.d $t2, $sp, 24 # 8-byte Folded Spill + st.d $a4, $sp, 48 # 8-byte Folded Spill rotri.w $a0, $s5, 6 rotri.w $s1, $s5, 11 xor $a0, $a0, $s1 @@ -612,6 +605,7 @@ sha256_transform: # @sha256_transform xor $t8, $s1, $t8 bstrpick.d $s1, $a2, 31, 10 xor $t8, $t8, $s1 + ld.d $t1, $sp, 88 # 8-byte Folded Reload rotri.w $s1, $t1, 7 rotri.w $a4, $t1, 18 xor $a4, $s1, $a4 @@ -620,506 +614,486 @@ sha256_transform: # @sha256_transform add.d $t8, $t8, $a3 add.d $a5, $t8, $a5 ld.w $t8, $t6, 0 - move $s1, $t0 add.w $t0, $a5, $a4 add.d $a4, $t5, $s2 add.d $a0, $a4, $a0 add.d $a0, $a0, $t8 - add.d $a4, $a0, $t0 - rotri.w $a0, $s3, 2 + add.d $a0, $a0, $t0 + rotri.w $a4, $s3, 2 rotri.w $t5, $s3, 13 - xor $a0, $a0, $t5 + xor $a4, $a4, $t5 rotri.w $t5, $s3, 22 - xor $a0, $a0, $t5 + xor $a4, $a4, $t5 and $t5, $s3, $s4 xor $t8, $s3, $s4 and $t8, $t8, $s6 xor $t5, $t8, $t5 - add.d $t5, $a0, $t5 - add.d $a0, $a4, $s7 - add.d $s7, $t5, $a4 - rotri.w $a4, $a0, 6 - rotri.w $t5, $a0, 11 - xor $a4, $a4, $t5 - rotri.w $t5, $a0, 25 - xor $a4, $a4, $t5 - andn $t5, $s8, $a0 - and $t8, $s5, $a0 - or $t5, $t8, $t5 - move $a5, $s1 - st.d $s1, $sp, 72 # 8-byte Folded Spill - rotri.w $t8, $s1, 17 - rotri.w $s1, $s1, 19 + add.d $a4, $a4, $t5 + add.d $s7, $a0, $s7 + add.d $s2, $a4, $a0 + rotri.w $a0, $s7, 6 + rotri.w $a4, $s7, 11 + xor $a0, $a0, $a4 + rotri.w $a4, $s7, 25 + xor $a0, $a0, $a4 + andn $a4, $s8, $s7 + and $t5, $s5, $s7 + or $a4, $t5, $a4 + rotri.w $t5, $s0, 17 + rotri.w $t8, $s0, 19 + xor $t5, $t5, $t8 + bstrpick.d $t8, $s0, 31, 10 + st.d $s0, $sp, 32 # 8-byte Folded Spill + xor $t5, $t5, $t8 + rotri.w $t8, $t7, 7 + rotri.w $s1, $t7, 18 xor $t8, $t8, $s1 - bstrpick.d $s1, $a5, 31, 10 + bstrpick.d $s1, $t7, 31, 3 xor $t8, $t8, $s1 - ld.d $a2, $sp, 56 # 8-byte Folded Reload - rotri.w $s1, $a2, 7 - rotri.w $s2, $a2, 18 - xor $s1, $s1, $s2 - bstrpick.d $s2, $a2, 31, 3 - xor $s1, $s1, $s2 st.w $t0, $a1, 0 - ld.w $s2, $t6, 4 + ld.w $s1, $t6, 4 add.d $t1, $a6, $t1 - add.d $t1, $t1, $t8 - add.w $a5, $t1, $s1 - add.d $t1, $s2, $ra add.d $t1, $t1, $t5 + move $fp, $a2 + st.d $a2, $sp, 40 # 8-byte Folded Spill + move $a2, $t7 + add.w $t7, $t1, $t8 + add.d $t1, $s1, $ra add.d $a4, $t1, $a4 - add.d $a4, $a4, $a5 - rotri.w $t1, $s7, 2 - rotri.w $t5, $s7, 13 - xor $t1, $t1, $t5 - rotri.w $t5, $s7, 22 - xor $t1, $t1, $t5 - and $t5, $s7, $s3 - xor $t8, $s7, $s3 - and $t8, $t8, $s4 - xor $t5, $t8, $t5 - add.d $t1, $t1, $t5 - add.d $ra, $a4, $s6 - add.d $s6, $t1, $a4 - rotri.w $a4, $ra, 6 - rotri.w $t1, $ra, 11 + add.d $a0, $a4, $a0 + add.d $a0, $a0, $t7 + rotri.w $a4, $s2, 2 + rotri.w $t1, $s2, 13 xor $a4, $a4, $t1 - rotri.w $t1, $ra, 25 + rotri.w $t1, $s2, 22 xor $a4, $a4, $t1 - and $t1, $a0, $ra - andn $t5, $s5, $ra - or $t1, $t1, $t5 - rotri.w $t5, $t0, 17 - rotri.w $t8, $t0, 19 + and $t1, $s2, $s3 + xor $t5, $s2, $s3 + and $t5, $t5, $s4 + xor $t1, $t5, $t1 + add.d $a4, $a4, $t1 + add.d $ra, $a0, $s6 + add.d $s6, $a4, $a0 + rotri.w $a0, $ra, 6 + rotri.w $a4, $ra, 11 + xor $a0, $a0, $a4 + rotri.w $a4, $ra, 25 + xor $a0, $a0, $a4 + and $a4, $s7, $ra + andn $t1, $s5, $ra + or $a4, $a4, $t1 + rotri.w $t1, $t0, 17 + rotri.w $t5, $t0, 19 + xor $t1, $t1, $t5 + bstrpick.d $t5, $t0, 31, 10 + bstrpick.d $t5, $t5, 53, 0 + xor $t1, $t1, $t5 + rotri.w $t5, $t4, 7 + rotri.w $t8, $t4, 18 xor $t5, $t5, $t8 - bstrpick.d $t8, $t0, 31, 10 - bstrpick.d $t8, $t8, 53, 0 + bstrpick.d $t8, $t4, 31, 3 xor $t5, $t5, $t8 - rotri.w $t8, $t4, 7 - rotri.w $s1, $t4, 18 - xor $t8, $t8, $s1 - bstrpick.d $s1, $t4, 31, 3 - xor $t8, $t8, $s1 - add.d $t5, $t5, $a2 - st.w $a5, $a1, 4 - st.d $a5, $sp, 80 # 8-byte Folded Spill - ld.w $s1, $t6, 8 - add.d $t5, $t5, $a7 - add.w $a2, $t5, $t8 + add.d $t1, $t1, $a2 + st.w $t7, $a1, 4 + ld.w $t8, $t6, 8 + add.d $t1, $t1, $a7 + add.w $a2, $t1, $t5 st.w $a2, $a1, 8 - add.d $t5, $s1, $s8 - add.d $t5, $t5, $a2 - add.d $t1, $t5, $t1 + add.d $t1, $t8, $s8 + add.d $t1, $t1, $a2 add.d $a4, $t1, $a4 - rotri.w $t1, $s6, 2 - rotri.w $t5, $s6, 13 - xor $t1, $t1, $t5 - rotri.w $t5, $s6, 22 - xor $t1, $t1, $t5 - and $t5, $s6, $s7 - xor $t8, $s6, $s7 - and $t8, $t8, $s3 - xor $t5, $t8, $t5 - add.d $t1, $t1, $t5 - add.d $s4, $a4, $s4 - add.d $s2, $t1, $a4 - rotri.w $a4, $s4, 6 - rotri.w $t1, $s4, 11 + add.d $a0, $a4, $a0 + rotri.w $a4, $s6, 2 + rotri.w $t1, $s6, 13 xor $a4, $a4, $t1 - rotri.w $t1, $s4, 25 + rotri.w $t1, $s6, 22 xor $a4, $a4, $t1 - andn $t1, $a0, $s4 - pcalau12i $t5, %pc_hi20(.LCPI4_0) - vld $vr1, $t5, %pc_lo12(.LCPI4_0) - and $t5, $ra, $s4 - or $t1, $t5, $t1 - vinsgr2vr.w $vr5, $a5, 0 - ld.d $s1, $sp, 48 # 8-byte Folded Reload - vinsgr2vr.w $vr5, $s1, 1 - vsrl.w $vr6, $vr5, $vr1 - pcalau12i $t5, %pc_hi20(.LCPI4_1) - vld $vr2, $t5, %pc_lo12(.LCPI4_1) - pcalau12i $t5, %pc_hi20(.LCPI4_2) - vld $vr3, $t5, %pc_lo12(.LCPI4_2) - pcalau12i $t5, %pc_hi20(.LCPI4_3) - vld $vr4, $t5, %pc_lo12(.LCPI4_3) - vsll.w $vr7, $vr5, $vr2 - vor.v $vr6, $vr7, $vr6 - vsrl.w $vr7, $vr5, $vr3 - vsll.w $vr8, $vr5, $vr4 - vor.v $vr7, $vr8, $vr7 - vxor.v $vr6, $vr6, $vr7 - vsrl.w $vr5, $vr5, $vr0 - vxor.v $vr5, $vr6, $vr5 - add.d $t4, $t2, $t4 - vpickve2gr.w $t5, $vr5, 0 - add.d $t4, $t4, $t5 + and $t1, $s6, $s2 + xor $t5, $s6, $s2 + and $t5, $t5, $s3 + xor $t1, $t5, $t1 + add.d $a4, $a4, $t1 + add.d $s8, $a0, $s4 + add.d $s4, $a4, $a0 + rotri.w $a0, $s8, 6 + rotri.w $a4, $s8, 11 + xor $a0, $a0, $a4 + rotri.w $a4, $s8, 25 + xor $a0, $a0, $a4 + andn $a4, $s7, $s8 + and $t1, $ra, $s8 + or $a4, $t1, $a4 + pcalau12i $t1, %pc_hi20(.LCPI4_0) + vld $vr1, $t1, %pc_lo12(.LCPI4_0) + pcalau12i $t1, %pc_hi20(.LCPI4_1) + vld $vr2, $t1, %pc_lo12(.LCPI4_1) + vinsgr2vr.w $vr3, $t7, 0 + ld.d $t8, $sp, 80 # 8-byte Folded Reload + vinsgr2vr.w $vr3, $t8, 1 + vrotr.w $vr4, $vr3, $vr1 + vrotr.w $vr5, $vr3, $vr2 + vxor.v $vr4, $vr4, $vr5 + vsrl.w $vr3, $vr3, $vr0 + vxor.v $vr3, $vr4, $vr3 + add.d $t1, $t2, $t4 + vpickve2gr.w $t4, $vr3, 0 + add.d $t1, $t1, $t4 ld.w $t5, $t6, 12 - vpickve2gr.w $t8, $vr5, 1 - add.w $a5, $t4, $t8 + vpickve2gr.w $t4, $vr3, 1 + add.w $a5, $t1, $t4 st.w $a5, $a1, 12 - add.d $t5, $t5, $s5 - add.d $t5, $t5, $a5 - add.d $t1, $t5, $t1 + add.d $t1, $t5, $s5 + add.d $t1, $t1, $a5 add.d $a4, $t1, $a4 - rotri.w $t1, $s2, 2 - rotri.w $t5, $s2, 13 - xor $t1, $t1, $t5 - rotri.w $t5, $s2, 22 - xor $t1, $t1, $t5 - and $t5, $s2, $s6 - xor $t8, $s2, $s6 - and $t8, $t8, $s7 - xor $t5, $t8, $t5 - add.d $t1, $t1, $t5 - add.d $s8, $a4, $s3 - add.d $s3, $t1, $a4 - rotri.w $a4, $s8, 6 - rotri.w $t1, $s8, 11 + add.d $a4, $a4, $a0 + rotri.w $a0, $s4, 2 + rotri.w $t1, $s4, 13 + xor $a0, $a0, $t1 + rotri.w $t1, $s4, 22 + xor $a0, $a0, $t1 + and $t1, $s4, $s6 + xor $t5, $s4, $s6 + and $t5, $t5, $s2 + xor $t1, $t5, $t1 + add.d $t1, $a0, $t1 + add.d $a0, $a4, $s3 + add.d $s5, $t1, $a4 + rotri.w $a4, $a0, 6 + rotri.w $t1, $a0, 11 xor $a4, $a4, $t1 - rotri.w $t1, $s8, 25 + rotri.w $t1, $a0, 25 xor $a4, $a4, $t1 - andn $t1, $ra, $s8 - and $t5, $s4, $s8 + andn $t1, $ra, $a0 + and $t5, $s8, $a0 or $t1, $t5, $t1 - vinsgr2vr.w $vr5, $a2, 0 - move $t4, $t3 - st.d $t3, $sp, 24 # 8-byte Folded Spill - ld.d $t3, $sp, 40 # 8-byte Folded Reload - vinsgr2vr.w $vr5, $t3, 1 - vsrl.w $vr6, $vr5, $vr1 - vsll.w $vr7, $vr5, $vr2 - vor.v $vr6, $vr7, $vr6 - vsrl.w $vr7, $vr5, $vr3 - vsll.w $vr8, $vr5, $vr4 - vor.v $vr7, $vr8, $vr7 - vxor.v $vr6, $vr6, $vr7 - vsrl.w $vr5, $vr5, $vr0 - vxor.v $vr5, $vr6, $vr5 - add.d $t5, $t4, $s1 - vpickve2gr.w $t8, $vr5, 0 + vinsgr2vr.w $vr3, $a2, 0 + ld.d $s3, $sp, 96 # 8-byte Folded Reload + vinsgr2vr.w $vr3, $s3, 1 + vrotr.w $vr4, $vr3, $vr1 + vrotr.w $vr5, $vr3, $vr2 + vxor.v $vr4, $vr4, $vr5 + vsrl.w $vr3, $vr3, $vr0 + vxor.v $vr3, $vr4, $vr3 + add.d $t5, $t3, $t8 + vpickve2gr.w $t8, $vr3, 0 add.d $t5, $t5, $t8 ld.w $t8, $t6, 16 - vpickve2gr.w $s1, $vr5, 1 + vpickve2gr.w $s1, $vr3, 1 add.d $t4, $t5, $s1 st.w $t4, $a1, 16 - add.d $a0, $t8, $a0 - add.d $a0, $a0, $t4 - add.d $a0, $a0, $t1 - add.d $a0, $a0, $a4 - rotri.w $a4, $s3, 2 - rotri.w $t1, $s3, 13 - xor $a4, $a4, $t1 - rotri.w $t1, $s3, 22 - xor $a4, $a4, $t1 - and $t1, $s3, $s2 - xor $t5, $s3, $s2 - and $t5, $t5, $s6 - xor $t1, $t5, $t1 - add.d $a4, $a4, $t1 - add.d $s7, $a0, $s7 - add.d $a0, $a4, $a0 + add.d $t5, $t8, $s7 + add.d $t5, $t5, $t4 + add.d $t1, $t5, $t1 + add.d $a4, $t1, $a4 + rotri.w $t1, $s5, 2 + rotri.w $t5, $s5, 13 + xor $t1, $t1, $t5 + rotri.w $t5, $s5, 22 + xor $t1, $t1, $t5 + and $t5, $s5, $s4 + xor $t8, $s5, $s4 + and $t8, $t8, $s6 + xor $t5, $t8, $t5 + add.d $t1, $t1, $t5 + add.d $s7, $a4, $s2 + add.d $s2, $t1, $a4 rotri.w $a4, $s7, 6 rotri.w $t1, $s7, 11 xor $a4, $a4, $t1 rotri.w $t1, $s7, 25 xor $a4, $a4, $t1 - andn $t1, $s4, $s7 - and $t5, $s8, $s7 + andn $t1, $s8, $s7 + and $t5, $a0, $s7 or $t1, $t5, $t1 - vinsgr2vr.w $vr5, $a5, 0 - vinsgr2vr.w $vr5, $s0, 1 - vsrl.w $vr6, $vr5, $vr1 - vsll.w $vr7, $vr5, $vr2 - vor.v $vr6, $vr7, $vr6 - vsrl.w $vr7, $vr5, $vr3 - vsll.w $vr8, $vr5, $vr4 - vor.v $vr7, $vr8, $vr7 - vxor.v $vr6, $vr6, $vr7 - vsrl.w $vr5, $vr5, $vr0 - vxor.v $vr5, $vr6, $vr5 - ld.d $t5, $sp, 64 # 8-byte Folded Reload - add.d $t5, $t3, $t5 - vpickve2gr.w $t8, $vr5, 0 + vinsgr2vr.w $vr3, $a5, 0 + move $t2, $a7 + move $a7, $a6 + move $a6, $a3 + ld.d $a3, $sp, 56 # 8-byte Folded Reload + vinsgr2vr.w $vr3, $a3, 1 + vrotr.w $vr4, $vr3, $vr1 + vrotr.w $vr5, $vr3, $vr2 + vxor.v $vr4, $vr4, $vr5 + vsrl.w $vr3, $vr3, $vr0 + vxor.v $vr3, $vr4, $vr3 + add.d $t5, $s3, $fp + vpickve2gr.w $t8, $vr3, 0 add.d $t5, $t5, $t8 ld.w $t8, $t6, 20 - vpickve2gr.w $s1, $vr5, 1 - add.d $t3, $t5, $s1 - st.w $t3, $a1, 20 + vpickve2gr.w $s1, $vr3, 1 + add.d $s1, $t5, $s1 + st.d $s1, $sp, 96 # 8-byte Folded Spill + st.w $s1, $a1, 20 add.d $t5, $t8, $ra - add.d $t5, $t5, $t3 + add.d $t5, $t5, $s1 add.d $t1, $t5, $t1 add.d $a4, $t1, $a4 - rotri.w $t1, $a0, 2 - rotri.w $t5, $a0, 13 + rotri.w $t1, $s2, 2 + rotri.w $t5, $s2, 13 xor $t1, $t1, $t5 - rotri.w $t5, $a0, 22 + rotri.w $t5, $s2, 22 xor $t1, $t1, $t5 - and $t5, $a0, $s3 - xor $t8, $a0, $s3 - and $t8, $t8, $s2 + and $t5, $s2, $s5 + xor $t8, $s2, $s5 + and $t8, $t8, $s4 xor $t5, $t8, $t5 add.d $t1, $t1, $t5 add.d $ra, $a4, $s6 - add.d $s5, $t1, $a4 + add.d $s3, $t1, $a4 rotri.w $a4, $ra, 6 rotri.w $t1, $ra, 11 xor $a4, $a4, $t1 rotri.w $t1, $ra, 25 xor $a4, $a4, $t1 - andn $t1, $s8, $ra + andn $t1, $a0, $ra and $t5, $s7, $ra or $t1, $t5, $t1 - vinsgr2vr.w $vr5, $t4, 0 - vinsgr2vr.w $vr5, $fp, 1 - vsrl.w $vr6, $vr5, $vr1 - vsll.w $vr7, $vr5, $vr2 - vor.v $vr6, $vr7, $vr6 - vsrl.w $vr7, $vr5, $vr3 - vsll.w $vr8, $vr5, $vr4 - vor.v $vr7, $vr8, $vr7 - vxor.v $vr6, $vr6, $vr7 - vsrl.w $vr5, $vr5, $vr0 - vxor.v $vr5, $vr6, $vr5 - ld.d $t5, $sp, 72 # 8-byte Folded Reload - add.d $t5, $s0, $t5 - vpickve2gr.w $t8, $vr5, 0 + vinsgr2vr.w $vr3, $t4, 0 + ld.d $fp, $sp, 64 # 8-byte Folded Reload + vinsgr2vr.w $vr3, $fp, 1 + vrotr.w $vr4, $vr3, $vr1 + vrotr.w $vr5, $vr3, $vr2 + vxor.v $vr4, $vr4, $vr5 + vsrl.w $vr3, $vr3, $vr0 + vxor.v $vr3, $vr4, $vr3 + add.d $t5, $a3, $s0 + vpickve2gr.w $t8, $vr3, 0 add.d $t5, $t5, $t8 ld.w $t8, $t6, 24 - vpickve2gr.w $s0, $vr5, 1 - add.w $s0, $t5, $s0 - st.w $s0, $a1, 24 - add.d $t5, $t8, $s4 - add.d $t5, $t5, $s0 + vpickve2gr.w $s0, $vr3, 1 + add.w $t3, $t5, $s0 + st.d $t3, $sp, 56 # 8-byte Folded Spill + st.w $t3, $a1, 24 + add.d $t5, $t8, $s8 + add.d $t5, $t5, $t3 add.d $t1, $t5, $t1 add.d $a4, $t1, $a4 - rotri.w $t1, $s5, 2 - rotri.w $t5, $s5, 13 + rotri.w $t1, $s3, 2 + rotri.w $t5, $s3, 13 xor $t1, $t1, $t5 - rotri.w $t5, $s5, 22 + rotri.w $t5, $s3, 22 xor $t1, $t1, $t5 - and $t5, $s5, $a0 - xor $t8, $s5, $a0 - and $t8, $t8, $s3 + and $t5, $s3, $s2 + xor $t8, $s3, $s2 + and $t8, $t8, $s5 xor $t5, $t8, $t5 add.d $t1, $t1, $t5 - add.d $s1, $a4, $s2 + add.d $s8, $a4, $s4 add.d $s4, $t1, $a4 - rotri.w $a4, $s1, 6 - rotri.w $t1, $s1, 11 + rotri.w $a4, $s8, 6 + rotri.w $t1, $s8, 11 xor $a4, $a4, $t1 - rotri.w $t1, $s1, 25 + rotri.w $t1, $s8, 25 xor $a4, $a4, $t1 - and $t1, $ra, $s1 - andn $t5, $s7, $s1 + and $t1, $ra, $s8 + andn $t5, $s7, $s8 or $t1, $t1, $t5 ld.w $t5, $t6, 28 - vinsgr2vr.w $vr5, $t3, 0 - vinsgr2vr.w $vr5, $t7, 1 - vsrl.w $vr1, $vr5, $vr1 - vsll.w $vr2, $vr5, $vr2 - vor.v $vr1, $vr2, $vr1 - vsrl.w $vr2, $vr5, $vr3 - vsll.w $vr3, $vr5, $vr4 - vor.v $vr2, $vr3, $vr2 + vinsgr2vr.w $vr3, $s1, 0 + ld.d $s1, $sp, 72 # 8-byte Folded Reload + vinsgr2vr.w $vr3, $s1, 1 + vrotr.w $vr1, $vr3, $vr1 + vrotr.w $vr2, $vr3, $vr2 vxor.v $vr1, $vr1, $vr2 - vsrl.w $vr2, $vr5, $vr0 + vsrl.w $vr2, $vr3, $vr0 vxor.v $vr1, $vr1, $vr2 add.d $t8, $fp, $t0 vpickve2gr.w $fp, $vr1, 0 add.d $t8, $t8, $fp vpickve2gr.w $fp, $vr1, 1 - add.w $fp, $t8, $fp - add.d $t5, $s8, $t5 - add.d $t5, $t5, $fp - add.d $t1, $t5, $t1 - add.d $a4, $t1, $a4 - rotri.w $t1, $s4, 2 - rotri.w $t5, $s4, 13 - xor $t1, $t1, $t5 - rotri.w $t5, $s4, 22 - xor $t1, $t1, $t5 - and $t5, $s4, $s5 - xor $t8, $s4, $s5 - and $t8, $t8, $a0 - xor $t5, $t8, $t5 - add.d $t1, $t1, $t5 - add.d $s8, $a4, $s3 - add.d $s6, $t1, $a4 - rotri.w $a4, $s8, 6 - rotri.w $t1, $s8, 11 + add.w $s0, $t8, $fp + add.d $a0, $a0, $t5 + add.d $a0, $a0, $s0 + add.d $a0, $a0, $t1 + add.d $a0, $a0, $a4 + rotri.w $a4, $s4, 2 + rotri.w $t1, $s4, 13 xor $a4, $a4, $t1 - rotri.w $t1, $s8, 25 + rotri.w $t1, $s4, 22 xor $a4, $a4, $t1 - andn $t1, $ra, $s8 - and $t5, $s1, $s8 - or $t1, $t5, $t1 - rotri.w $t5, $s0, 17 - rotri.w $t8, $s0, 19 - xor $t5, $t5, $t8 - bstrpick.d $t8, $s0, 31, 10 - xor $t5, $t5, $t8 - rotri.w $t8, $a3, 7 - rotri.w $s2, $a3, 18 - xor $t8, $t8, $s2 - bstrpick.d $s2, $a3, 31, 3 - xor $t8, $t8, $s2 - st.w $fp, $a1, 28 - ld.w $s2, $t6, 32 - ld.d $s3, $sp, 80 # 8-byte Folded Reload - add.d $t8, $s3, $t8 - add.d $t7, $t8, $t7 - add.w $t7, $t7, $t5 - add.d $t5, $t7, $s2 - add.d $t5, $t5, $s7 - add.d $t1, $t5, $t1 - add.d $a4, $t1, $a4 - rotri.w $t1, $s6, 2 - rotri.w $t5, $s6, 13 - xor $t1, $t1, $t5 - rotri.w $t5, $s6, 22 - xor $t1, $t1, $t5 - and $t5, $s6, $s4 - xor $t8, $s6, $s4 - and $t8, $t8, $s5 - xor $t5, $t8, $t5 - add.d $t1, $t1, $t5 - add.d $s7, $a4, $a0 - add.d $s2, $t1, $a4 - rotri.w $a0, $s7, 6 - rotri.w $a4, $s7, 11 + and $t1, $s4, $s3 + xor $t5, $s4, $s3 + and $t5, $t5, $s2 + xor $t1, $t5, $t1 + add.d $a4, $a4, $t1 + add.d $s6, $a0, $s5 + add.d $s5, $a4, $a0 + rotri.w $a0, $s6, 6 + rotri.w $a4, $s6, 11 xor $a0, $a0, $a4 - rotri.w $a4, $s7, 25 + rotri.w $a4, $s6, 25 xor $a0, $a0, $a4 - andn $a4, $s1, $s7 - and $t1, $s8, $s7 + andn $a4, $ra, $s6 + and $t1, $s8, $s6 or $a4, $t1, $a4 - rotri.w $t1, $fp, 17 - rotri.w $t5, $fp, 19 + rotri.w $t1, $t3, 17 + rotri.w $t5, $t3, 19 xor $t1, $t1, $t5 - bstrpick.d $t5, $fp, 31, 10 + bstrpick.d $t5, $t3, 31, 10 xor $t1, $t1, $t5 rotri.w $t5, $a6, 7 rotri.w $t8, $a6, 18 xor $t5, $t5, $t8 bstrpick.d $t8, $a6, 31, 3 xor $t5, $t5, $t8 - st.w $t7, $a1, 32 - ld.w $t8, $t6, 36 - add.d $a3, $t5, $a3 - st.d $a2, $sp, 56 # 8-byte Folded Spill - add.d $a3, $a3, $a2 - add.w $a3, $a3, $t1 - add.d $t1, $a3, $t8 - add.d $t1, $t1, $ra + st.w $s0, $a1, 28 + ld.w $t8, $t6, 32 + st.d $t7, $sp, 88 # 8-byte Folded Spill + add.d $t5, $t7, $t5 + move $t7, $a2 + add.d $t5, $t5, $s1 + add.w $fp, $t5, $t1 + add.d $t1, $fp, $t8 + add.d $t1, $t1, $s7 add.d $a4, $t1, $a4 add.d $a0, $a4, $a0 - rotri.w $a4, $s2, 2 - rotri.w $t1, $s2, 13 + rotri.w $a4, $s5, 2 + rotri.w $t1, $s5, 13 xor $a4, $a4, $t1 - rotri.w $t1, $s2, 22 + rotri.w $t1, $s5, 22 xor $a4, $a4, $t1 - and $t1, $s2, $s6 - xor $t5, $s2, $s6 - and $t5, $t5, $s4 + and $t1, $s5, $s4 + xor $t5, $s5, $s4 + and $t5, $t5, $s3 xor $t1, $t5, $t1 add.d $a4, $a4, $t1 - add.d $s5, $a0, $s5 - add.d $s3, $a4, $a0 - rotri.w $a0, $s5, 6 - rotri.w $a4, $s5, 11 + add.d $s7, $a0, $s2 + add.d $s2, $a4, $a0 + rotri.w $a0, $s7, 6 + rotri.w $a4, $s7, 11 xor $a0, $a0, $a4 - rotri.w $a4, $s5, 25 + rotri.w $a4, $s7, 25 xor $a0, $a0, $a4 - andn $a4, $s8, $s5 - and $t1, $s7, $s5 + andn $a4, $s8, $s7 + and $t1, $s6, $s7 or $a4, $t1, $a4 - rotri.w $t1, $t7, 17 - rotri.w $t5, $t7, 19 + rotri.w $t1, $s0, 17 + rotri.w $t5, $s0, 19 xor $t1, $t1, $t5 - bstrpick.d $t5, $t7, 31, 10 + bstrpick.d $t5, $s0, 31, 10 xor $t1, $t1, $t5 rotri.w $t5, $a7, 7 rotri.w $t8, $a7, 18 xor $t5, $t5, $t8 bstrpick.d $t8, $a7, 31, 3 xor $t5, $t5, $t8 - st.w $a3, $a1, 36 - ld.w $t8, $t6, 40 - add.d $a6, $t5, $a6 - add.d $a6, $a6, $a5 - add.w $a6, $a6, $t1 - add.d $t1, $a6, $t8 - add.d $t1, $t1, $s1 + st.w $fp, $a1, 32 + ld.w $t8, $t6, 36 + add.d $a3, $t5, $a6 + add.d $a3, $a3, $a2 + add.w $a3, $a3, $t1 + add.d $t1, $a3, $t8 + add.d $t1, $t1, $ra add.d $a4, $t1, $a4 - add.d $a0, $a4, $a0 - rotri.w $a4, $s3, 2 - rotri.w $t1, $s3, 13 + add.d $a4, $a4, $a0 + rotri.w $a0, $s2, 2 + rotri.w $t1, $s2, 13 + xor $a0, $a0, $t1 + rotri.w $t1, $s2, 22 + xor $a0, $a0, $t1 + and $t1, $s2, $s5 + xor $t5, $s2, $s5 + and $t5, $t5, $s4 + xor $t1, $t5, $t1 + add.d $t1, $a0, $t1 + add.d $a0, $a4, $s3 + add.d $s3, $t1, $a4 + rotri.w $a4, $a0, 6 + rotri.w $t1, $a0, 11 xor $a4, $a4, $t1 - rotri.w $t1, $s3, 22 + rotri.w $t1, $a0, 25 xor $a4, $a4, $t1 - and $t1, $s3, $s2 - xor $t5, $s3, $s2 - and $t5, $t5, $s6 - xor $t1, $t5, $t1 - add.d $a4, $a4, $t1 - add.d $s1, $a0, $s4 - add.d $s4, $a4, $a0 - rotri.w $a0, $s1, 6 - rotri.w $a4, $s1, 11 - xor $a0, $a0, $a4 - rotri.w $a4, $s1, 25 - xor $a0, $a0, $a4 - andn $a4, $s7, $s1 - and $t1, $s5, $s1 - or $a4, $t1, $a4 - rotri.w $t1, $a3, 17 - rotri.w $t5, $a3, 19 + andn $t1, $s6, $a0 + and $t5, $s7, $a0 + or $t1, $t5, $t1 + rotri.w $t5, $fp, 17 + rotri.w $t8, $fp, 19 + xor $t5, $t5, $t8 + bstrpick.d $t8, $fp, 31, 10 + xor $t5, $t5, $t8 + rotri.w $t8, $t2, 7 + rotri.w $s1, $t2, 18 + xor $t8, $t8, $s1 + bstrpick.d $s1, $t2, 31, 3 + xor $t8, $t8, $s1 + st.w $a3, $a1, 36 + ld.w $s1, $t6, 40 + add.d $a6, $t8, $a7 + add.d $a6, $a6, $a5 + add.w $a6, $a6, $t5 + add.d $t5, $a6, $s1 + add.d $t5, $t5, $s8 + add.d $t1, $t5, $t1 + add.d $a4, $t1, $a4 + rotri.w $t1, $s3, 2 + rotri.w $t5, $s3, 13 xor $t1, $t1, $t5 - bstrpick.d $t5, $a3, 31, 10 + rotri.w $t5, $s3, 22 xor $t1, $t1, $t5 - rotri.w $t5, $t2, 7 - rotri.w $t8, $t2, 18 + and $t5, $s3, $s2 + xor $t8, $s3, $s2 + and $t8, $t8, $s5 + xor $t5, $t8, $t5 + add.d $t1, $t1, $t5 + add.d $s8, $a4, $s4 + add.d $s4, $t1, $a4 + rotri.w $a4, $s8, 6 + rotri.w $t1, $s8, 11 + xor $a4, $a4, $t1 + rotri.w $t1, $s8, 25 + xor $a4, $a4, $t1 + andn $t1, $s7, $s8 + and $t5, $a0, $s8 + or $t1, $t5, $t1 + rotri.w $t5, $a3, 17 + rotri.w $t8, $a3, 19 xor $t5, $t5, $t8 - bstrpick.d $t8, $t2, 31, 3 + bstrpick.d $t8, $a3, 31, 10 xor $t5, $t5, $t8 + ld.d $a2, $sp, 24 # 8-byte Folded Reload + rotri.w $t8, $a2, 7 + rotri.w $s1, $a2, 18 + xor $t8, $t8, $s1 + bstrpick.d $s1, $a2, 31, 3 + xor $t8, $t8, $s1 st.w $a6, $a1, 40 - ld.w $t8, $t6, 44 - add.d $a7, $t5, $a7 - st.d $t4, $sp, 48 # 8-byte Folded Spill + ld.w $s1, $t6, 44 + add.d $a7, $t8, $t2 + st.d $t4, $sp, 80 # 8-byte Folded Spill add.d $a7, $a7, $t4 move $t4, $a5 move $a5, $t0 - add.w $a7, $a7, $t1 - add.d $t1, $a7, $t8 - add.d $t1, $t1, $s8 + add.w $a7, $a7, $t5 + add.d $t5, $a7, $s1 + add.d $t5, $t5, $s6 + add.d $t1, $t5, $t1 add.d $a4, $t1, $a4 - add.d $a0, $a4, $a0 - rotri.w $a4, $s4, 2 - rotri.w $t1, $s4, 13 - xor $a4, $a4, $t1 - rotri.w $t1, $s4, 22 - xor $a4, $a4, $t1 - and $t1, $s4, $s3 - xor $t5, $s4, $s3 - and $t5, $t5, $s2 - xor $t1, $t5, $t1 - add.d $a4, $a4, $t1 - add.d $s8, $a0, $s6 - add.d $a0, $a4, $a0 - rotri.w $a4, $s8, 6 - rotri.w $t1, $s8, 11 + rotri.w $t1, $s4, 2 + rotri.w $t5, $s4, 13 + xor $t1, $t1, $t5 + rotri.w $t5, $s4, 22 + xor $t1, $t1, $t5 + and $t5, $s4, $s3 + xor $t8, $s4, $s3 + and $t8, $t8, $s2 + xor $t5, $t8, $t5 + add.d $t1, $t1, $t5 + add.d $s1, $a4, $s5 + add.d $s5, $t1, $a4 + rotri.w $a4, $s1, 6 + rotri.w $t1, $s1, 11 xor $a4, $a4, $t1 - rotri.w $t1, $s8, 25 + rotri.w $t1, $s1, 25 xor $a4, $a4, $t1 - andn $t1, $s5, $s8 - and $t5, $s1, $s8 + andn $t1, $a0, $s1 + and $t5, $s8, $s1 or $t1, $t5, $t1 rotri.w $t5, $a6, 17 rotri.w $t8, $a6, 19 xor $t5, $t5, $t8 bstrpick.d $t8, $a6, 31, 10 xor $t5, $t5, $t8 - ld.d $t0, $sp, 24 # 8-byte Folded Reload + ld.d $t0, $sp, 16 # 8-byte Folded Reload rotri.w $t8, $t0, 7 rotri.w $s6, $t0, 18 xor $t8, $t8, $s6 @@ -1127,21 +1101,21 @@ sha256_transform: # @sha256_transform xor $t8, $t8, $s6 st.w $a7, $a1, 44 ld.w $s6, $t6, 48 - add.d $t2, $t8, $t2 - st.d $t3, $sp, 40 # 8-byte Folded Spill - add.d $t2, $t2, $t3 + add.d $t2, $t8, $a2 + ld.d $a2, $sp, 96 # 8-byte Folded Reload + add.d $t2, $t2, $a2 add.w $t2, $t2, $t5 add.d $t5, $t2, $s6 add.d $t5, $t5, $s7 add.d $t1, $t5, $t1 add.d $s6, $t1, $a4 - rotri.w $a4, $a0, 2 - rotri.w $t1, $a0, 13 + rotri.w $a4, $s5, 2 + rotri.w $t1, $s5, 13 xor $a4, $a4, $t1 - rotri.w $t1, $a0, 22 + rotri.w $t1, $s5, 22 xor $a4, $a4, $t1 - and $t1, $a0, $s4 - xor $t5, $a0, $s4 + and $t1, $s5, $s4 + xor $t5, $s5, $s4 and $t5, $t5, $s3 xor $t1, $t5, $t1 add.d $s7, $a4, $t1 @@ -1150,132 +1124,133 @@ sha256_transform: # @sha256_transform xor $a4, $a4, $t1 bstrpick.d $t1, $a7, 31, 10 xor $a4, $a4, $t1 - ld.d $ra, $sp, 64 # 8-byte Folded Reload - rotri.w $t1, $ra, 7 - rotri.w $t5, $ra, 18 + ld.d $t8, $sp, 40 # 8-byte Folded Reload + rotri.w $t1, $t8, 7 + rotri.w $t5, $t8, 18 xor $t1, $t1, $t5 - bstrpick.d $t5, $ra, 31, 3 + bstrpick.d $t5, $t8, 31, 3 bstrpick.d $t5, $t5, 60, 0 xor $t1, $t1, $t5 st.w $t2, $a1, 48 ld.w $t5, $t6, 52 add.d $t1, $t0, $t1 - add.d $t1, $t1, $s0 + add.d $t1, $t1, $t3 add.w $t3, $t1, $a4 add.d $a4, $t3, $t5 - add.d $a4, $a4, $s5 - rotri.w $t1, $t2, 17 - rotri.w $t5, $t2, 19 - xor $t1, $t1, $t5 - bstrpick.d $t5, $t2, 31, 10 + add.d $a0, $a4, $a0 + rotri.w $a4, $t2, 17 + rotri.w $t1, $t2, 19 + xor $a4, $a4, $t1 + bstrpick.d $t1, $t2, 31, 10 + xor $a4, $a4, $t1 + ld.d $ra, $sp, 32 # 8-byte Folded Reload + rotri.w $t1, $ra, 7 + rotri.w $t5, $ra, 18 xor $t1, $t1, $t5 - ld.d $t0, $sp, 72 # 8-byte Folded Reload - rotri.w $t5, $t0, 7 - rotri.w $t8, $t0, 18 - xor $t5, $t5, $t8 - bstrpick.d $t8, $t0, 31, 3 - xor $t5, $t5, $t8 - add.d $a2, $t5, $ra - add.d $a2, $a2, $fp - add.w $a2, $a2, $t1 - rotri.w $t1, $t3, 17 - rotri.w $t5, $t3, 19 + bstrpick.d $t5, $ra, 31, 3 xor $t1, $t1, $t5 - bstrpick.d $t5, $t3, 31, 10 + add.d $a2, $t1, $t8 + st.d $s0, $sp, 64 # 8-byte Folded Spill + add.d $a2, $a2, $s0 + add.w $a2, $a2, $a4 + rotri.w $a4, $t3, 17 + rotri.w $t1, $t3, 19 + xor $a4, $a4, $t1 + bstrpick.d $t1, $t3, 31, 10 + xor $a4, $a4, $t1 + rotri.w $t1, $a5, 7 + rotri.w $t5, $a5, 18 xor $t1, $t1, $t5 - rotri.w $t5, $a5, 7 - rotri.w $t8, $a5, 18 - xor $t5, $t5, $t8 st.w $t3, $a1, 52 - ld.w $t8, $t6, 56 - bstrpick.d $s5, $a5, 31, 3 - xor $t5, $t5, $s5 - add.d $t0, $t5, $t0 + ld.w $t5, $t6, 56 + bstrpick.d $t8, $a5, 31, 3 + xor $t1, $t1, $t8 + add.d $t0, $t1, $ra st.w $a2, $a1, 56 - add.d $t5, $a2, $t8 - add.d $t5, $t5, $s1 - add.d $t0, $t0, $t7 - add.w $t0, $t0, $t1 + add.d $t1, $a2, $t5 + add.d $t1, $t1, $s8 + st.d $fp, $sp, 72 # 8-byte Folded Spill + add.d $t0, $t0, $fp + add.w $s0, $t0, $a4 add.d $s2, $s6, $s2 add.d $s7, $s7, $s6 - rotri.w $t1, $s2, 6 - rotri.w $t8, $s2, 11 - xor $t1, $t1, $t8 - rotri.w $t8, $s2, 25 - xor $t1, $t1, $t8 - andn $t8, $s1, $s2 - and $s1, $s8, $s2 - or $t8, $s1, $t8 - ld.w $s1, $t6, 60 - add.d $a4, $a4, $t8 - add.d $a4, $a4, $t1 - rotri.w $t1, $s7, 2 - rotri.w $t8, $s7, 13 - xor $t1, $t1, $t8 - rotri.w $t8, $s7, 22 - xor $t1, $t1, $t8 - xor $t8, $s7, $a0 - and $t8, $t8, $s4 - and $s5, $s7, $a0 - xor $t8, $t8, $s5 - add.d $s1, $t0, $s1 - add.d $s1, $s1, $s8 - add.d $t1, $t1, $t8 - add.d $ra, $a4, $s3 - add.d $s6, $t1, $a4 - rotri.w $a4, $ra, 6 - rotri.w $t1, $ra, 11 + rotri.w $a4, $s2, 6 + rotri.w $t5, $s2, 11 + xor $a4, $a4, $t5 + rotri.w $t5, $s2, 25 + xor $a4, $a4, $t5 + andn $t5, $s8, $s2 + and $t8, $s1, $s2 + or $t5, $t8, $t5 + ld.w $t8, $t6, 60 + add.d $a0, $a0, $t5 + add.d $a0, $a0, $a4 + rotri.w $a4, $s7, 2 + rotri.w $t5, $s7, 13 + xor $a4, $a4, $t5 + rotri.w $t5, $s7, 22 + xor $a4, $a4, $t5 + xor $t5, $s7, $s5 + and $t5, $t5, $s4 + and $s6, $s7, $s5 + xor $t5, $t5, $s6 + add.d $t8, $s0, $t8 + add.d $t8, $t8, $s1 + add.d $a4, $a4, $t5 + add.d $ra, $a0, $s3 + add.d $s6, $a4, $a0 + rotri.w $a0, $ra, 6 + rotri.w $a4, $ra, 11 + xor $a0, $a0, $a4 + rotri.w $a4, $ra, 25 + xor $a0, $a0, $a4 + andn $a4, $s1, $ra + and $t5, $s2, $ra + or $a4, $t5, $a4 + add.d $a4, $t1, $a4 + add.d $a0, $a4, $a0 + rotri.w $a4, $s6, 2 + rotri.w $t1, $s6, 13 xor $a4, $a4, $t1 - rotri.w $t1, $ra, 25 + rotri.w $t1, $s6, 22 xor $a4, $a4, $t1 - andn $t1, $s8, $ra - and $t8, $s2, $ra - or $t1, $t8, $t1 - add.d $t1, $t5, $t1 - add.d $a4, $t1, $a4 - rotri.w $t1, $s6, 2 - rotri.w $t5, $s6, 13 - xor $t1, $t1, $t5 - rotri.w $t5, $s6, 22 - xor $t1, $t1, $t5 - and $t5, $s6, $s7 - xor $t8, $s6, $s7 - and $t8, $t8, $a0 - xor $t5, $t8, $t5 - add.d $t1, $t1, $t5 - add.d $s8, $a4, $s4 - add.d $s4, $t1, $a4 - rotri.w $a4, $s8, 6 - rotri.w $t1, $s8, 11 + and $t1, $s6, $s7 + xor $t5, $s6, $s7 + and $t5, $t5, $s5 + xor $t1, $t5, $t1 + add.d $a4, $a4, $t1 + add.d $s8, $a0, $s4 + add.d $s4, $a4, $a0 + rotri.w $a0, $s8, 6 + rotri.w $a4, $s8, 11 + xor $a0, $a0, $a4 + rotri.w $a4, $s8, 25 + xor $a0, $a0, $a4 + andn $a4, $s2, $s8 + and $t1, $ra, $s8 + or $a4, $t1, $a4 + add.d $a4, $t8, $a4 + add.d $a0, $a4, $a0 + rotri.w $a4, $s4, 2 + rotri.w $t1, $s4, 13 xor $a4, $a4, $t1 - rotri.w $t1, $s8, 25 + rotri.w $t1, $s4, 22 xor $a4, $a4, $t1 - andn $t1, $s2, $s8 - and $t5, $ra, $s8 - or $t1, $t5, $t1 - add.d $t1, $s1, $t1 - add.d $a4, $t1, $a4 - rotri.w $t1, $s4, 2 - rotri.w $t5, $s4, 13 - xor $t1, $t1, $t5 - rotri.w $t5, $s4, 22 - xor $t1, $t1, $t5 - and $t5, $s4, $s6 - xor $t8, $s4, $s6 - and $t8, $t8, $s7 - xor $t5, $t8, $t5 - add.d $t1, $t1, $t5 - ld.d $t5, $sp, 32 # 8-byte Folded Reload - st.w $t0, $a1, 60 - add.d $s5, $a4, $a0 - add.d $s3, $t1, $a4 - ld.d $t1, $sp, 80 # 8-byte Folded Reload - addi.d $a4, $t5, 16 + and $t1, $s4, $s6 + xor $t5, $s4, $s6 + and $t5, $t5, $s7 + xor $t1, $t5, $t1 + add.d $a4, $a4, $t1 + ld.d $t1, $sp, 48 # 8-byte Folded Reload + st.w $s0, $a1, 60 + add.d $s5, $a0, $s5 + add.d $s3, $a4, $a0 + addi.d $a4, $t1, 16 addi.d $t6, $t6, 64 ori $a0, $zero, 48 - bltu $t5, $a0, .LBB4_3 + bltu $t1, $a0, .LBB4_3 # %bb.4: # %._crit_edge - ld.d $a0, $sp, 16 # 8-byte Folded Reload + ld.d $a0, $sp, 8 # 8-byte Folded Reload xvld $xr0, $a0, 0 vinsgr2vr.w $vr1, $s5, 0 vinsgr2vr.w $vr1, $s8, 1 @@ -1288,18 +1263,18 @@ sha256_transform: # @sha256_transform xvpermi.q $xr2, $xr1, 2 xvadd.w $xr0, $xr0, $xr2 xvst $xr0, $a0, 0 - ld.d $s8, $sp, 88 # 8-byte Folded Reload - ld.d $s7, $sp, 96 # 8-byte Folded Reload - ld.d $s6, $sp, 104 # 8-byte Folded Reload - ld.d $s5, $sp, 112 # 8-byte Folded Reload - ld.d $s4, $sp, 120 # 8-byte Folded Reload - ld.d $s3, $sp, 128 # 8-byte Folded Reload - ld.d $s2, $sp, 136 # 8-byte Folded Reload - ld.d $s1, $sp, 144 # 8-byte Folded Reload - ld.d $s0, $sp, 152 # 8-byte Folded Reload - ld.d $fp, $sp, 160 # 8-byte Folded Reload - ld.d $ra, $sp, 168 # 8-byte Folded Reload - addi.d $sp, $sp, 176 + ld.d $s8, $sp, 104 # 8-byte Folded Reload + ld.d $s7, $sp, 112 # 8-byte Folded Reload + ld.d $s6, $sp, 120 # 8-byte Folded Reload + ld.d $s5, $sp, 128 # 8-byte Folded Reload + ld.d $s4, $sp, 136 # 8-byte Folded Reload + ld.d $s3, $sp, 144 # 8-byte Folded Reload + ld.d $s2, $sp, 152 # 8-byte Folded Reload + ld.d $s1, $sp, 160 # 8-byte Folded Reload + ld.d $s0, $sp, 168 # 8-byte Folded Reload + ld.d $fp, $sp, 176 # 8-byte Folded Reload + ld.d $ra, $sp, 184 # 8-byte Folded Reload + addi.d $sp, $sp, 192 ret .Lfunc_end4: .size sha256_transform, .Lfunc_end4-sha256_transform diff --git a/results/SingleSource/Regression/C/gcc-c-torture/execute/CMakeFiles/GCC-C-execute-pr56866.dir/pr56866.s b/results/SingleSource/Regression/C/gcc-c-torture/execute/CMakeFiles/GCC-C-execute-pr56866.dir/pr56866.s index a202520..db49a05 100644 --- a/results/SingleSource/Regression/C/gcc-c-torture/execute/CMakeFiles/GCC-C-execute-pr56866.dir/pr56866.s +++ b/results/SingleSource/Regression/C/gcc-c-torture/execute/CMakeFiles/GCC-C-execute-pr56866.dir/pr56866.s @@ -75,15 +75,11 @@ main: # @main .p2align 4, , 16 .LBB0_1: # %vector.body # =>This Inner Loop Header: Depth=1 - xvldx $xr0, $a1, $a0 add.d $a4, $a0, $a1 + xvldx $xr0, $a1, $a0 xvld $xr1, $a4, 32 - xvsrli.d $xr2, $xr0, 8 - xvslli.d $xr0, $xr0, 56 - xvor.v $xr0, $xr0, $xr2 - xvsrli.d $xr2, $xr1, 8 - xvslli.d $xr1, $xr1, 56 - xvor.v $xr1, $xr1, $xr2 + xvrotri.d $xr0, $xr0, 8 + xvrotri.d $xr1, $xr1, 8 add.d $a4, $a2, $a1 xvstx $xr0, $a1, $a2 addi.d $a1, $a1, 64 @@ -96,380 +92,268 @@ main: # @main ori $a0, $zero, 2664 add.d $a0, $sp, $a0 xvld $xr1, $a0, 0 - xvsrli.w $xr2, $xr0, 8 - xvslli.w $xr0, $xr0, 24 - xvor.v $xr0, $xr0, $xr2 - xvsrli.w $xr2, $xr1, 8 - xvslli.w $xr1, $xr1, 24 - xvor.v $xr1, $xr1, $xr2 + xvrotri.w $xr0, $xr0, 8 + xvst $xr0, $sp, 1608 ori $a0, $zero, 2696 add.d $a0, $sp, $a0 - xvld $xr2, $a0, 0 - xvst $xr0, $sp, 1608 - xvst $xr1, $sp, 1640 + xvld $xr0, $a0, 0 ori $a0, $zero, 2728 add.d $a0, $sp, $a0 - xvld $xr0, $a0, 0 - xvsrli.w $xr1, $xr2, 8 - xvslli.w $xr2, $xr2, 24 - xvor.v $xr1, $xr2, $xr1 - xvsrli.w $xr2, $xr0, 8 - xvslli.w $xr0, $xr0, 24 - xvor.v $xr0, $xr0, $xr2 + xvld $xr2, $a0, 0 + xvrotri.w $xr1, $xr1, 8 + xvst $xr1, $sp, 1640 + xvrotri.w $xr0, $xr0, 8 + xvrotri.w $xr1, $xr2, 8 ori $a0, $zero, 2760 add.d $a0, $sp, $a0 xvld $xr2, $a0, 0 - xvst $xr1, $sp, 1672 - xvst $xr0, $sp, 1704 + xvst $xr0, $sp, 1672 + xvst $xr1, $sp, 1704 ori $a0, $zero, 2792 add.d $a0, $sp, $a0 xvld $xr0, $a0, 0 - xvsrli.w $xr1, $xr2, 8 - xvslli.w $xr2, $xr2, 24 - xvor.v $xr1, $xr2, $xr1 - xvsrli.w $xr2, $xr0, 8 - xvslli.w $xr0, $xr0, 24 - xvor.v $xr0, $xr0, $xr2 + xvrotri.w $xr1, $xr2, 8 + xvst $xr1, $sp, 1736 ori $a0, $zero, 2824 add.d $a0, $sp, $a0 - xvld $xr2, $a0, 0 - xvst $xr1, $sp, 1736 - xvst $xr0, $sp, 1768 + xvld $xr1, $a0, 0 ori $a0, $zero, 2856 add.d $a0, $sp, $a0 - xvld $xr0, $a0, 0 - xvsrli.w $xr1, $xr2, 8 - xvslli.w $xr2, $xr2, 24 - xvor.v $xr1, $xr2, $xr1 - xvsrli.w $xr2, $xr0, 8 - xvslli.w $xr0, $xr0, 24 - xvor.v $xr0, $xr0, $xr2 + xvld $xr2, $a0, 0 + xvrotri.w $xr0, $xr0, 8 + xvst $xr0, $sp, 1768 + xvrotri.w $xr0, $xr1, 8 + xvrotri.w $xr1, $xr2, 8 ori $a0, $zero, 2888 add.d $a0, $sp, $a0 xvld $xr2, $a0, 0 - xvst $xr1, $sp, 1800 - xvst $xr0, $sp, 1832 + xvst $xr0, $sp, 1800 + xvst $xr1, $sp, 1832 ori $a0, $zero, 2920 add.d $a0, $sp, $a0 xvld $xr0, $a0, 0 - xvsrli.w $xr1, $xr2, 8 - xvslli.w $xr2, $xr2, 24 - xvor.v $xr1, $xr2, $xr1 - xvsrli.w $xr2, $xr0, 8 - xvslli.w $xr0, $xr0, 24 - xvor.v $xr0, $xr0, $xr2 + xvrotri.w $xr1, $xr2, 8 + xvst $xr1, $sp, 1864 ori $a0, $zero, 2952 add.d $a0, $sp, $a0 - xvld $xr2, $a0, 0 - xvst $xr1, $sp, 1864 - xvst $xr0, $sp, 1896 + xvld $xr1, $a0, 0 ori $a0, $zero, 2984 add.d $a0, $sp, $a0 - xvld $xr0, $a0, 0 - xvsrli.w $xr1, $xr2, 8 - xvslli.w $xr2, $xr2, 24 - xvor.v $xr1, $xr2, $xr1 - xvsrli.w $xr2, $xr0, 8 - xvslli.w $xr0, $xr0, 24 - xvor.v $xr0, $xr0, $xr2 + xvld $xr2, $a0, 0 + xvrotri.w $xr0, $xr0, 8 + xvst $xr0, $sp, 1896 + xvrotri.w $xr0, $xr1, 8 + xvrotri.w $xr1, $xr2, 8 ori $a0, $zero, 3016 add.d $a0, $sp, $a0 xvld $xr2, $a0, 0 - xvst $xr1, $sp, 1928 - xvst $xr0, $sp, 1960 + xvst $xr0, $sp, 1928 + xvst $xr1, $sp, 1960 ori $a0, $zero, 3048 add.d $a0, $sp, $a0 xvld $xr0, $a0, 0 - xvsrli.w $xr1, $xr2, 8 - xvslli.w $xr2, $xr2, 24 - xvor.v $xr1, $xr2, $xr1 - xvsrli.w $xr2, $xr0, 8 - xvslli.w $xr0, $xr0, 24 - xvor.v $xr0, $xr0, $xr2 + xvrotri.w $xr1, $xr2, 8 + xvst $xr1, $sp, 1992 ori $a0, $zero, 3080 add.d $a0, $sp, $a0 - xvld $xr2, $a0, 0 - xvst $xr1, $sp, 1992 - xvst $xr0, $sp, 2024 + xvld $xr1, $a0, 0 ori $a0, $zero, 3112 add.d $a0, $sp, $a0 - xvld $xr0, $a0, 0 - xvsrli.w $xr1, $xr2, 8 - xvslli.w $xr2, $xr2, 24 - xvor.v $xr1, $xr2, $xr1 - xvsrli.w $xr2, $xr0, 8 - xvslli.w $xr0, $xr0, 24 - xvor.v $xr0, $xr0, $xr2 + xvld $xr2, $a0, 0 + xvrotri.w $xr0, $xr0, 8 + xvst $xr0, $sp, 2024 + xvrotri.w $xr0, $xr1, 8 + xvrotri.w $xr1, $xr2, 8 ori $a0, $zero, 3144 add.d $a0, $sp, $a0 xvld $xr2, $a0, 0 ori $a0, $zero, 2056 add.d $a0, $sp, $a0 - xvst $xr1, $a0, 0 + xvst $xr0, $a0, 0 ori $a0, $zero, 2088 add.d $a0, $sp, $a0 - xvst $xr0, $a0, 0 + xvst $xr1, $a0, 0 ori $a0, $zero, 3176 add.d $a0, $sp, $a0 xvld $xr0, $a0, 0 - xvsrli.w $xr1, $xr2, 8 - xvslli.w $xr2, $xr2, 24 - xvor.v $xr1, $xr2, $xr1 - xvsrli.w $xr2, $xr0, 8 - xvslli.w $xr0, $xr0, 24 - xvor.v $xr0, $xr0, $xr2 - ori $a0, $zero, 3208 - add.d $a0, $sp, $a0 - xvld $xr2, $a0, 0 + xvrotri.w $xr1, $xr2, 8 ori $a0, $zero, 2120 add.d $a0, $sp, $a0 xvst $xr1, $a0, 0 - ori $a0, $zero, 2152 + ori $a0, $zero, 3208 add.d $a0, $sp, $a0 - xvst $xr0, $a0, 0 + xvld $xr1, $a0, 0 ori $a0, $zero, 3240 add.d $a0, $sp, $a0 - xvld $xr0, $a0, 0 - xvsrli.w $xr1, $xr2, 8 - xvslli.w $xr2, $xr2, 24 - xvor.v $xr1, $xr2, $xr1 - xvsrli.w $xr2, $xr0, 8 - xvslli.w $xr0, $xr0, 24 - xvor.v $xr0, $xr0, $xr2 + xvld $xr2, $a0, 0 + xvrotri.w $xr0, $xr0, 8 + ori $a0, $zero, 2152 + add.d $a0, $sp, $a0 + xvst $xr0, $a0, 0 + xvrotri.w $xr0, $xr1, 8 + xvrotri.w $xr1, $xr2, 8 ori $a0, $zero, 3272 add.d $a0, $sp, $a0 xvld $xr2, $a0, 0 ori $a0, $zero, 2184 add.d $a0, $sp, $a0 - xvst $xr1, $a0, 0 + xvst $xr0, $a0, 0 ori $a0, $zero, 2216 add.d $a0, $sp, $a0 - xvst $xr0, $a0, 0 + xvst $xr1, $a0, 0 ori $a0, $zero, 3304 add.d $a0, $sp, $a0 xvld $xr0, $a0, 0 - xvsrli.w $xr1, $xr2, 8 - xvslli.w $xr2, $xr2, 24 - xvor.v $xr1, $xr2, $xr1 - xvsrli.w $xr2, $xr0, 8 - xvslli.w $xr0, $xr0, 24 - xvor.v $xr0, $xr0, $xr2 - ori $a0, $zero, 3336 - add.d $a0, $sp, $a0 - xvld $xr2, $a0, 0 + xvrotri.w $xr1, $xr2, 8 ori $a0, $zero, 2248 add.d $a0, $sp, $a0 xvst $xr1, $a0, 0 - ori $a0, $zero, 2280 + ori $a0, $zero, 3336 add.d $a0, $sp, $a0 - xvst $xr0, $a0, 0 + xvld $xr1, $a0, 0 ori $a0, $zero, 3368 add.d $a0, $sp, $a0 - xvld $xr0, $a0, 0 - xvsrli.w $xr1, $xr2, 8 - xvslli.w $xr2, $xr2, 24 - xvor.v $xr1, $xr2, $xr1 - xvsrli.w $xr2, $xr0, 8 - xvslli.w $xr0, $xr0, 24 - xvor.v $xr0, $xr0, $xr2 + xvld $xr2, $a0, 0 + xvrotri.w $xr0, $xr0, 8 + ori $a0, $zero, 2280 + add.d $a0, $sp, $a0 + xvst $xr0, $a0, 0 + xvrotri.w $xr0, $xr1, 8 + xvrotri.w $xr1, $xr2, 8 ori $a0, $zero, 3400 add.d $a0, $sp, $a0 xvld $xr2, $a0, 0 ori $a0, $zero, 2312 add.d $a0, $sp, $a0 - xvst $xr1, $a0, 0 + xvst $xr0, $a0, 0 ori $a0, $zero, 2344 add.d $a0, $sp, $a0 - xvst $xr0, $a0, 0 + xvst $xr1, $a0, 0 ori $a0, $zero, 3432 add.d $a0, $sp, $a0 xvld $xr0, $a0, 0 - xvsrli.w $xr1, $xr2, 8 - xvslli.w $xr2, $xr2, 24 - xvor.v $xr1, $xr2, $xr1 - xvsrli.w $xr2, $xr0, 8 - xvslli.w $xr0, $xr0, 24 - xvor.v $xr0, $xr0, $xr2 - ori $a0, $zero, 3464 - add.d $a0, $sp, $a0 - xvld $xr2, $a0, 0 + xvrotri.w $xr1, $xr2, 8 ori $a0, $zero, 2376 add.d $a0, $sp, $a0 xvst $xr1, $a0, 0 - ori $a0, $zero, 2408 + ori $a0, $zero, 3464 add.d $a0, $sp, $a0 - xvst $xr0, $a0, 0 + xvld $xr1, $a0, 0 ori $a0, $zero, 3496 add.d $a0, $sp, $a0 - xvld $xr0, $a0, 0 - xvsrli.w $xr1, $xr2, 8 - xvslli.w $xr2, $xr2, 24 - xvor.v $xr1, $xr2, $xr1 - xvsrli.w $xr2, $xr0, 8 - xvslli.w $xr0, $xr0, 24 - xvor.v $xr0, $xr0, $xr2 + xvld $xr2, $a0, 0 + xvrotri.w $xr0, $xr0, 8 + ori $a0, $zero, 2408 + add.d $a0, $sp, $a0 + xvst $xr0, $a0, 0 + xvrotri.w $xr0, $xr1, 8 + xvrotri.w $xr1, $xr2, 8 ori $a0, $zero, 3528 add.d $a0, $sp, $a0 xvld $xr2, $a0, 0 ori $a0, $zero, 2440 add.d $a0, $sp, $a0 - xvst $xr1, $a0, 0 + xvst $xr0, $a0, 0 ori $a0, $zero, 2472 add.d $a0, $sp, $a0 - xvst $xr0, $a0, 0 + xvst $xr1, $a0, 0 ori $a0, $zero, 3560 add.d $a0, $sp, $a0 xvld $xr0, $a0, 0 - xvsrli.w $xr1, $xr2, 8 - xvslli.w $xr2, $xr2, 24 - xvor.v $xr1, $xr2, $xr1 - xvsrli.w $xr2, $xr0, 8 - xvslli.w $xr0, $xr0, 24 - xvor.v $xr0, $xr0, $xr2 - ori $a0, $zero, 3592 - add.d $a0, $sp, $a0 - xvld $xr2, $a0, 0 + xvrotri.w $xr1, $xr2, 8 ori $a0, $zero, 2504 add.d $a0, $sp, $a0 xvst $xr1, $a0, 0 - ori $a0, $zero, 2536 + ori $a0, $zero, 3592 add.d $a0, $sp, $a0 - xvst $xr0, $a0, 0 + xvld $xr1, $a0, 0 ori $a0, $zero, 3624 add.d $a0, $sp, $a0 - xvld $xr0, $a0, 0 - xvsrli.w $xr1, $xr2, 8 - xvslli.w $xr2, $xr2, 24 - xvor.v $xr1, $xr2, $xr1 - xvsrli.w $xr2, $xr0, 8 - xvslli.w $xr0, $xr0, 24 - xvor.v $xr0, $xr0, $xr2 + xvld $xr2, $a0, 0 + xvrotri.w $xr0, $xr0, 8 + ori $a0, $zero, 2536 + add.d $a0, $sp, $a0 + xvst $xr0, $a0, 0 + xvrotri.w $xr0, $xr1, 8 + xvrotri.w $xr1, $xr2, 8 xvld $xr2, $sp, 1096 ori $a0, $zero, 2568 add.d $a0, $sp, $a0 - xvst $xr1, $a0, 0 + xvst $xr0, $a0, 0 ori $a0, $zero, 2600 add.d $a0, $sp, $a0 - xvst $xr0, $a0, 0 + xvst $xr1, $a0, 0 xvld $xr0, $sp, 1128 - xvsrli.h $xr1, $xr2, 9 - xvslli.h $xr2, $xr2, 7 - xvor.v $xr1, $xr2, $xr1 - xvsrli.h $xr2, $xr0, 9 - xvslli.h $xr0, $xr0, 7 - xvor.v $xr0, $xr0, $xr2 - xvld $xr2, $sp, 1160 + xvrotri.h $xr1, $xr2, 9 xvst $xr1, $sp, 584 + xvld $xr1, $sp, 1160 + xvld $xr2, $sp, 1192 + xvrotri.h $xr0, $xr0, 9 xvst $xr0, $sp, 616 - xvld $xr0, $sp, 1192 - xvsrli.h $xr1, $xr2, 9 - xvslli.h $xr2, $xr2, 7 - xvor.v $xr1, $xr2, $xr1 - xvsrli.h $xr2, $xr0, 9 - xvslli.h $xr0, $xr0, 7 - xvor.v $xr0, $xr0, $xr2 + xvrotri.h $xr0, $xr1, 9 + xvrotri.h $xr1, $xr2, 9 xvld $xr2, $sp, 1224 - xvst $xr1, $sp, 648 - xvst $xr0, $sp, 680 + xvst $xr0, $sp, 648 + xvst $xr1, $sp, 680 xvld $xr0, $sp, 1256 - xvsrli.h $xr1, $xr2, 9 - xvslli.h $xr2, $xr2, 7 - xvor.v $xr1, $xr2, $xr1 - xvsrli.h $xr2, $xr0, 9 - xvslli.h $xr0, $xr0, 7 - xvor.v $xr0, $xr0, $xr2 - xvld $xr2, $sp, 1288 + xvrotri.h $xr1, $xr2, 9 xvst $xr1, $sp, 712 + xvld $xr1, $sp, 1288 + xvld $xr2, $sp, 1320 + xvrotri.h $xr0, $xr0, 9 xvst $xr0, $sp, 744 - xvld $xr0, $sp, 1320 - xvsrli.h $xr1, $xr2, 9 - xvslli.h $xr2, $xr2, 7 - xvor.v $xr1, $xr2, $xr1 - xvsrli.h $xr2, $xr0, 9 - xvslli.h $xr0, $xr0, 7 - xvor.v $xr0, $xr0, $xr2 + xvrotri.h $xr0, $xr1, 9 + xvrotri.h $xr1, $xr2, 9 xvld $xr2, $sp, 1352 - xvst $xr1, $sp, 776 - xvst $xr0, $sp, 808 + xvst $xr0, $sp, 776 + xvst $xr1, $sp, 808 xvld $xr0, $sp, 1384 - xvsrli.h $xr1, $xr2, 9 - xvslli.h $xr2, $xr2, 7 - xvor.v $xr1, $xr2, $xr1 - xvsrli.h $xr2, $xr0, 9 - xvslli.h $xr0, $xr0, 7 - xvor.v $xr0, $xr0, $xr2 - xvld $xr2, $sp, 1416 + xvrotri.h $xr1, $xr2, 9 xvst $xr1, $sp, 840 + xvld $xr1, $sp, 1416 + xvld $xr2, $sp, 1448 + xvrotri.h $xr0, $xr0, 9 xvst $xr0, $sp, 872 - xvld $xr0, $sp, 1448 - xvsrli.h $xr1, $xr2, 9 - xvslli.h $xr2, $xr2, 7 - xvor.v $xr1, $xr2, $xr1 - xvsrli.h $xr2, $xr0, 9 - xvslli.h $xr0, $xr0, 7 - xvor.v $xr0, $xr0, $xr2 + xvrotri.h $xr0, $xr1, 9 + xvrotri.h $xr1, $xr2, 9 xvld $xr2, $sp, 1480 - xvst $xr1, $sp, 904 - xvst $xr0, $sp, 936 + xvst $xr0, $sp, 904 + xvst $xr1, $sp, 936 xvld $xr0, $sp, 1512 - xvsrli.h $xr1, $xr2, 9 - xvslli.h $xr2, $xr2, 7 - xvor.v $xr1, $xr2, $xr1 - xvsrli.h $xr2, $xr0, 9 - xvslli.h $xr0, $xr0, 7 - xvor.v $xr0, $xr0, $xr2 - xvld $xr2, $sp, 1544 + xvrotri.h $xr1, $xr2, 9 xvst $xr1, $sp, 968 + xvld $xr1, $sp, 1544 + xvld $xr2, $sp, 1576 + xvrotri.h $xr0, $xr0, 9 xvst $xr0, $sp, 1000 - xvld $xr0, $sp, 1576 - xvsrli.h $xr1, $xr2, 9 - xvslli.h $xr2, $xr2, 7 - xvor.v $xr1, $xr2, $xr1 - xvsrli.h $xr2, $xr0, 9 - xvslli.h $xr0, $xr0, 7 - xvor.v $xr0, $xr0, $xr2 + xvrotri.h $xr0, $xr1, 9 + xvrotri.h $xr1, $xr2, 9 xvld $xr2, $sp, 328 - xvst $xr1, $sp, 1032 - xvst $xr0, $sp, 1064 + xvst $xr0, $sp, 1032 + xvst $xr1, $sp, 1064 xvld $xr0, $sp, 360 - xvsrli.b $xr1, $xr2, 5 - xvslli.b $xr2, $xr2, 3 - xvor.v $xr1, $xr2, $xr1 - xvsrli.b $xr2, $xr0, 5 - xvslli.b $xr0, $xr0, 3 - xvor.v $xr0, $xr0, $xr2 - xvld $xr2, $sp, 392 + xvrotri.b $xr1, $xr2, 5 xvst $xr1, $sp, 72 + xvld $xr1, $sp, 392 + xvld $xr2, $sp, 424 + xvrotri.b $xr0, $xr0, 5 xvst $xr0, $sp, 104 - xvld $xr0, $sp, 424 - xvsrli.b $xr1, $xr2, 5 - xvslli.b $xr2, $xr2, 3 - xvor.v $xr1, $xr2, $xr1 - xvsrli.b $xr2, $xr0, 5 - xvslli.b $xr0, $xr0, 3 - xvor.v $xr0, $xr0, $xr2 + xvrotri.b $xr0, $xr1, 5 + xvrotri.b $xr1, $xr2, 5 xvld $xr2, $sp, 456 - xvst $xr1, $sp, 136 - xvst $xr0, $sp, 168 + xvst $xr0, $sp, 136 + xvst $xr1, $sp, 168 xvld $xr0, $sp, 488 - xvsrli.b $xr1, $xr2, 5 - xvslli.b $xr2, $xr2, 3 - xvor.v $xr1, $xr2, $xr1 - xvsrli.b $xr2, $xr0, 5 - xvslli.b $xr0, $xr0, 3 - xvor.v $xr0, $xr0, $xr2 - xvld $xr2, $sp, 520 + xvrotri.b $xr1, $xr2, 5 xvst $xr1, $sp, 200 + xvld $xr1, $sp, 520 + xvld $xr2, $sp, 552 + xvrotri.b $xr0, $xr0, 5 xvst $xr0, $sp, 232 - xvld $xr0, $sp, 552 - xvsrli.b $xr1, $xr2, 5 - xvslli.b $xr2, $xr2, 3 - xvor.v $xr1, $xr2, $xr1 - xvsrli.b $xr2, $xr0, 5 - xvslli.b $xr0, $xr0, 3 - xvor.v $xr0, $xr0, $xr2 - xvst $xr1, $sp, 264 - xvst $xr0, $sp, 296 + xvrotri.b $xr0, $xr1, 5 + xvrotri.b $xr1, $xr2, 5 + xvst $xr0, $sp, 264 + xvst $xr1, $sp, 296 ori $a0, $zero, 3656 add.d $a0, $sp, $a0 st.d $a0, $sp, 32