@@ -602,11 +602,11 @@ body: |
602
602
; GCN: [[COPY:%[0-9]+]]:vgpr_32 = COPY $vgpr0
603
603
; GCN: [[V_MOV_B32_e32_:%[0-9]+]]:vgpr_32 = V_MOV_B32_e32 518144, implicit $exec
604
604
; GCN: [[V_ADD_I32_e32_:%[0-9]+]]:vgpr_32 = V_ADD_I32_e32 killed [[V_MOV_B32_e32_]], [[COPY]], implicit-def $vcc, implicit $exec
605
- ; GCN: S_ENDPGM 0, implicit [[V_ADD_I32_e32_]]
605
+ ; GCN: S_ENDPGM implicit [[V_ADD_I32_e32_]]
606
606
%0:vgpr_32 = COPY $vgpr0
607
607
%1:vgpr_32 = V_MOV_B32_e32 518144, implicit $exec
608
- %2:vgpr_32, %3:sreg_64_xexec = V_ADD_I32_e64 killed %1, %0, 0, implicit $exec
609
- S_ENDPGM 0, implicit %2
608
+ %2:vgpr_32, %3:sreg_64_xexec = V_ADD_I32_e64 killed %1, %0, implicit $exec
609
+ S_ENDPGM implicit %2
610
610
...
611
611
612
612
---
@@ -620,11 +620,11 @@ body: |
620
620
; GCN: [[COPY:%[0-9]+]]:vgpr_32 = COPY $vgpr0
621
621
; GCN: [[V_MOV_B32_e32_:%[0-9]+]]:vgpr_32 = V_MOV_B32_e32 518144, implicit $exec
622
622
; GCN: [[V_ADD_I32_e32_:%[0-9]+]]:vgpr_32 = V_ADD_I32_e32 [[V_MOV_B32_e32_]], killed [[COPY]], implicit-def $vcc, implicit $exec
623
- ; GCN: S_ENDPGM 0, implicit [[V_ADD_I32_e32_]]
623
+ ; GCN: S_ENDPGM implicit [[V_ADD_I32_e32_]]
624
624
%0:vgpr_32 = COPY $vgpr0
625
625
%1:vgpr_32 = V_MOV_B32_e32 518144, implicit $exec
626
- %2:vgpr_32, %3:sreg_64_xexec = V_ADD_I32_e64 %1, killed %0, 0, implicit $exec
627
- S_ENDPGM 0, implicit %2
626
+ %2:vgpr_32, %3:sreg_64_xexec = V_ADD_I32_e64 %1, killed %0, implicit $exec
627
+ S_ENDPGM implicit %2
628
628
...
629
629
630
630
---
@@ -638,11 +638,11 @@ body: |
638
638
; GCN: [[COPY:%[0-9]+]]:vgpr_32 = COPY $vgpr0
639
639
; GCN: [[V_MOV_B32_e32_:%[0-9]+]]:vgpr_32 = V_MOV_B32_e32 518144, implicit $exec
640
640
; GCN: [[COPY1:%[0-9]+]]:sreg_64_xexec = COPY $vcc
641
- ; GCN: [[V_ADDC_U32_e64_:%[0-9]+]]:vgpr_32, [[V_ADDC_U32_e64_1:%[0-9]+]]:sreg_64_xexec = V_ADDC_U32_e64 [[V_MOV_B32_e32_]], [[COPY]], [[COPY1]], 0, implicit $exec
642
- ; GCN: S_ENDPGM 0, implicit [[V_ADDC_U32_e64_]]
641
+ ; GCN: [[V_ADDC_U32_e64_:%[0-9]+]]:vgpr_32, [[V_ADDC_U32_e64_1:%[0-9]+]]:sreg_64_xexec = V_ADDC_U32_e64 [[V_MOV_B32_e32_]], [[COPY]], [[COPY1]], implicit $exec
642
+ ; GCN: S_ENDPGM implicit [[V_ADDC_U32_e64_]]
643
643
%0:vgpr_32 = COPY $vgpr0
644
644
%1:vgpr_32 = V_MOV_B32_e32 518144, implicit $exec
645
645
%2:sreg_64_xexec = COPY $vcc
646
- %3:vgpr_32, %4:sreg_64_xexec = V_ADDC_U32_e64 %1, %0, %2, 0, implicit $exec
647
- S_ENDPGM 0, implicit %3
646
+ %3:vgpr_32, %4:sreg_64_xexec = V_ADDC_U32_e64 %1, %0, %2, implicit $exec
647
+ S_ENDPGM implicit %3
648
648
...
0 commit comments