@@ -1052,31 +1052,15 @@ define <4 x float> @PR34724_add_v4f32_0u23(<4 x float> %0, <4 x float> %1) {
10521052}
10531053
10541054define <4 x float > @PR34724_add_v4f32_01u3 (<4 x float > %0 , <4 x float > %1 ) {
1055- ; SSE-SLOW-LABEL: PR34724_add_v4f32_01u3:
1056- ; SSE-SLOW: # %bb.0:
1057- ; SSE-SLOW-NEXT: haddps %xmm0, %xmm0
1058- ; SSE-SLOW-NEXT: movsldup {{.*#+}} xmm2 = xmm1[0,0,2,2]
1059- ; SSE-SLOW-NEXT: addps %xmm1, %xmm2
1060- ; SSE-SLOW-NEXT: shufps {{.*#+}} xmm0 = xmm0[0,1],xmm2[2,3]
1061- ; SSE-SLOW-NEXT: retq
1062- ;
1063- ; SSE-FAST-LABEL: PR34724_add_v4f32_01u3:
1064- ; SSE-FAST: # %bb.0:
1065- ; SSE-FAST-NEXT: haddps %xmm1, %xmm0
1066- ; SSE-FAST-NEXT: retq
1067- ;
1068- ; AVX-SLOW-LABEL: PR34724_add_v4f32_01u3:
1069- ; AVX-SLOW: # %bb.0:
1070- ; AVX-SLOW-NEXT: vhaddps %xmm0, %xmm0, %xmm0
1071- ; AVX-SLOW-NEXT: vmovsldup {{.*#+}} xmm2 = xmm1[0,0,2,2]
1072- ; AVX-SLOW-NEXT: vaddps %xmm1, %xmm2, %xmm1
1073- ; AVX-SLOW-NEXT: vblendps {{.*#+}} xmm0 = xmm0[0,1],xmm1[2,3]
1074- ; AVX-SLOW-NEXT: retq
1055+ ; SSE-LABEL: PR34724_add_v4f32_01u3:
1056+ ; SSE: # %bb.0:
1057+ ; SSE-NEXT: haddps %xmm1, %xmm0
1058+ ; SSE-NEXT: retq
10751059;
1076- ; AVX-FAST- LABEL: PR34724_add_v4f32_01u3:
1077- ; AVX-FAST : # %bb.0:
1078- ; AVX-FAST- NEXT: vhaddps %xmm1, %xmm0, %xmm0
1079- ; AVX-FAST- NEXT: retq
1060+ ; AVX-LABEL: PR34724_add_v4f32_01u3:
1061+ ; AVX: # %bb.0:
1062+ ; AVX-NEXT: vhaddps %xmm1, %xmm0, %xmm0
1063+ ; AVX-NEXT: retq
10801064 %3 = shufflevector <4 x float > %0 , <4 x float > undef , <2 x i32 > <i32 0 , i32 2 >
10811065 %4 = shufflevector <4 x float > %0 , <4 x float > undef , <2 x i32 > <i32 1 , i32 3 >
10821066 %5 = fadd <2 x float > %3 , %4
@@ -1088,31 +1072,15 @@ define <4 x float> @PR34724_add_v4f32_01u3(<4 x float> %0, <4 x float> %1) {
10881072}
10891073
10901074define <4 x float > @PR34724_add_v4f32_012u (<4 x float > %0 , <4 x float > %1 ) {
1091- ; SSE-SLOW-LABEL: PR34724_add_v4f32_012u:
1092- ; SSE-SLOW: # %bb.0:
1093- ; SSE-SLOW-NEXT: haddps %xmm0, %xmm0
1094- ; SSE-SLOW-NEXT: movshdup {{.*#+}} xmm2 = xmm1[1,1,3,3]
1095- ; SSE-SLOW-NEXT: addps %xmm1, %xmm2
1096- ; SSE-SLOW-NEXT: movlhps {{.*#+}} xmm0 = xmm0[0],xmm2[0]
1097- ; SSE-SLOW-NEXT: retq
1098- ;
1099- ; SSE-FAST-LABEL: PR34724_add_v4f32_012u:
1100- ; SSE-FAST: # %bb.0:
1101- ; SSE-FAST-NEXT: haddps %xmm1, %xmm0
1102- ; SSE-FAST-NEXT: retq
1103- ;
1104- ; AVX-SLOW-LABEL: PR34724_add_v4f32_012u:
1105- ; AVX-SLOW: # %bb.0:
1106- ; AVX-SLOW-NEXT: vhaddps %xmm0, %xmm0, %xmm0
1107- ; AVX-SLOW-NEXT: vmovshdup {{.*#+}} xmm2 = xmm1[1,1,3,3]
1108- ; AVX-SLOW-NEXT: vaddps %xmm1, %xmm2, %xmm1
1109- ; AVX-SLOW-NEXT: vmovlhps {{.*#+}} xmm0 = xmm0[0],xmm1[0]
1110- ; AVX-SLOW-NEXT: retq
1075+ ; SSE-LABEL: PR34724_add_v4f32_012u:
1076+ ; SSE: # %bb.0:
1077+ ; SSE-NEXT: haddps %xmm1, %xmm0
1078+ ; SSE-NEXT: retq
11111079;
1112- ; AVX-FAST- LABEL: PR34724_add_v4f32_012u:
1113- ; AVX-FAST : # %bb.0:
1114- ; AVX-FAST- NEXT: vhaddps %xmm1, %xmm0, %xmm0
1115- ; AVX-FAST- NEXT: retq
1080+ ; AVX-LABEL: PR34724_add_v4f32_012u:
1081+ ; AVX: # %bb.0:
1082+ ; AVX-NEXT: vhaddps %xmm1, %xmm0, %xmm0
1083+ ; AVX-NEXT: retq
11161084 %3 = shufflevector <4 x float > %0 , <4 x float > undef , <2 x i32 > <i32 0 , i32 2 >
11171085 %4 = shufflevector <4 x float > %0 , <4 x float > undef , <2 x i32 > <i32 1 , i32 3 >
11181086 %5 = fadd <2 x float > %3 , %4
0 commit comments