-
Notifications
You must be signed in to change notification settings - Fork 15.2k
[X86] combinePMULH - combine mulhu + srl
#132548
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from 3 commits
df60a87
94472a1
c6b4c64
d380d8a
c0aa0cd
d611c2a
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -54021,7 +54021,7 @@ static SDValue combineTruncatedArithmetic(SDNode *N, SelectionDAG &DAG, | |
| } | ||
|
|
||
| // Try to form a MULHU or MULHS node by looking for | ||
| // (trunc (srl (mul ext, ext), 16)) | ||
| // (trunc (srl (mul ext, ext), >= 16)) | ||
| // TODO: This is X86 specific because we want to be able to handle wide types | ||
| // before type legalization. But we can only do it if the vector will be | ||
| // legalized via widening/splitting. Type legalization can't handle promotion | ||
|
|
@@ -54046,10 +54046,16 @@ static SDValue combinePMULH(SDValue Src, EVT VT, const SDLoc &DL, | |
|
|
||
| // First instruction should be a right shift by 16 of a multiply. | ||
| SDValue LHS, RHS; | ||
| APInt ShiftAmt; | ||
| if (!sd_match(Src, | ||
| m_Srl(m_Mul(m_Value(LHS), m_Value(RHS)), m_SpecificInt(16)))) | ||
| m_Srl(m_Mul(m_Value(LHS), m_Value(RHS)), m_ConstInt(ShiftAmt)))) | ||
| return SDValue(); | ||
|
|
||
| if (ShiftAmt.ult(16)) | ||
| return SDValue(); | ||
|
|
||
| APInt AdditionalShift = (ShiftAmt - 16).trunc(16); | ||
|
|
||
|
Collaborator
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. You need to do some testing to see what happens with MULHS - probably limit AdditionalShift != 0 to just the IsUnsigned case for starters? |
||
| // Count leading sign/zero bits on both inputs - if there are enough then | ||
| // truncation back to vXi16 will be cheap - either as a pack/shuffle | ||
| // sequence or using AVX512 truncations. If the inputs are sext/zext then the | ||
|
|
@@ -54087,15 +54093,19 @@ static SDValue combinePMULH(SDValue Src, EVT VT, const SDLoc &DL, | |
| InVT.getSizeInBits() / 16); | ||
| SDValue Res = DAG.getNode(ISD::MULHU, DL, BCVT, DAG.getBitcast(BCVT, LHS), | ||
| DAG.getBitcast(BCVT, RHS)); | ||
| return DAG.getNode(ISD::TRUNCATE, DL, VT, DAG.getBitcast(InVT, Res)); | ||
| Res = DAG.getNode(ISD::TRUNCATE, DL, VT, DAG.getBitcast(InVT, Res)); | ||
abhishek-kaushik22 marked this conversation as resolved.
Show resolved
Hide resolved
|
||
| return DAG.getNode(ISD::SRL, DL, VT, Res, | ||
| DAG.getShiftAmountConstant(AdditionalShift, VT, DL)); | ||
abhishek-kaushik22 marked this conversation as resolved.
Outdated
Show resolved
Hide resolved
|
||
| } | ||
|
|
||
| // Truncate back to source type. | ||
| LHS = DAG.getNode(ISD::TRUNCATE, DL, VT, LHS); | ||
| RHS = DAG.getNode(ISD::TRUNCATE, DL, VT, RHS); | ||
|
|
||
| unsigned Opc = IsSigned ? ISD::MULHS : ISD::MULHU; | ||
| return DAG.getNode(Opc, DL, VT, LHS, RHS); | ||
| SDValue Res = DAG.getNode(Opc, DL, VT, LHS, RHS); | ||
| return DAG.getNode(ISD::SRL, DL, VT, Res, | ||
| DAG.getShiftAmountConstant(AdditionalShift, VT, DL)); | ||
|
||
| } | ||
|
|
||
| // Attempt to match PMADDUBSW, which multiplies corresponding unsigned bytes | ||
|
|
||
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -2166,3 +2166,260 @@ define <8 x i16> @sse2_pmulhu_w_const(<8 x i16> %a0, <8 x i16> %a1) { | |
| } | ||
| declare <8 x i16> @llvm.x86.sse2.pmulhu.w(<8 x i16>, <8 x i16>) | ||
|
|
||
| define <8 x i16> @zext_mul_and_shift17(<8 x i16> %a, <8 x i16> %b) { | ||
| ; SSE-LABEL: zext_mul_and_shift17: | ||
| ; SSE: # %bb.0: | ||
| ; SSE-NEXT: pmulhuw %xmm1, %xmm0 | ||
| ; SSE-NEXT: psrlw $1, %xmm0 | ||
| ; SSE-NEXT: retq | ||
| ; | ||
| ; AVX-LABEL: zext_mul_and_shift17: | ||
| ; AVX: # %bb.0: | ||
| ; AVX-NEXT: vpmulhuw %xmm1, %xmm0, %xmm0 | ||
| ; AVX-NEXT: vpsrlw $1, %xmm0, %xmm0 | ||
| ; AVX-NEXT: retq | ||
| %a.ext = zext <8 x i16> %a to <8 x i32> | ||
| %b.ext = zext <8 x i16> %b to <8 x i32> | ||
| %mul = mul <8 x i32> %a.ext, %b.ext | ||
| %shift = lshr <8 x i32> %mul, splat(i32 17) | ||
| %trunc = trunc <8 x i32> %shift to <8 x i16> | ||
| ret <8 x i16> %trunc | ||
| } | ||
|
Collaborator
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. It'd be better if we tested a range of shift values as well as the shift by 17 from the original bug report - shifts by 24 and 31 will be interesting for instance. |
||
|
|
||
| define <8 x i16> @sext_mul_and_shift17(<8 x i16> %a, <8 x i16> %b) { | ||
| ; SSE-LABEL: sext_mul_and_shift17: | ||
| ; SSE: # %bb.0: | ||
| ; SSE-NEXT: pmulhw %xmm1, %xmm0 | ||
| ; SSE-NEXT: psrlw $1, %xmm0 | ||
| ; SSE-NEXT: retq | ||
| ; | ||
| ; AVX-LABEL: sext_mul_and_shift17: | ||
| ; AVX: # %bb.0: | ||
| ; AVX-NEXT: vpmulhw %xmm1, %xmm0, %xmm0 | ||
| ; AVX-NEXT: vpsrlw $1, %xmm0, %xmm0 | ||
| ; AVX-NEXT: retq | ||
| %a.ext = sext <8 x i16> %a to <8 x i32> | ||
| %b.ext = sext <8 x i16> %b to <8 x i32> | ||
| %mul = mul <8 x i32> %a.ext, %b.ext | ||
| %shift = lshr <8 x i32> %mul, splat(i32 17) | ||
| %trunc = trunc <8 x i32> %shift to <8 x i16> | ||
| ret <8 x i16> %trunc | ||
| } | ||
|
|
||
| define <4 x i16> @sext_mulhw_v4i16_shift17(<4 x i16> %a, <4 x i16> %b) { | ||
| ; SSE-LABEL: sext_mulhw_v4i16_shift17: | ||
| ; SSE: # %bb.0: | ||
| ; SSE-NEXT: pmulhw %xmm1, %xmm0 | ||
| ; SSE-NEXT: psrlw $1, %xmm0 | ||
| ; SSE-NEXT: retq | ||
| ; | ||
| ; AVX-LABEL: sext_mulhw_v4i16_shift17: | ||
| ; AVX: # %bb.0: | ||
| ; AVX-NEXT: vpmulhw %xmm1, %xmm0, %xmm0 | ||
| ; AVX-NEXT: vpsrlw $1, %xmm0, %xmm0 | ||
| ; AVX-NEXT: retq | ||
| %a1 = sext <4 x i16> %a to <4 x i32> | ||
| %b1 = sext <4 x i16> %b to <4 x i32> | ||
| %c = mul <4 x i32> %a1, %b1 | ||
| %d = lshr <4 x i32> %c, splat (i32 17) | ||
| %e = trunc <4 x i32> %d to <4 x i16> | ||
| ret <4 x i16> %e | ||
| } | ||
|
|
||
| define <4 x i16> @and_mulhuw_v4i16_shift17(<4 x i64> %a, <4 x i64> %b) { | ||
| ; SSE2-LABEL: and_mulhuw_v4i16_shift17: | ||
| ; SSE2: # %bb.0: | ||
| ; SSE2-NEXT: shufps {{.*#+}} xmm2 = xmm2[0,2],xmm3[0,2] | ||
| ; SSE2-NEXT: pslld $16, %xmm2 | ||
| ; SSE2-NEXT: psrad $16, %xmm2 | ||
| ; SSE2-NEXT: xorps %xmm3, %xmm3 | ||
| ; SSE2-NEXT: packssdw %xmm3, %xmm2 | ||
| ; SSE2-NEXT: shufps {{.*#+}} xmm0 = xmm0[0,2],xmm1[0,2] | ||
| ; SSE2-NEXT: pslld $16, %xmm0 | ||
| ; SSE2-NEXT: psrad $16, %xmm0 | ||
| ; SSE2-NEXT: packssdw %xmm3, %xmm0 | ||
| ; SSE2-NEXT: pmulhuw %xmm2, %xmm0 | ||
| ; SSE2-NEXT: psrlw $1, %xmm0 | ||
| ; SSE2-NEXT: retq | ||
| ; | ||
| ; SSE41-LABEL: and_mulhuw_v4i16_shift17: | ||
| ; SSE41: # %bb.0: | ||
| ; SSE41-NEXT: pxor %xmm4, %xmm4 | ||
| ; SSE41-NEXT: pblendw {{.*#+}} xmm3 = xmm3[0],xmm4[1,2,3],xmm3[4],xmm4[5,6,7] | ||
| ; SSE41-NEXT: pblendw {{.*#+}} xmm2 = xmm2[0],xmm4[1,2,3],xmm2[4],xmm4[5,6,7] | ||
| ; SSE41-NEXT: packusdw %xmm3, %xmm2 | ||
| ; SSE41-NEXT: packusdw %xmm4, %xmm2 | ||
| ; SSE41-NEXT: pblendw {{.*#+}} xmm1 = xmm1[0],xmm4[1,2,3],xmm1[4],xmm4[5,6,7] | ||
| ; SSE41-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0],xmm4[1,2,3],xmm0[4],xmm4[5,6,7] | ||
| ; SSE41-NEXT: packusdw %xmm1, %xmm0 | ||
| ; SSE41-NEXT: packusdw %xmm4, %xmm0 | ||
| ; SSE41-NEXT: pmulhuw %xmm2, %xmm0 | ||
| ; SSE41-NEXT: psrlw $1, %xmm0 | ||
| ; SSE41-NEXT: retq | ||
| ; | ||
| ; AVX2-LABEL: and_mulhuw_v4i16_shift17: | ||
| ; AVX2: # %bb.0: | ||
| ; AVX2-NEXT: vpmulhuw %ymm1, %ymm0, %ymm0 | ||
| ; AVX2-NEXT: vpxor %xmm1, %xmm1, %xmm1 | ||
| ; AVX2-NEXT: vpblendw {{.*#+}} ymm0 = ymm0[0],ymm1[1,2,3],ymm0[4],ymm1[5,6,7],ymm0[8],ymm1[9,10,11],ymm0[12],ymm1[13,14,15] | ||
| ; AVX2-NEXT: vextracti128 $1, %ymm0, %xmm1 | ||
| ; AVX2-NEXT: vpackusdw %xmm1, %xmm0, %xmm0 | ||
| ; AVX2-NEXT: vpackusdw %xmm0, %xmm0, %xmm0 | ||
| ; AVX2-NEXT: vpsrlw $1, %xmm0, %xmm0 | ||
| ; AVX2-NEXT: vzeroupper | ||
| ; AVX2-NEXT: retq | ||
| ; | ||
| ; AVX512-LABEL: and_mulhuw_v4i16_shift17: | ||
| ; AVX512: # %bb.0: | ||
| ; AVX512-NEXT: vpmulhuw %ymm1, %ymm0, %ymm0 | ||
| ; AVX512-NEXT: vpmovqw %zmm0, %xmm0 | ||
| ; AVX512-NEXT: vpsrlw $1, %xmm0, %xmm0 | ||
| ; AVX512-NEXT: vzeroupper | ||
| ; AVX512-NEXT: retq | ||
| %a1 = and <4 x i64> %a, <i64 65535, i64 65535, i64 65535, i64 65535> | ||
| %b1 = and <4 x i64> %b, <i64 65535, i64 65535, i64 65535, i64 65535> | ||
| %c = mul <4 x i64> %a1, %b1 | ||
| %d = lshr <4 x i64> %c, splat (i64 17) | ||
| %e = trunc <4 x i64> %d to <4 x i16> | ||
| ret <4 x i16> %e | ||
| } | ||
|
|
||
| define <8 x i16> @lshr_mulhuw_v8i16_shift17(<8 x i32> %a, <8 x i32> %b) { | ||
| ; SSE2-LABEL: lshr_mulhuw_v8i16_shift17: | ||
| ; SSE2: # %bb.0: | ||
| ; SSE2-NEXT: psrad $16, %xmm3 | ||
| ; SSE2-NEXT: psrad $16, %xmm2 | ||
| ; SSE2-NEXT: packssdw %xmm3, %xmm2 | ||
| ; SSE2-NEXT: psrad $16, %xmm1 | ||
| ; SSE2-NEXT: psrad $16, %xmm0 | ||
| ; SSE2-NEXT: packssdw %xmm1, %xmm0 | ||
| ; SSE2-NEXT: pmulhuw %xmm2, %xmm0 | ||
| ; SSE2-NEXT: psrlw $1, %xmm0 | ||
| ; SSE2-NEXT: retq | ||
| ; | ||
| ; SSE41-LABEL: lshr_mulhuw_v8i16_shift17: | ||
| ; SSE41: # %bb.0: | ||
| ; SSE41-NEXT: psrld $16, %xmm1 | ||
| ; SSE41-NEXT: psrld $16, %xmm0 | ||
| ; SSE41-NEXT: packusdw %xmm1, %xmm0 | ||
| ; SSE41-NEXT: psrld $16, %xmm3 | ||
| ; SSE41-NEXT: psrld $16, %xmm2 | ||
| ; SSE41-NEXT: packusdw %xmm3, %xmm2 | ||
| ; SSE41-NEXT: pmulhuw %xmm2, %xmm0 | ||
| ; SSE41-NEXT: psrlw $1, %xmm0 | ||
| ; SSE41-NEXT: retq | ||
| ; | ||
| ; AVX2-LABEL: lshr_mulhuw_v8i16_shift17: | ||
| ; AVX2: # %bb.0: | ||
| ; AVX2-NEXT: vpsrld $16, %ymm0, %ymm0 | ||
| ; AVX2-NEXT: vpsrld $16, %ymm1, %ymm1 | ||
| ; AVX2-NEXT: vpmulhuw %ymm1, %ymm0, %ymm0 | ||
| ; AVX2-NEXT: vextracti128 $1, %ymm0, %xmm1 | ||
| ; AVX2-NEXT: vpackusdw %xmm1, %xmm0, %xmm0 | ||
| ; AVX2-NEXT: vpsrlw $1, %xmm0, %xmm0 | ||
| ; AVX2-NEXT: vzeroupper | ||
| ; AVX2-NEXT: retq | ||
| ; | ||
| ; AVX512-LABEL: lshr_mulhuw_v8i16_shift17: | ||
| ; AVX512: # %bb.0: | ||
| ; AVX512-NEXT: vpsrld $16, %ymm0, %ymm0 | ||
| ; AVX512-NEXT: vpsrld $16, %ymm1, %ymm1 | ||
| ; AVX512-NEXT: vpmulhuw %ymm1, %ymm0, %ymm0 | ||
| ; AVX512-NEXT: vpmovdw %zmm0, %ymm0 | ||
| ; AVX512-NEXT: vpsrlw $1, %xmm0, %xmm0 | ||
| ; AVX512-NEXT: vzeroupper | ||
| ; AVX512-NEXT: retq | ||
| %a1 = lshr <8 x i32> %a, <i32 16, i32 16, i32 16, i32 16, i32 16, i32 16, i32 16, i32 16> | ||
| %b1 = lshr <8 x i32> %b, <i32 16, i32 16, i32 16, i32 16, i32 16, i32 16, i32 16, i32 16> | ||
| %c = mul <8 x i32> %a1, %b1 | ||
| %d = lshr <8 x i32> %c, splat (i32 17) | ||
| %e = trunc <8 x i32> %d to <8 x i16> | ||
| ret <8 x i16> %e | ||
| } | ||
|
|
||
| define <16 x i16> @and_mulhuw_v16i16_shift17(<16 x i32> %a, <16 x i32> %b) { | ||
| ; SSE2-LABEL: and_mulhuw_v16i16_shift17: | ||
| ; SSE2: # %bb.0: | ||
| ; SSE2-NEXT: movdqa {{.*#+}} xmm8 = [32767,32767,32767,32767] | ||
| ; SSE2-NEXT: pand %xmm8, %xmm3 | ||
| ; SSE2-NEXT: pand %xmm8, %xmm2 | ||
| ; SSE2-NEXT: packssdw %xmm3, %xmm2 | ||
| ; SSE2-NEXT: pand %xmm8, %xmm1 | ||
| ; SSE2-NEXT: pand %xmm8, %xmm0 | ||
| ; SSE2-NEXT: packssdw %xmm1, %xmm0 | ||
| ; SSE2-NEXT: pand %xmm8, %xmm7 | ||
| ; SSE2-NEXT: pand %xmm8, %xmm6 | ||
| ; SSE2-NEXT: packssdw %xmm7, %xmm6 | ||
| ; SSE2-NEXT: pmulhw %xmm2, %xmm6 | ||
| ; SSE2-NEXT: pand %xmm8, %xmm5 | ||
| ; SSE2-NEXT: pand %xmm4, %xmm8 | ||
| ; SSE2-NEXT: packssdw %xmm5, %xmm8 | ||
| ; SSE2-NEXT: pmulhw %xmm8, %xmm0 | ||
| ; SSE2-NEXT: psrlw $1, %xmm0 | ||
| ; SSE2-NEXT: psrlw $1, %xmm6 | ||
| ; SSE2-NEXT: movdqa %xmm6, %xmm1 | ||
| ; SSE2-NEXT: retq | ||
| ; | ||
| ; SSE41-LABEL: and_mulhuw_v16i16_shift17: | ||
| ; SSE41: # %bb.0: | ||
| ; SSE41-NEXT: pmovsxwd {{.*#+}} xmm8 = [32767,32767,32767,32767] | ||
| ; SSE41-NEXT: pand %xmm8, %xmm3 | ||
| ; SSE41-NEXT: pand %xmm8, %xmm2 | ||
| ; SSE41-NEXT: packusdw %xmm3, %xmm2 | ||
| ; SSE41-NEXT: pand %xmm8, %xmm1 | ||
| ; SSE41-NEXT: pand %xmm8, %xmm0 | ||
| ; SSE41-NEXT: packusdw %xmm1, %xmm0 | ||
| ; SSE41-NEXT: pand %xmm8, %xmm7 | ||
| ; SSE41-NEXT: pand %xmm8, %xmm6 | ||
| ; SSE41-NEXT: packusdw %xmm7, %xmm6 | ||
| ; SSE41-NEXT: pmulhw %xmm2, %xmm6 | ||
| ; SSE41-NEXT: pand %xmm8, %xmm5 | ||
| ; SSE41-NEXT: pand %xmm4, %xmm8 | ||
| ; SSE41-NEXT: packusdw %xmm5, %xmm8 | ||
| ; SSE41-NEXT: pmulhw %xmm8, %xmm0 | ||
| ; SSE41-NEXT: psrlw $1, %xmm0 | ||
| ; SSE41-NEXT: psrlw $1, %xmm6 | ||
| ; SSE41-NEXT: movdqa %xmm6, %xmm1 | ||
| ; SSE41-NEXT: retq | ||
| ; | ||
| ; AVX2-LABEL: and_mulhuw_v16i16_shift17: | ||
| ; AVX2: # %bb.0: | ||
| ; AVX2-NEXT: vpbroadcastd {{.*#+}} ymm4 = [32767,32767,32767,32767,32767,32767,32767,32767] | ||
| ; AVX2-NEXT: vpand %ymm4, %ymm0, %ymm0 | ||
| ; AVX2-NEXT: vpand %ymm4, %ymm1, %ymm1 | ||
| ; AVX2-NEXT: vpand %ymm4, %ymm2, %ymm2 | ||
| ; AVX2-NEXT: vpmulhuw %ymm2, %ymm0, %ymm0 | ||
| ; AVX2-NEXT: vpand %ymm4, %ymm3, %ymm2 | ||
| ; AVX2-NEXT: vpmulhuw %ymm2, %ymm1, %ymm1 | ||
| ; AVX2-NEXT: vpackusdw %ymm1, %ymm0, %ymm0 | ||
| ; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,2,1,3] | ||
| ; AVX2-NEXT: vpsrlw $1, %ymm0, %ymm0 | ||
| ; AVX2-NEXT: retq | ||
| ; | ||
| ; AVX512F-LABEL: and_mulhuw_v16i16_shift17: | ||
| ; AVX512F: # %bb.0: | ||
| ; AVX512F-NEXT: vpmovdw %zmm1, %ymm1 | ||
| ; AVX512F-NEXT: vpbroadcastw {{.*#+}} ymm2 = [32767,32767,32767,32767,32767,32767,32767,32767,32767,32767,32767,32767,32767,32767,32767,32767] | ||
| ; AVX512F-NEXT: vpand %ymm2, %ymm1, %ymm1 | ||
| ; AVX512F-NEXT: vpmovdw %zmm0, %ymm0 | ||
| ; AVX512F-NEXT: vpand %ymm2, %ymm0, %ymm0 | ||
| ; AVX512F-NEXT: vpmulhw %ymm1, %ymm0, %ymm0 | ||
| ; AVX512F-NEXT: vpsrlw $1, %ymm0, %ymm0 | ||
| ; AVX512F-NEXT: retq | ||
| ; | ||
| ; AVX512BW-LABEL: and_mulhuw_v16i16_shift17: | ||
| ; AVX512BW: # %bb.0: | ||
| ; AVX512BW-NEXT: vpbroadcastd {{.*#+}} zmm2 = [32767,32767,32767,32767,32767,32767,32767,32767,32767,32767,32767,32767,32767,32767,32767,32767] | ||
| ; AVX512BW-NEXT: vpandd %zmm2, %zmm0, %zmm0 | ||
| ; AVX512BW-NEXT: vpandd %zmm2, %zmm1, %zmm1 | ||
| ; AVX512BW-NEXT: vpmulhuw %zmm1, %zmm0, %zmm0 | ||
| ; AVX512BW-NEXT: vpmovdw %zmm0, %ymm0 | ||
| ; AVX512BW-NEXT: vpsrlw $1, %ymm0, %ymm0 | ||
| ; AVX512BW-NEXT: retq | ||
| %a1 = and <16 x i32> %a, <i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767> | ||
| %b1 = and <16 x i32> %b, <i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767> | ||
| %c = mul <16 x i32> %a1, %b1 | ||
| %d = lshr <16 x i32> %c, splat (i32 17) | ||
| %e = trunc <16 x i32> %d to <16 x i16> | ||
| ret <16 x i16> %e | ||
| } | ||
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
maybe worth checking for
ShiftAmt.uge(InVT.getScalarSizeInBits())as well?Uh oh!
There was an error while loading. Please reload this page.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I tried using a shift value of that size, but the DAG builder replaces that node with
undefso the condition was not being hit