/freebsd/contrib/llvm-project/llvm/lib/Transforms/AggressiveInstCombine/ |
H A D | AggressiveInstCombine.cpp | 87 m_Sub(m_SpecificInt(Width), m_Deferred(ShAmt))))))) { in foldGuardedFunnelShift() 94 m_OneUse(m_c_Or(m_Shl(m_Value(ShVal0), m_Sub(m_SpecificInt(Width), in foldGuardedFunnelShift() 317 if ((match(Op0, m_Mul(m_Value(MulOp0), m_SpecificInt(Mask01)))) && in tryToRecognizePopCount() 318 match(Op1, m_SpecificInt(MaskShift))) { in tryToRecognizePopCount() 321 if (match(MulOp0, m_And(m_c_Add(m_LShr(m_Value(ShiftOp0), m_SpecificInt(4)), in tryToRecognizePopCount() 323 m_SpecificInt(Mask0F)))) { in tryToRecognizePopCount() 327 m_c_Add(m_And(m_Value(AndOp0), m_SpecificInt(Mask33)), in tryToRecognizePopCount() 328 m_And(m_LShr(m_Deferred(AndOp0), m_SpecificInt(2)), in tryToRecognizePopCount() 329 m_SpecificInt(Mask33))))) { in tryToRecognizePopCount() 333 match(SubOp1, m_And(m_LShr(m_Specific(Root), m_SpecificInt( in tryToRecognizePopCount() [all...] |
/freebsd/contrib/llvm-project/llvm/lib/Transforms/InstCombine/ |
H A D | InstCombineSelect.cpp | 950 match(TrueVal, m_Add(m_Specific(B), m_SpecificInt(-*C))))) in canonicalizeSaturatedSubtract() 954 match(TrueVal, m_Add(m_Specific(A), m_SpecificInt(-*C))))) in canonicalizeSaturatedSubtract() 1103 m_Xor(m_Value(Ctlz), m_SpecificInt(BitWidth - 1)))) in foldSelectCtlzToCttz() 1109 if (TrueVal != Ctlz && !match(TrueVal, m_SpecificInt(BitWidth))) in foldSelectCtlzToCttz() 1173 if (match(ValueOnZero, m_SpecificInt(SizeOfInBits))) { in foldSelectCttzCtlz() 1681 if (CmpC->isMinValue() && match(TVal, m_SpecificInt(*CmpC + 1))) in foldSelectInstWithICmpConst() 1684 if (CmpC->isMaxValue() && match(TVal, m_SpecificInt(*CmpC - 1))) in foldSelectInstWithICmpConst() 1687 if (CmpC->isMinSignedValue() && match(TVal, m_SpecificInt(*CmpC + 1))) in foldSelectInstWithICmpConst() 1690 if (CmpC->isMaxSignedValue() && match(TVal, m_SpecificInt(*CmpC - 1))) in foldSelectInstWithICmpConst() 2126 return match(Min, m_SpecificInt(MinVal)) && in foldOverflowingAddSubSelect() [all …]
|
H A D | InstCombineAndOrXor.cpp | 915 m_SpecificInt(1))) || in foldIsPowerOf2OrZero() 945 m_SpecificInt(2))) && in foldIsPowerOf2() 956 m_SpecificInt(1))) && in foldIsPowerOf2() 1761 m_SpecificInt(Op0->getType()->getScalarSizeInBits() - 1)))) && in foldCastedBitwiseLogic() 2908 if (match(R, m_OneUse(m_Sub(m_SpecificInt(Width), m_Specific(L))))) { in convertOrOfShiftsToFunnelShift() 2928 if (match(L, m_And(m_Value(X), m_SpecificInt(Mask))) && in convertOrOfShiftsToFunnelShift() 2929 match(R, m_And(m_Neg(m_Specific(X)), m_SpecificInt(Mask)))) in convertOrOfShiftsToFunnelShift() 2933 if (match(R, m_And(m_Neg(m_Specific(L)), m_SpecificInt(Mask)))) in convertOrOfShiftsToFunnelShift() 2938 if (match(L, m_ZExt(m_And(m_Value(X), m_SpecificInt(Mask)))) && in convertOrOfShiftsToFunnelShift() 2940 m_And(m_Neg(m_ZExt(m_And(m_Specific(X), m_SpecificInt(Mask)))), in convertOrOfShiftsToFunnelShift() [all …]
|
H A D | InstCombineCasts.cpp | 486 if (match(R, m_OneUse(m_Sub(m_SpecificInt(Width), m_Specific(L))))) in narrowFunnelShift() 498 if (match(L, m_And(m_Value(X), m_SpecificInt(Mask))) && in narrowFunnelShift() 499 match(R, m_And(m_Neg(m_Specific(X)), m_SpecificInt(Mask)))) in narrowFunnelShift() 503 if (match(L, m_ZExt(m_And(m_Value(X), m_SpecificInt(Mask)))) && in narrowFunnelShift() 504 match(R, m_ZExt(m_And(m_Neg(m_Specific(X)), m_SpecificInt(Mask))))) in narrowFunnelShift() 1538 m_SpecificInt(SrcBitSize - 1))))) { in visitSExt()
|
H A D | InstCombineAddSub.cpp | 1054 if (matchesSquareSum</*FP*/ false>(I, m_SpecificInt(1), A, B)) { in foldSquareSumInt() 1461 if (!match(&I, m_c_Add(m_Shl(m_Value(CrossSum), m_SpecificInt(HalfBits)), in foldBoxMultiply() 1470 if (!match(XLo, m_And(m_Value(X), m_SpecificInt(HalfMask))) || in foldBoxMultiply() 1471 !match(YLo, m_And(m_Value(Y), m_SpecificInt(HalfMask)))) in foldBoxMultiply() 1477 m_c_Add(m_c_Mul(m_LShr(m_Specific(Y), m_SpecificInt(HalfBits)), in foldBoxMultiply() 1479 m_c_Mul(m_LShr(m_Specific(X), m_SpecificInt(HalfBits)), in foldBoxMultiply() 2629 if (match(Op0, m_SpecificInt(BitWidth)) && in visitSub()
|
H A D | InstCombineShifts.cpp | 511 if (match(Op1, m_Or(m_Value(), m_SpecificInt(BitWidth - 1)))) in commonShiftTransforms() 1240 if (match(Op1, m_Sub(m_SpecificInt(BitWidth - 1), m_Value(X)))) in visitShl() 1462 if (match(Op0, m_OneUse(m_SRem(m_Value(X), m_SpecificInt(2))))) { in visitLShr()
|
H A D | InstCombineSimplifyDemanded.cpp | 1050 !match(I->getOperand(0), m_SpecificInt(LHSKnown.One))) { in SimplifyDemandedUseBits() 1057 !match(I->getOperand(1), m_SpecificInt(RHSKnown.One))) { in SimplifyDemandedUseBits() 1553 m_ExtractElt(m_Value(Vec), m_SpecificInt(IdxNo))) && in SimplifyDemandedVectorElts()
|
H A D | InstCombineMulDivRem.cpp | 1260 match(Op0, m_NSWAddLike(m_NSWMul(m_Value(X), m_SpecificInt(*C2)), in commonIDivTransforms() 1266 match(Op0, m_NUWAddLike(m_NUWMul(m_Value(X), m_SpecificInt(*C2)), in commonIDivTransforms()
|
H A D | InstCombineCompares.cpp | 1941 m_Select(m_Value(B), m_SpecificInt(*TC), m_SpecificInt(*FC))) && in foldICmpAndConstant() 5745 if (match(Op0, m_AShr(m_Trunc(m_Value(A)), m_SpecificInt(BitWidth - 1))) && in foldICmpEquality() 5746 match(Op1, m_Trunc(m_LShr(m_Specific(A), m_SpecificInt(BitWidth)))) && in foldICmpEquality()
|
H A D | InstructionCombining.cpp | 2950 m_SpecificInt(countr_zero(TyAllocSize))))) || in visitGetElementPtrInst() 2952 m_IDiv(m_Value(V), m_SpecificInt(TyAllocSize)))) { in visitGetElementPtrInst()
|
H A D | InstCombineCalls.cpp | 1851 match(I1, m_OneUse(m_And(m_Value(Y), m_SpecificInt(*RHSC))))) in visitCallInst() 2088 if (Op0 == Op1 && BitWidth == 16 && match(ShAmtC, m_SpecificInt(8))) { in visitCallInst()
|
H A D | InstCombineVectorOps.cpp | 1345 if (!match(Scalar, m_ExtractElt(m_Specific(X), m_SpecificInt(IdxC)))) in foldInsEltIntoIdentityShuffle()
|
/freebsd/contrib/llvm-project/llvm/include/llvm/CodeGen/ |
H A D | SDPatternMatch.h | 768 inline SpecificInt_match m_SpecificInt(APInt V) { 771 inline SpecificInt_match m_SpecificInt(uint64_t V) { 775 inline SpecificInt_match m_Zero() { return m_SpecificInt(0U); } 776 inline SpecificInt_match m_One() { return m_SpecificInt(1U); }
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/Vectorize/ |
H A D | VPlanPatternMatch.h | 79 inline specific_intval<0> m_SpecificInt(uint64_t V) { in m_SpecificInt() function
|
H A D | VPlanTransforms.cpp | 1064 if (match(&R, m_c_Mul(m_VPValue(A), m_SpecificInt(1)))) in simplifyRecipe()
|
H A D | VPlan.cpp | 1640 (match(A, m_ScalarIVSteps(m_CanonicalIV(), m_SpecificInt(1))) || in isHeaderMask()
|
H A D | VectorCombine.cpp | 647 m_ExtractElt(m_Value(SrcVec), m_SpecificInt(Index)))))) in foldInsExtFNeg()
|
/freebsd/contrib/llvm-project/llvm/include/llvm/IR/ |
H A D | PatternMatch.h | 972 inline specific_intval<false> m_SpecificInt(const APInt &V) { in m_SpecificInt() function 976 inline specific_intval64<false> m_SpecificInt(uint64_t V) { in m_SpecificInt() function 2826 auto LHS = m_AShr(m_Value(OpL), m_SpecificInt(ShiftWidth)); 2827 auto RHS = m_LShr(m_Neg(m_Value(OpR)), m_SpecificInt(ShiftWidth)); 2914 m_SpecificInt(1).match(GEP->idx_begin()->get()))
|
/freebsd/contrib/llvm-project/llvm/lib/Analysis/ |
H A D | InstructionSimplify.cpp | 1474 match(Op1, m_SpecificInt(Ty->getScalarSizeInBits() - 1))) in simplifyShlInst() 2222 match(Op1, m_Xor(m_Specific(A), m_SpecificInt(~*C1)))) in simplifyAndInst() 2494 match(Op1, m_Xor(m_Specific(A), m_SpecificInt(~*C1)))) in simplifyOrInst() 4669 if (match(Y, m_SpecificInt(LimitC))) in simplifySelectWithICmpCond() 5061 m_SpecificInt(TyAllocSize))) && in simplifyGEPInst() 6866 if (ScaledOne.isNonNegative() && match(Op1, m_SpecificInt(ScaledOne))) in simplifyIntrinsic()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/AMDGPU/ |
H A D | AMDGPUPromoteAlloca.cpp | 385 match(I->getOperand(2), m_SpecificInt(Size)) && !I->isVolatile(); in isSupportedMemset()
|
/freebsd/contrib/llvm-project/llvm/lib/CodeGen/ |
H A D | CodeGenPrepare.cpp | 8003 m_SpecificInt(HalfValBitSize)))))) in splitMergedValStore() 8267 match(UI, m_Shr(m_Specific(X), m_SpecificInt(CmpC.logBase2())))) { in optimizeBranch() 8280 (match(UI, m_Add(m_Specific(X), m_SpecificInt(-CmpC))) || in optimizeBranch() 8281 match(UI, m_Sub(m_Specific(X), m_SpecificInt(CmpC))))) { in optimizeBranch()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/AArch64/ |
H A D | AArch64TargetTransformInfo.cpp | 1840 m_Value(IndexBase), m_SpecificInt(1)))) { in instCombineLD1GatherIndex() 1870 m_Value(IndexBase), m_SpecificInt(1)))) { in instCombineST1ScatterIndex() 2465 m_c_Add(m_Instruction(Ex2), m_SpecificInt(1)))))) in isExtPartOfAvgExpr()
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/Scalar/ |
H A D | LowerMatrixIntrinsics.cpp | 1362 m_Value(), m_SpecificInt(1)))))); in lowerDotProduct()
|
/freebsd/contrib/llvm-project/llvm/lib/CodeGen/SelectionDAG/ |
H A D | DAGCombiner.cpp | 2572 m_SpecificInt(1))))) { in foldSubToAvg() 2578 m_SpecificInt(1))))) { in foldSubToAvg() 2937 m_SpecificInt(1))))) { in foldAddToAvg() 2943 m_SpecificInt(1))))) { in foldAddToAvg() 4031 sd_match(N1, m_Sra(m_Value(A), m_SpecificInt(BitWidth - 1))) && in visitSUB() 4516 m_Mul(m_Or(m_Sra(m_Value(X), m_SpecificInt(BitWidth - 1)), m_One()), in visitMUL() 6844 if (!sd_match(N, m_And(m_OneUse(m_Xor(m_Value(X), m_SpecificInt(SignMask))), in foldAndToUsubsat() 6846 m_SpecificInt(BitWidth - 1))))) && in foldAndToUsubsat() 6847 !sd_match(N, m_And(m_OneUse(m_Add(m_Value(X), m_SpecificInt(SignMask))), in foldAndToUsubsat() 6849 m_SpecificInt(BitWidth - 1)))))) in foldAndToUsubsat() [all …]
|
/freebsd/contrib/llvm-project/llvm/lib/Target/X86/ |
H A D | X86ISelLowering.cpp | 34281 match(Op.get(), m_AShr(m_Shl(m_Value(), m_SpecificInt(32)), in shouldSinkOperands() 34282 m_SpecificInt(32)))) { in shouldSinkOperands() 34287 m_And(m_Value(), m_SpecificInt(UINT64_C(0xffffffff))))) { in shouldSinkOperands() 46270 m_SpecificInt(VT.getScalarSizeInBits()), in combineSelect() 46282 m_SpecificInt(VT.getScalarSizeInBits()), in combineSelect() 48120 sd_match(Cond, m_SetCC(m_Specific(N1), m_SpecificInt(EltSizeInBits), in combineShiftLeft() 48126 sd_match(Cond, m_SetCC(m_Specific(N1), m_SpecificInt(EltSizeInBits), in combineShiftLeft() 48185 m_SpecificInt(VT.getScalarSizeInBits() - 1)))) in combineShiftRightArithmetic() 48259 sd_match(Cond, m_SetCC(m_Specific(N1), m_SpecificInt(EltSizeInBits), in combineShiftRightLogical() 48265 sd_match(Cond, m_SetCC(m_Specific(N1), m_SpecificInt(EltSizeInBits), in combineShiftRightLogical()
|