/freebsd/contrib/llvm-project/llvm/lib/Target/SystemZ/ |
H A D | SystemZTargetTransformInfo.cpp | 408 static unsigned getScalarSizeInBits(Type *Ty) { in getScalarSizeInBits() function 410 (Ty->isPtrOrPtrVectorTy() ? 64U : Ty->getScalarSizeInBits()); in getScalarSizeInBits() 420 unsigned WideBits = getScalarSizeInBits(Ty) * VTy->getNumElements(); in getNumVectorRegs() 442 unsigned ScalarBits = Ty->getScalarSizeInBits(); in getArithmeticInstrCost() 648 unsigned Bits0 = Ty0->getScalarSizeInBits(); in getElSizeLog2Diff() 649 unsigned Bits1 = Ty1->getScalarSizeInBits(); in getElSizeLog2Diff() 691 if (VF == 8 && SrcTy->getScalarSizeInBits() == 64 && in getVectorTruncCost() 692 DstTy->getScalarSizeInBits() == 8) in getVectorTruncCost() 706 unsigned SrcScalarBits = SrcTy->getScalarSizeInBits(); in getVectorBitmaskConversionCost() 707 unsigned DstScalarBits = DstTy->getScalarSizeInBits(); in getVectorBitmaskConversionCost() [all …]
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/InstCombine/ |
H A D | InstCombineCasts.cpp | 290 uint32_t OrigBitWidth = OrigTy->getScalarSizeInBits(); in canEvaluateTruncated() 291 uint32_t BitWidth = Ty->getScalarSizeInBits(); in canEvaluateTruncated() 306 uint32_t BitWidth = Ty->getScalarSizeInBits(); in canEvaluateTruncated() 320 uint32_t OrigBitWidth = OrigTy->getScalarSizeInBits(); in canEvaluateTruncated() 321 uint32_t BitWidth = Ty->getScalarSizeInBits(); in canEvaluateTruncated() 338 uint32_t OrigBitWidth = OrigTy->getScalarSizeInBits(); in canEvaluateTruncated() 339 uint32_t BitWidth = Ty->getScalarSizeInBits(); in canEvaluateTruncated() 382 return Ty->getScalarSizeInBits() >= MinBitWidth; in canEvaluateTruncated() 449 unsigned NarrowWidth = DestTy->getScalarSizeInBits(); in narrowFunnelShift() 450 unsigned WideWidth = Trunc.getSrcTy()->getScalarSizeInBits(); in narrowFunnelShift() [all …]
|
H A D | InstCombineShifts.cpp | 40 (Sh0->getType()->getScalarSizeInBits() - 1) + in canTryToConstantAddTwoShiftAmounts() 41 (Sh1->getType()->getScalarSizeInBits() - 1); in canTryToConstantAddTwoShiftAmounts() 43 APInt::getAllOnes(ShAmt0->getType()->getScalarSizeInBits()); in canTryToConstantAddTwoShiftAmounts() 114 unsigned NewShAmtBitWidth = NewShAmt->getType()->getScalarSizeInBits(); in reassociateShiftAmtsOfTwoSameDirectionShifts() 115 unsigned XBitWidth = X->getType()->getScalarSizeInBits(); in reassociateShiftAmtsOfTwoSameDirectionShifts() 252 ExtendedTy->getScalarSizeInBits())); in dropRedundantMaskingOfLeftShiftInput() 289 unsigned WidestTyBitWidth = WidestTy->getScalarSizeInBits(); in dropRedundantMaskingOfLeftShiftInput() 376 unsigned Size = Ty->getScalarSizeInBits(); in foldShiftOfShiftedBinOp() 455 unsigned BitWidth = Ty->getScalarSizeInBits(); in commonShiftTransforms() 548 unsigned TypeWidth = InnerShift->getType()->getScalarSizeInBits(); in canEvaluateShiftedShift() [all …]
|
/freebsd/contrib/llvm-project/llvm/include/llvm/CodeGenTypes/ |
H A D | LowLevelType.h | 195 return TypeSize::getFixed(getScalarSizeInBits()); in getSizeInBits() 197 return TypeSize(getScalarSizeInBits() * EC.getKnownMinValue(), in getSizeInBits() 239 assert((!isScalar() || getScalarSizeInBits() != 0) && in divide() 247 assert(getScalarSizeInBits() % Factor == 0); in divide() 248 return scalar(getScalarSizeInBits() / Factor); in divide() 267 constexpr unsigned getScalarSizeInBits() const { in getScalarSizeInBits() function 293 return pointer(getAddressSpace(), getScalarSizeInBits()); in getElementType() 295 return scalar(getScalarSizeInBits()); in getElementType()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/RISCV/ |
H A D | RISCVTargetTransformInfo.cpp | 348 assert((DataVT.getScalarSizeInBits() != 8 || in getVRGatherIndexType() 379 if (EltTp.getScalarSizeInBits() < ST->getELen()) { in getShuffleCost() 401 (LT.second.getScalarSizeInBits() != 8 || in getShuffleCost() 416 (LT.second.getScalarSizeInBits() != 8 || in getShuffleCost() 494 SubLT.second.getScalarSizeInBits() * Index % MinVLen == 0 && in getShuffleCost() 527 if (LT.second.getScalarSizeInBits() == 1) { in getShuffleCost() 1034 (Src->getScalarSizeInBits() <= ST->getELen()) && in getCastInstrCost() 1035 (Dst->getScalarSizeInBits() <= ST->getELen()); in getCastInstrCost() 1047 int PowDiff = (int)Log2_32(Dst->getScalarSizeInBits()) - in getCastInstrCost() 1048 (int)Log2_32(Src->getScalarSizeInBits()); in getCastInstrCost() [all …]
|
/freebsd/contrib/llvm-project/llvm/lib/Target/AArch64/GISel/ |
H A D | AArch64PreLegalizerCombiner.cpp | 247 if (DstTy.getScalarSizeInBits() != 32 || MidTy.getScalarSizeInBits() != 32) in matchExtAddvToUdotAddv() 284 if (SrcTy.getScalarSizeInBits() != 8 || SrcTy.getNumElements() % 8 != 0) in matchExtAddvToUdotAddv() 435 if ((DstTy.getScalarSizeInBits() == 16 && in matchExtUaddvToUaddlv() 437 (DstTy.getScalarSizeInBits() == 32 && in matchExtUaddvToUaddlv() 439 (DstTy.getScalarSizeInBits() == 64 && in matchExtUaddvToUaddlv() 463 unsigned SrcScalSize = SrcTy.getScalarSizeInBits(); in applyExtUaddvToUaddlv() 492 unsigned MidScalarSize = MainTy.getScalarSizeInBits() * 2; in applyExtUaddvToUaddlv() 499 if ((WorkingRegTy.getScalarSizeInBits() == 8) && in applyExtUaddvToUaddlv() 544 if (DstTy.getScalarSizeInBits() > MidScalarSize) { in applyExtUaddvToUaddlv() 576 unsigned ExtDstScal = MRI.getType(ExtDstReg).getScalarSizeInBits(); in matchPushAddSubExt() [all …]
|
H A D | AArch64LegalizerInfo.cpp | 610 return (Query.Types[0].getScalarSizeInBits() > in AArch64LegalizerInfo() 611 Query.Types[1].getScalarSizeInBits() * 2) && in AArch64LegalizerInfo() 613 (Query.Types[1].getScalarSizeInBits() == 8 || in AArch64LegalizerInfo() 614 Query.Types[1].getScalarSizeInBits() == 16); in AArch64LegalizerInfo() 632 DstTy.getScalarSizeInBits() * 2 <= SrcTy.getScalarSizeInBits(); in AArch64LegalizerInfo() 696 return Query.Types[0].getScalarSizeInBits() <= 64 && in AArch64LegalizerInfo() 697 Query.Types[0].getScalarSizeInBits() > in AArch64LegalizerInfo() 698 Query.Types[1].getScalarSizeInBits(); in AArch64LegalizerInfo() 703 return Query.Types[1].getScalarSizeInBits() <= 64 && in AArch64LegalizerInfo() 704 Query.Types[0].getScalarSizeInBits() < in AArch64LegalizerInfo() [all …]
|
H A D | AArch64PostLegalizerLowering.cpp | 164 unsigned EltSize = Ty.getScalarSizeInBits(); in matchREV() 359 uint64_t ExtFactor = MRI.getType(V1).getScalarSizeInBits() / 8; in matchEXT() 519 int64_t ElementBits = Ty.getScalarSizeInBits(); in isVShiftRImm() 984 unsigned EltSize = MRI.getType(LHS).getScalarSizeInBits(); in matchLowerVectorFCMP() 1152 (MRI.getType(I1->getOperand(0).getReg()).getScalarSizeInBits() == in matchExtMulToMULL() 1153 MRI.getType(I1->getOperand(1).getReg()).getScalarSizeInBits() * 2) && in matchExtMulToMULL() 1154 (MRI.getType(I2->getOperand(0).getReg()).getScalarSizeInBits() == in matchExtMulToMULL() 1155 MRI.getType(I2->getOperand(1).getReg()).getScalarSizeInBits() * 2)) { in matchExtMulToMULL() 1181 (MRI.getType(I1->getOperand(0).getReg()).getScalarSizeInBits() == in applyExtMulToMULL() 1182 MRI.getType(I1->getOperand(1).getReg()).getScalarSizeInBits() * 2) && in applyExtMulToMULL() [all …]
|
/freebsd/contrib/llvm-project/llvm/lib/Target/ARM/ |
H A D | MVEGatherScatterLowering.cpp | 199 ->getScalarSizeInBits(); in checkOffsetSize() 235 MemoryTy->getScalarSizeInBits()); in decomposePtr() 244 if (PtrTy->getNumElements() != 4 || MemoryTy->getScalarSizeInBits() == 32) in decomposePtr() 291 ->getScalarSizeInBits() != 32) in decomposeGEP() 298 if ((Ty->getElementType()->getScalarSizeInBits() < in decomposeGEP() 299 OffsetType->getElementType()->getScalarSizeInBits())) { in decomposeGEP() 415 if (!isLegalTypeAndAlignment(Ty->getNumElements(), Ty->getScalarSizeInBits(), in lowerGather() 459 if (Ty->getNumElements() != 4 || Ty->getScalarSizeInBits() != 32) in tryCreateMaskedGatherBase() 480 if (Ty->getNumElements() != 4 || Ty->getScalarSizeInBits() != 32) in tryCreateMaskedGatherBaseWB() 562 {BasePtr, Offsets, Builder.getInt32(MemoryTy->getScalarSizeInBits()), in tryCreateMaskedGatherOffset() [all …]
|
H A D | MVELaneInterleavingPass.cpp | 306 unsigned BaseElts = VT->getScalarSizeInBits() == 16 in tryInterleave() 308 : (VT->getScalarSizeInBits() == 8 ? 16 : 0); in tryInterleave() 313 if (Start->getOperand(0)->getType()->getScalarSizeInBits() != in tryInterleave() 314 VT->getScalarSizeInBits() * 2) { in tryInterleave()
|
H A D | ARMTargetTransformInfo.cpp | 222 assert(II.getArgOperand(CarryOp)->getType()->getScalarSizeInBits() == 32 && in instCombineIntrinsic() 889 ValTy->isVectorTy() && ValTy->getScalarSizeInBits() <= 32) in getVectorInstrCost() 902 ValTy->getScalarSizeInBits() <= 32) in getVectorInstrCost() 1122 unsigned EltWidth = DataTy->getScalarSizeInBits(); in isLegalMaskedLoad() 1131 unsigned EltWidth = Ty->getScalarSizeInBits(); in isLegalMaskedGather() 1591 unsigned EltSize = VTy->getScalarSizeInBits(); in getGatherScatterOpCost() 1628 cast<Instruction>(Us)->getType()->getScalarSizeInBits(); in getGatherScatterOpCost() 1642 unsigned TypeSize = T->getOperand(0)->getType()->getScalarSizeInBits(); in getGatherScatterOpCost() 1673 if (ZExt->getOperand(0)->getType()->getScalarSizeInBits() <= ExtSize) in getGatherScatterOpCost() 1688 unsigned EltSize = ValVT.getScalarSizeInBits(); in getArithmeticReductionCost() [all …]
|
/freebsd/contrib/llvm-project/llvm/lib/CodeGen/GlobalISel/ |
H A D | LegalizeMutations.cpp | 72 const LLT NewEltTy = LLT::scalar(NewTy.getScalarSizeInBits()); in changeElementSizeTo() 82 std::max(1u << Log2_32_Ceil(Ty.getScalarSizeInBits()), Min); in widenScalarOrEltToNextPow2() 92 unsigned NewEltSizeInBits = alignTo(Ty.getScalarSizeInBits(), Size); in widenScalarOrEltToNextMultipleOf()
|
H A D | GISelKnownBits.cpp | 90 unsigned BitWidth = Ty.getScalarSizeInBits(); in signBitIsZero() 160 unsigned BitWidth = DstTy.getScalarSizeInBits(); in computeKnownBitsImpl() 410 KnownBits KnownRange(MMO->getMemoryType().getScalarSizeInBits()); in computeKnownBitsImpl() 421 KnownBits KnownRange(MMO->getMemoryType().getScalarSizeInBits()); in computeKnownBitsImpl() 672 const unsigned TyBits = DstTy.getScalarSizeInBits(); in computeNumSignBits() 696 unsigned Tmp = DstTy.getScalarSizeInBits() - SrcTy.getScalarSizeInBits(); in computeNumSignBits() 763 unsigned DstTyBits = DstTy.getScalarSizeInBits(); in computeNumSignBits() 764 unsigned NumSrcBits = SrcTy.getScalarSizeInBits(); in computeNumSignBits()
|
H A D | CallLowering.cpp | 394 PartLLT.getScalarSizeInBits() > LLTy.getScalarSizeInBits() && in buildCopyFromRegs() 403 SrcReg = B.buildAssertSExt(LocTy, SrcReg, LLTy.getScalarSizeInBits()) in buildCopyFromRegs() 406 SrcReg = B.buildAssertZExt(LocTy, SrcReg, LLTy.getScalarSizeInBits()) in buildCopyFromRegs() 446 PartLLT.getScalarSizeInBits() == LLTy.getScalarSizeInBits() * 2 && in buildCopyFromRegs() 567 PartTy.getScalarSizeInBits() > SrcTy.getScalarSizeInBits()) { in buildCopyToRegs() 600 SrcTy.getScalarSizeInBits() > PartTy.getSizeInBits()) { in buildCopyToRegs() 603 LLT::scalar(PartTy.getScalarSizeInBits() * DstRegs.size() / in buildCopyToRegs() 860 int FrameIdx = MFI.CreateStackObject(OrigTy.getScalarSizeInBits(), in handleAssignments() 1256 ValTy.getScalarSizeInBits()); in getStackValueStoreType() 1349 NarrowTy.getScalarSizeInBits()) in buildExtensionHint() [all …]
|
H A D | LegalityPredicates.cpp | 137 return QueryTy.getScalarSizeInBits() < Size; in scalarOrEltNarrowerThan() 145 return QueryTy.getScalarSizeInBits() > Size; in scalarOrEltWiderThan() 152 return !isPowerOf2_32(QueryTy.getScalarSizeInBits()); in scalarOrEltSizeNotPow2()
|
H A D | LegalizerInfo.cpp | 168 if (NewTy.getScalarSizeInBits() >= OldTy.getScalarSizeInBits()) in mutationIsSane() 172 if (NewTy.getScalarSizeInBits() <= OldTy.getScalarSizeInBits()) in mutationIsSane()
|
H A D | Utils.cpp | 534 RegTy.getScalarSizeInBits() == MainTy.getScalarSizeInBits() && in extractParts() 537 LLT::fixed_vector(LeftoverNumElts, RegTy.getScalarSizeInBits()); in extractParts() 946 return MaybeOp1Cst->trunc(Imm).sext(Ty.getScalarSizeInBits()); in ConstantFoldExtOp() 960 const unsigned DstSize = DstTy.getScalarSizeInBits(); in ConstantFoldCastOp() 1101 unsigned BitWidth = Ty.getScalarSizeInBits(); in isKnownToBeAPowerOfTwo() 1137 const unsigned BitWidth = Ty.getScalarSizeInBits(); in isKnownToBeAPowerOfTwo() 1240 (OrigTy.getScalarSizeInBits() != TargetTy.getScalarSizeInBits())) in getCoverTy() 1524 const unsigned ScalarSize = MRI.getType(Def).getScalarSizeInBits(); in isConstantOrConstantSplatVector() 1754 return Val->Value.ult(Ty.getScalarSizeInBits()); in shiftAmountKnownInRange() 1767 if (!Val->Value.ult(Ty.getScalarSizeInBits())) in shiftAmountKnownInRange() [all …]
|
/freebsd/contrib/llvm-project/llvm/lib/Target/AMDGPU/ |
H A D | AMDGPULateCodeGenPrepare.cpp | 134 EltTy->getScalarSizeInBits() > ConvertToScalar->getScalarSizeInBits()) in shouldReplace() 192 assert(OriginalType->getScalarSizeInBits() <= in calculateConvertType() 193 ConvertToScalar->getScalarSizeInBits()); in calculateConvertType() 225 uint64_t ExpandedVecElementCount = NewSize / VTy->getScalarSizeInBits(); in convertToOptType() 265 Type::getIntNTy(Mod->getContext(), NewVTy->getScalarSizeInBits()), in convertFromOptType() 266 (OriginalSize / NewVTy->getScalarSizeInBits()), false); in convertFromOptType()
|
/freebsd/contrib/llvm-project/llvm/include/llvm/CodeGen/GlobalISel/ |
H A D | LegalizerInfo.h | 948 scalarOrEltNarrowerThan(TypeIdx, Ty.getScalarSizeInBits()), in minScalarOrElt() 959 TypeIdx, Ty.getScalarSizeInBits())), in minScalarOrEltIf() 1015 scalarOrEltWiderThan(TypeIdx, Ty.getScalarSizeInBits()), in maxScalarOrElt() 1064 return Query.Types[LargeTypeIdx].getScalarSizeInBits() > in minScalarSameAs() 1075 return Query.Types[NarrowTypeIdx].getScalarSizeInBits() < in maxScalarSameAs() 1094 return Query.Types[LargeTypeIdx].getScalarSizeInBits() > in minScalarEltSameAsIf() 1095 Query.Types[TypeIdx].getScalarSizeInBits() && in minScalarEltSameAsIf() 1101 T = T.changeElementType(LLT::scalar(T.getScalarSizeInBits())); in minScalarEltSameAsIf() 1113 return Query.Types[SmallTypeIdx].getScalarSizeInBits() < in maxScalarEltSameAsIf() 1114 Query.Types[TypeIdx].getScalarSizeInBits() && in maxScalarEltSameAsIf()
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/AggressiveInstCombine/ |
H A D | TruncInstCombine.cpp | 181 unsigned TruncBitWidth = DstTy->getScalarSizeInBits(); in getMinBitWidth() 183 CurrentTruncInst->getOperand(0)->getType()->getScalarSizeInBits(); in getMinBitWidth() 252 MinBitWidth = Ty ? Ty->getScalarSizeInBits() : OrigBitWidth; in getMinBitWidth() 289 I->getOperand(0)->getType()->getScalarSizeInBits(); in getBestTruncatedType() 297 CurrentTruncInst->getOperand(0)->getType()->getScalarSizeInBits(); in getBestTruncatedType()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/AArch64/ |
H A D | AArch64TargetTransformInfo.cpp | 292 FVTy->getScalarSizeInBits() * FVTy->getNumElements() > 128; in areTypesABICompatible() 523 (EltTy->getScalarSizeInBits() != 32 && in getHistogramCost() 524 EltTy->getScalarSizeInBits() != 64)) in getHistogramCost() 585 LT.second.getScalarSizeInBits() == RetTy->getScalarSizeInBits() ? 1 : 4; in getIntrinsicInstrCost() 604 LT.second.getScalarSizeInBits() == RetTy->getScalarSizeInBits()) in getIntrinsicInstrCost() 702 int ExtraCost = MTy.isVector() && MTy.getScalarSizeInBits() != in getIntrinsicInstrCost() 703 RetTy->getScalarSizeInBits() in getIntrinsicInstrCost() 761 (LT.second.getScalarSizeInBits() == MTy.getScalarSizeInBits() || in getIntrinsicInstrCost() 769 (LT.second.getScalarSizeInBits() == MTy.getScalarSizeInBits())))) in getIntrinsicInstrCost() 776 LT.second.getScalarSizeInBits() >= MTy.getScalarSizeInBits()) { in getIntrinsicInstrCost() [all …]
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/Scalar/ |
H A D | BDCE.cpp | 119 const uint32_t SrcBitSize = SE->getSrcTy()->getScalarSizeInBits(); in bitTrackingDCE() 121 const uint32_t DestBitSize = DstTy->getScalarSizeInBits(); in bitTrackingDCE()
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/Vectorize/ |
H A D | VPlanTransforms.cpp | 542 assert(ResultTy->getScalarSizeInBits() > TruncTy->getScalarSizeInBits() && in createScalarIVSteps() 553 assert(StepTy->getScalarSizeInBits() > ResultTy->getScalarSizeInBits() && in createScalarIVSteps() 1015 if (ATy->getScalarSizeInBits() < TruncTy->getScalarSizeInBits()) { in simplifyRecipe() 1028 } else if (ATy->getScalarSizeInBits() > TruncTy->getScalarSizeInBits()) { in simplifyRecipe() 1143 unsigned OldResSizeInBits = OldResTy->getScalarSizeInBits(); in truncateToMinimalBitwidths() 1179 TypeInfo.inferScalarType(Op)->getScalarSizeInBits(); in truncateToMinimalBitwidths() 1462 if (unsigned IVSize = CanonicalIVPHI->getScalarType()->getScalarSizeInBits(); in tryAddExplicitVectorLength()
|
/freebsd/contrib/llvm-project/llvm/include/llvm/IR/ |
H A D | MatrixBuilder.h | 234 B.getIntN(Idx->getType()->getScalarSizeInBits(), NumElements); 246 unsigned MaxWidth = std::max(RowIdx->getType()->getScalarSizeInBits(), 247 ColumnIdx->getType()->getScalarSizeInBits());
|
/freebsd/contrib/llvm-project/llvm/lib/CodeGen/ |
H A D | TypePromotion.cpp | 208 return V->getType()->getScalarSizeInBits() == TypeSize; in EqualTypeSize() 212 return V->getType()->getScalarSizeInBits() <= TypeSize; in LessOrEqualTypeSize() 216 return V->getType()->getScalarSizeInBits() > TypeSize; in GreaterThanTypeSize() 220 return V->getType()->getScalarSizeInBits() < TypeSize; in LessThanTypeSize() 579 if (ZExt->getType()->getScalarSizeInBits() >= PromotedWidth) in TruncateSinks() 642 unsigned NumBits = DestTy->getScalarSizeInBits(); in ConvertTruncs()
|