| /freebsd/contrib/llvm-project/llvm/lib/Target/X86/MCTargetDesc/ |
| H A D | X86ShuffleDecode.cpp | 50 void DecodeInsertElementMask(unsigned NumElts, unsigned Idx, unsigned Len, in DecodeInsertElementMask() argument 52 assert((Idx + Len) <= NumElts && "Insertion out of range"); in DecodeInsertElementMask() 54 for (unsigned i = 0; i != NumElts; ++i) in DecodeInsertElementMask() 57 ShuffleMask[Idx + i] = NumElts + i; in DecodeInsertElementMask() 78 void DecodeMOVSLDUPMask(unsigned NumElts, SmallVectorImpl<int> &ShuffleMask) { in DecodeMOVSLDUPMask() argument 79 for (int i = 0, e = NumElts / 2; i < e; ++i) { in DecodeMOVSLDUPMask() 85 void DecodeMOVSHDUPMask(unsigned NumElts, SmallVectorImpl<int> &ShuffleMask) { in DecodeMOVSHDUPMask() argument 86 for (int i = 0, e = NumElts / 2; i < e; ++i) { in DecodeMOVSHDUPMask() 92 void DecodeMOVDDUPMask(unsigned NumElts, SmallVectorImpl<int> &ShuffleMask) { in DecodeMOVDDUPMask() argument 95 for (unsigned l = 0; l < NumElts; l += NumLaneElts) in DecodeMOVDDUPMask() [all …]
|
| H A D | X86ShuffleDecode.h | 36 void DecodeInsertElementMask(unsigned NumElts, unsigned Idx, unsigned Len, 47 void DecodeMOVSLDUPMask(unsigned NumElts, SmallVectorImpl<int> &ShuffleMask); 49 void DecodeMOVSHDUPMask(unsigned NumElts, SmallVectorImpl<int> &ShuffleMask); 51 void DecodeMOVDDUPMask(unsigned NumElts, SmallVectorImpl<int> &ShuffleMask); 53 void DecodePSLLDQMask(unsigned NumElts, unsigned Imm, 56 void DecodePSRLDQMask(unsigned NumElts, unsigned Imm, 59 void DecodePALIGNRMask(unsigned NumElts, unsigned Imm, 62 void DecodeVALIGNMask(unsigned NumElts, unsigned Imm, 66 void DecodePSHUFMask(unsigned NumElts, unsigned ScalarBits, unsigned Imm, 70 void DecodePSHUFHWMask(unsigned NumElts, unsigned Imm, [all …]
|
| H A D | X86IntelInstPrinter.cpp | 189 unsigned NumElts; in printVecCompareInstr() local 191 NumElts = (Desc.TSFlags & X86II::REX_W) ? 8 : 16; in printVecCompareInstr() 193 NumElts = (Desc.TSFlags & X86II::REX_W) ? 4 : 8; in printVecCompareInstr() 195 NumElts = (Desc.TSFlags & X86II::REX_W) ? 2 : 4; in printVecCompareInstr() 198 NumElts *= 2; in printVecCompareInstr() 200 OS << "{1to" << NumElts << "}"; in printVecCompareInstr() 338 unsigned NumElts; in printVecCompareInstr() local 340 NumElts = (Desc.TSFlags & X86II::REX_W) ? 8 : 16; in printVecCompareInstr() 342 NumElts = (Desc.TSFlags & X86II::REX_W) ? 4 : 8; in printVecCompareInstr() 344 NumElts = (Desc.TSFlags & X86II::REX_W) ? 2 : 4; in printVecCompareInstr() [all …]
|
| H A D | X86ATTInstPrinter.cpp | 215 unsigned NumElts; in printVecCompareInstr() local 217 NumElts = (Desc.TSFlags & X86II::REX_W) ? 8 : 16; in printVecCompareInstr() 219 NumElts = (Desc.TSFlags & X86II::REX_W) ? 4 : 8; in printVecCompareInstr() 221 NumElts = (Desc.TSFlags & X86II::REX_W) ? 2 : 4; in printVecCompareInstr() 224 NumElts *= 2; in printVecCompareInstr() 226 OS << "{1to" << NumElts << "}"; in printVecCompareInstr() 366 unsigned NumElts; in printVecCompareInstr() local 368 NumElts = (Desc.TSFlags & X86II::REX_W) ? 8 : 16; in printVecCompareInstr() 370 NumElts = (Desc.TSFlags & X86II::REX_W) ? 4 : 8; in printVecCompareInstr() 372 NumElts = (Desc.TSFlags & X86II::REX_W) ? 2 : 4; in printVecCompareInstr() [all …]
|
| /freebsd/contrib/llvm-project/llvm/lib/Target/AArch64/ |
| H A D | AArch64PerfectShuffle.h | 6627 inline bool isZIPMask(ArrayRef<int> M, unsigned NumElts, in isZIPMask() argument 6629 if (NumElts % 2 != 0) in isZIPMask() 6633 for (unsigned i = 0; i != NumElts / 2; i++) { in isZIPMask() 6638 WhichResult = ((unsigned)M[i * 2 + 1] == NumElts + i ? 0 : 1); in isZIPMask() 6646 unsigned Idx = WhichResult * NumElts / 2; in isZIPMask() 6647 for (unsigned i = 0; i != NumElts; i += 2) { in isZIPMask() 6649 (M[i + 1] >= 0 && (unsigned)M[i + 1] != Idx + NumElts)) in isZIPMask() 6660 inline bool isUZPMask(ArrayRef<int> M, unsigned NumElts, in isUZPMask() argument 6664 for (unsigned i = 0; i != NumElts; i++) { in isUZPMask() 6674 for (unsigned i = 0; i != NumElts; ++i) { in isUZPMask() [all …]
|
| /freebsd/contrib/llvm-project/llvm/lib/Target/X86/ |
| H A D | X86ShuffleDecodeConstantPool.cpp | 128 unsigned NumElts = Width / 8; in DecodePSHUFBMask() local 129 assert((NumElts == 16 || NumElts == 32 || NumElts == 64) && in DecodePSHUFBMask() 132 for (unsigned i = 0; i != NumElts; ++i) { in DecodePSHUFBMask() 167 unsigned NumElts = Width / ElSize; in DecodeVPERMILPMask() local 169 assert((NumElts == 2 || NumElts == 4 || NumElts == 8 || NumElts == 16) && in DecodeVPERMILPMask() 172 for (unsigned i = 0; i != NumElts; ++i) { in DecodeVPERMILPMask() 203 unsigned NumElts = Width / ElSize; in DecodeVPERMIL2PMask() local 205 assert((NumElts == 2 || NumElts == 4 || NumElts == 8) && in DecodeVPERMIL2PMask() 208 for (unsigned i = 0; i != NumElts; ++i) { in DecodeVPERMIL2PMask() 239 Index += Src * NumElts; in DecodeVPERMIL2PMask() [all …]
|
| H A D | X86PartialReduction.cpp | 183 unsigned NumElts = MulTy->getNumElements(); in tryMAddReplacement() local 188 SmallVector<int, 16> EvenMask(NumElts / 2); in tryMAddReplacement() 189 SmallVector<int, 16> OddMask(NumElts / 2); in tryMAddReplacement() 190 for (int i = 0, e = NumElts / 2; i != e; ++i) { in tryMAddReplacement() 202 SmallVector<int, 32> ConcatMask(NumElts); in tryMAddReplacement() 263 unsigned NumElts = OpTy->getNumElements(); in trySADReplacement() local 267 if (ST->hasBWI() && NumElts >= 64) { in trySADReplacement() 270 } else if (ST->hasAVX2() && NumElts >= 32) { in trySADReplacement() 280 if (NumElts < 16) { in trySADReplacement() 283 for (unsigned i = 0; i != NumElts; ++i) in trySADReplacement() [all …]
|
| H A D | X86ISelLowering.cpp | 3207 unsigned NumElts = std::min(DataVT.getVectorNumElements(), in getTgtMemIntrinsic() local 3209 Info.memVT = MVT::getVectorVT(DataVT.getVectorElementType(), NumElts); in getTgtMemIntrinsic() 3219 unsigned NumElts = std::min(DataVT.getVectorNumElements(), in getTgtMemIntrinsic() local 3221 Info.memVT = MVT::getVectorVT(DataVT.getVectorElementType(), NumElts); in getTgtMemIntrinsic() 3738 unsigned NumElts = Mask.size(); in isUndefLowerHalf() local 3739 return isUndefInRange(Mask, 0, NumElts / 2); in isUndefLowerHalf() 3744 unsigned NumElts = Mask.size(); in isUndefUpperHalf() local 3745 return isUndefInRange(Mask, NumElts / 2, NumElts / 2); in isUndefUpperHalf() 3793 unsigned NumElts = Mask.size(); in isBlendOrUndef() local 3795 if (!isUndefOrEqual(M, I) && !isUndefOrEqual(M, I + NumElts)) in isBlendOrUndef() [all …]
|
| /freebsd/contrib/llvm-project/llvm/lib/Bitstream/Reader/ |
| H A D | BitstreamReader.cpp | 106 unsigned NumElts = MaybeVBR.get(); in skipRecord() local 107 for (unsigned i = 0; i != NumElts; ++i) in skipRecord() 154 unsigned NumElts = MaybeNum.get(); in skipRecord() local 168 JumpToBit(GetCurrentBitNo() + static_cast<uint64_t>(NumElts) * in skipRecord() 174 for (; NumElts; --NumElts) in skipRecord() 182 if (Error Err = JumpToBit(GetCurrentBitNo() + NumElts * 6)) in skipRecord() 194 unsigned NumElts = MaybeNum.get(); in skipRecord() local 198 const size_t NewEnd = GetCurrentBitNo() + alignTo(NumElts, 4) * 8; in skipRecord() 227 uint32_t NumElts = MaybeNumElts.get(); in readRecord() local 228 if (!isSizePlausible(NumElts)) in readRecord() [all …]
|
| /freebsd/contrib/llvm-project/clang/lib/CodeGen/TargetBuiltins/ |
| H A D | X86.cpp | 59 unsigned NumElts) { in getMaskVecValue() argument 68 if (NumElts < 8) { in getMaskVecValue() 70 for (unsigned i = 0; i != NumElts; ++i) in getMaskVecValue() 73 MaskVec, MaskVec, ArrayRef(Indices, NumElts), "extract"); in getMaskVecValue() 141 unsigned NumElts = Ops[0]->getType()->getIntegerBitWidth(); in EmitX86MaskLogic() local 142 Value *LHS = getMaskVecValue(CGF, Ops[0], NumElts); in EmitX86MaskLogic() 143 Value *RHS = getMaskVecValue(CGF, Ops[1], NumElts); in EmitX86MaskLogic() 160 unsigned NumElts = cast<llvm::FixedVectorType>(Ty)->getNumElements(); in EmitX86FunnelShift() local 162 Amt = CGF.Builder.CreateVectorSplat(NumElts, Amt); in EmitX86FunnelShift() 239 unsigned NumElts, Value *MaskIn) { in EmitX86MaskedCompareResult() argument [all …]
|
| /freebsd/contrib/llvm-project/llvm/lib/Transforms/InstCombine/ |
| H A D | InstCombineVectorOps.cpp | 190 ElementCount NumElts = in foldBitcastExtElt() local 206 ExtIndexC = NumElts.getKnownMinValue() - 1 - ExtIndexC; in foldBitcastExtElt() 230 if (NumSrcElts == NumElts) in foldBitcastExtElt() 234 assert(NumSrcElts.isScalable() == NumElts.isScalable() && in foldBitcastExtElt() 239 if (NumSrcElts.getKnownMinValue() < NumElts.getKnownMinValue()) { in foldBitcastExtElt() 252 NumElts.getKnownMinValue() / NumSrcElts.getKnownMinValue(); in foldBitcastExtElt() 429 unsigned NumElts = EC.getKnownMinValue(); in visitExtractElementInst() local 430 HasKnownValidIndex = IndexC->getValue().ult(NumElts); in visitExtractElementInst() 436 if (IID == Intrinsic::stepvector && IndexC->getValue().ult(NumElts)) { in visitExtractElementInst() 452 if (!EC.isScalable() && IndexC->getValue().uge(NumElts)) in visitExtractElementInst() [all …]
|
| /freebsd/contrib/llvm-project/clang/lib/CodeGen/Targets/ |
| H A D | RISCV.cpp | 456 unsigned NumElts = STy->getStructNumElements(); in detectVLSCCEligibleStruct() local 457 if (NumElts > 8) in detectVLSCCEligibleStruct() 466 if (NumElts != 1) in detectVLSCCEligibleStruct() 469 NumElts = ArrayTy->getNumElements(); in detectVLSCCEligibleStruct() 477 if (NumElts * llvm::divideCeil( in detectVLSCCEligibleStruct() 485 if (NumElts == 1) { in detectVLSCCEligibleStruct() 510 NumElts); in detectVLSCCEligibleStruct() 525 unsigned NumElts = VT->getNumElements(); in coerceVLSVector() local 531 NumElts *= 2; in coerceVLSVector() 534 NumElts *= 4; in coerceVLSVector() [all …]
|
| /freebsd/contrib/llvm-project/llvm/lib/IR/ |
| H A D | AutoUpgrade.cpp | 1683 unsigned NumElts = ResultTy->getNumElements() * 8; in upgradeX86PSLLDQIntrinsics() local 1686 Type *VecTy = FixedVectorType::get(Builder.getInt8Ty(), NumElts); in upgradeX86PSLLDQIntrinsics() 1697 for (unsigned l = 0; l != NumElts; l += 16) in upgradeX86PSLLDQIntrinsics() 1699 unsigned Idx = NumElts + i - Shift; in upgradeX86PSLLDQIntrinsics() 1700 if (Idx < NumElts) in upgradeX86PSLLDQIntrinsics() 1701 Idx -= NumElts - 16; // end of lane, switch operand. in upgradeX86PSLLDQIntrinsics() 1705 Res = Builder.CreateShuffleVector(Res, Op, ArrayRef(Idxs, NumElts)); in upgradeX86PSLLDQIntrinsics() 1717 unsigned NumElts = ResultTy->getNumElements() * 8; in upgradeX86PSRLDQIntrinsics() local 1720 Type *VecTy = FixedVectorType::get(Builder.getInt8Ty(), NumElts); in upgradeX86PSRLDQIntrinsics() 1731 for (unsigned l = 0; l != NumElts; l += 16) in upgradeX86PSRLDQIntrinsics() [all …]
|
| /freebsd/contrib/llvm-project/llvm/lib/Target/AMDGPU/ |
| H A D | AMDGPUImageIntrinsicOptimizer.cpp | 206 unsigned NumElts = popcount(DMaskVal); in optimizeSection() local 211 unsigned NumMsaas = NumElts; in optimizeSection() 213 unsigned NumVDataLoads = divideCeil(NumElts, isD16 ? 2 : 1) * NumLoads; in optimizeSection() 255 if (NumElts == 1) { in optimizeSection() 260 for (unsigned I = 0; I < NumElts; ++I) { in optimizeSection()
|
| /freebsd/contrib/llvm-project/llvm/lib/CodeGen/SelectionDAG/ |
| H A D | LegalizeTypesGeneric.cpp | 344 unsigned NumElts = 2; in ExpandOp_BITCAST() local 348 NumElts); in ExpandOp_BITCAST() 352 NumElts = N->getValueType(0).getVectorNumElements(); in ExpandOp_BITCAST() 357 IntegerToVector(N->getOperand(0), NumElts, Ops, NVT.getVectorElementType()); in ExpandOp_BITCAST() 359 SDValue Vec = DAG.getBuildVector(NVT, dl, ArrayRef(Ops.data(), NumElts)); in ExpandOp_BITCAST() 370 unsigned NumElts = VecVT.getVectorNumElements(); in ExpandOp_BUILD_VECTOR() local 390 NewElts.reserve(NumElts*2); in ExpandOp_BUILD_VECTOR() 392 for (unsigned i = 0; i < NumElts; ++i) { in ExpandOp_BUILD_VECTOR() 428 unsigned NumElts = VecVT.getVectorNumElements(); in ExpandOp_INSERT_VECTOR_ELT() local 440 EVT NewVecVT = EVT::getVectorVT(*DAG.getContext(), NewEVT, NumElts*2); in ExpandOp_INSERT_VECTOR_ELT() [all …]
|
| /freebsd/contrib/llvm-project/llvm/lib/Target/AArch64/GISel/ |
| H A D | AArch64PostLegalizerLowering.cpp | 81 unsigned NumElts) { in getExtMask() argument 88 unsigned MaskBits = APInt(32, NumElts * 2).logBase2(); in getExtMask() 113 if (Imm < NumElts) in getExtMask() 116 Imm -= NumElts; in getExtMask() 166 unsigned NumElts = Ty.getNumElements(); in matchREV() local 170 if (isREVMask(ShuffleMask, EltSize, NumElts, LaneSize)) { in matchREV() 195 unsigned NumElts = MRI.getType(Dst).getNumElements(); in matchTRN() local 196 if (!isTRNMask(ShuffleMask, NumElts, WhichResult)) in matchTRN() 216 unsigned NumElts = MRI.getType(Dst).getNumElements(); in matchUZP() local 217 if (!isUZPMask(ShuffleMask, NumElts, WhichResult)) in matchUZP() [all …]
|
| /freebsd/contrib/llvm-project/llvm/lib/Target/NVPTX/ |
| H A D | NVPTXISelDAGToDAG.cpp | 1182 unsigned NumElts; in tryLDG() local 1185 NumElts = 1; in tryLDG() 1188 NumElts = getLoadStoreVectorNumElts(LD); in tryLDG() 1194 const unsigned FromTypeWidth = TotalWidth / NumElts; in tryLDG() 1243 unsigned NumElts; in tryLDU() local 1248 NumElts = 1; in tryLDU() 1251 NumElts = 2; in tryLDU() 1254 NumElts = 4; in tryLDU() 1259 MVT::getIntegerVT(LD->getMemoryVT().getSizeInBits() / NumElts).SimpleTy; in tryLDU() 1380 const unsigned NumElts = getLoadStoreVectorNumElts(ST); in tryStoreVector() local [all …]
|
| /freebsd/contrib/llvm-project/llvm/lib/SandboxIR/ |
| H A D | Type.cpp | 113 FixedVectorType *FixedVectorType::get(Type *ElementType, unsigned NumElts) { in get() argument 115 llvm::FixedVectorType::get(ElementType->LLVMTy, NumElts))); in get() 119 unsigned NumElts) { in get() argument 121 llvm::ScalableVectorType::get(ElementType->LLVMTy, NumElts))); in get()
|
| /freebsd/contrib/llvm-project/llvm/lib/Target/DirectX/ |
| H A D | CBufferDataLayout.cpp | 77 unsigned NumElts = AT->getNumElements(); in getTypeAllocSize() local 78 if (NumElts == 0) in getTypeAllocSize() 84 return TypeSize::getFixed(AlignedEltSize * (NumElts - 1) + EltSize); in getTypeAllocSize()
|
| H A D | DXILCBufferAccess.cpp | 30 unsigned int NumElts; member 40 NumElts = 8; in CBufferRowIntrin() 46 NumElts = 4; in CBufferRowIntrin() 52 NumElts = 2; in CBufferRowIntrin() 162 CurrentIndex %= Intrin.NumElts; in loadValue()
|
| /freebsd/contrib/llvm-project/llvm/lib/Analysis/ |
| H A D | VectorUtils.cpp | 496 bool llvm::isMaskedSlidePair(ArrayRef<int> Mask, int NumElts, in isMaskedSlidePair() argument 498 const int SignalValue = NumElts * 2; in isMaskedSlidePair() 504 int Src = M >= (int)NumElts; in isMaskedSlidePair() 505 int Diff = (int)i - (M % NumElts); in isMaskedSlidePair() 558 int NumElts = Mask.size(); in widenShuffleMaskElts() local 559 if (NumElts % Scale != 0) in widenShuffleMaskElts() 563 ScaledMask.reserve(NumElts / Scale); in widenShuffleMaskElts() 591 assert((int)ScaledMask.size() * Scale == NumElts && "Unexpected scaled mask"); in widenShuffleMaskElts() 600 unsigned NumElts = M.size(); in widenShuffleMaskElts() local 601 if (NumElts % 2 != 0) in widenShuffleMaskElts() [all …]
|
| /freebsd/contrib/llvm-project/llvm/include/llvm/ADT/ |
| H A D | SmallVector.h | 392 void growAndAssign(size_t NumElts, const T &Elt) { in growAndAssign() argument 395 T *NewElts = mallocForGrow(NumElts, NewCapacity); in growAndAssign() 396 std::uninitialized_fill_n(NewElts, NumElts, Elt); in growAndAssign() 399 this->set_size(NumElts); in growAndAssign() 544 void growAndAssign(size_t NumElts, T Elt) { in growAndAssign() argument 548 this->grow(NumElts); in growAndAssign() 549 std::uninitialized_fill_n(this->begin(), NumElts, Elt); in growAndAssign() 550 this->set_size(NumElts); in growAndAssign() 705 void assign(size_type NumElts, ValueParamT Elt) { in assign() argument 707 if (NumElts > this->capacity()) { in assign() [all …]
|
| /freebsd/contrib/llvm-project/llvm/include/llvm/Transforms/Vectorize/SandboxVectorizer/ |
| H A D | VecUtils.h | 114 static Type *getWideType(Type *ElemTy, unsigned NumElts) { 118 NumElts = VecTy->getNumElements() * NumElts; 120 return FixedVectorType::get(ElemTy, NumElts);
|
| /freebsd/contrib/llvm-project/llvm/lib/CodeGen/ |
| H A D | ExpandReductions.cpp | 101 unsigned NumElts = FTy->getNumElements(); in expandReductions() local 102 if (!isPowerOf2_32(NumElts)) in expandReductions() 106 Rdx = Builder.CreateBitCast(Vec, Builder.getIntNTy(NumElts)); in expandReductions()
|
| /freebsd/contrib/llvm-project/llvm/lib/Target/ARM/ |
| H A D | ARMTargetTransformInfo.cpp | 288 unsigned NumElts = cast<FixedVectorType>(II.getType())->getNumElements(); in simplifyDemandedVectorEltsIntrinsic() local 294 APInt::getSplat(NumElts, IsTop ? APInt::getLowBitsSet(2, 1) in simplifyDemandedVectorEltsIntrinsic() 298 UndefElts &= APInt::getSplat(NumElts, IsTop ? APInt::getLowBitsSet(2, 1) in simplifyDemandedVectorEltsIntrinsic() 1631 unsigned NumElts = cast<FixedVectorType>(VecTy)->getNumElements(); in getInterleavedMemoryOpCost() local 1633 FixedVectorType::get(VecTy->getScalarType(), NumElts / Factor); in getInterleavedMemoryOpCost() 1640 if (NumElts % Factor == 0 && in getInterleavedMemoryOpCost() 1649 if (ST->hasMVEIntegerOps() && Factor == 2 && NumElts / Factor > 2 && in getInterleavedMemoryOpCost() 1780 unsigned NumElts = cast<FixedVectorType>(ValTy)->getNumElements(); in getArithmeticReductionCost() local 1783 while (!TTI::requiresOrderedReduction(FMF) && isPowerOf2_32(NumElts) && in getArithmeticReductionCost() 1784 NumElts * EltSize > VecLimit) { in getArithmeticReductionCost() [all …]
|