| /freebsd/contrib/llvm-project/llvm/lib/Target/AMDGPU/ |
| H A D | AMDGPUImageIntrinsicOptimizer.cpp | 210 unsigned NumLoads = IIList.size(); in optimizeSection() local 212 unsigned NumVAddrLoads = 3 * NumLoads; in optimizeSection() 213 unsigned NumVDataLoads = divideCeil(NumElts, isD16 ? 2 : 1) * NumLoads; in optimizeSection() 217 if (NumLoads < NumMsaas || in optimizeSection()
|
| H A D | AMDGPURegisterBankInfo.cpp | 1378 int NumLoads = 1; in applyMappingSBufferLoad() local 1380 NumLoads = LoadSize / 128; in applyMappingSBufferLoad() 1381 Ty = Ty.divide(NumLoads); in applyMappingSBufferLoad() 1386 const Align Alignment = NumLoads > 1 ? Align(16 * NumLoads) : Align(1); in applyMappingSBufferLoad() 1416 SmallVector<Register, 4> LoadParts(NumLoads); in applyMappingSBufferLoad() 1421 for (int i = 0; i < NumLoads; ++i) { in applyMappingSBufferLoad() 1422 if (NumLoads == 1) { in applyMappingSBufferLoad() 1461 if (NumLoads != 1) { in applyMappingSBufferLoad()
|
| H A D | SIInstrInfo.h | 265 int64_t Offset1, unsigned NumLoads) const override;
|
| H A D | SIISelLowering.cpp | 8919 unsigned NumLoads = 1; in lowerSBuffer() local 8926 NumLoads = NumElts / 4; in lowerSBuffer() 8935 NumLoads > 1 ? Align(16 * NumLoads) : Align(4)); in lowerSBuffer() 8938 for (unsigned i = 0; i < NumLoads; ++i) { in lowerSBuffer()
|
| H A D | SIInstrInfo.cpp | 613 unsigned NumLoads) const { in shouldScheduleLoadsNear() 620 return (NumLoads <= 16 && (Offset1 - Offset0) < 64); in shouldScheduleLoadsNear()
|
| /freebsd/contrib/llvm-project/llvm/lib/Target/Hexagon/ |
| H A D | HexagonTargetTransformInfo.cpp | 196 unsigned NumLoads = alignTo(VecWidth, AlignWidth) / AlignWidth; in getMemoryOpCost() local 197 return 3 * NumLoads; in getMemoryOpCost() 208 unsigned NumLoads = alignTo(VecWidth, AlignWidth) / AlignWidth; in getMemoryOpCost() local 210 return Cost * NumLoads; in getMemoryOpCost() 214 return (3 - LogA) * Cost * NumLoads; in getMemoryOpCost()
|
| /freebsd/contrib/llvm-project/llvm/lib/Target/NVPTX/ |
| H A D | NVPTXLowerAggrCopies.cpp | 107 unsigned NumLoads = DL.getTypeStoreSize(LI->getType()); in runOnFunction() local 109 ConstantInt::get(Type::getInt32Ty(Context), NumLoads); in runOnFunction()
|
| /freebsd/contrib/llvm-project/llvm/lib/Target/SystemZ/ |
| H A D | SystemZTargetTransformInfo.cpp | 58 unsigned &NumLoads, const Function *F) { in countNumMemAccesses() argument 69 NumLoads++; in countNumMemAccesses() 72 countNumMemAccesses(GEP, NumStores, NumLoads, F); in countNumMemAccesses() 114 unsigned NumStores = 0, NumLoads = 0; in adjustInliningThreshold() local 115 countNumMemAccesses(GEP, NumStores, NumLoads, Callee); in adjustInliningThreshold() 116 Ptr2NumUses[GV] += NumLoads + NumStores; in adjustInliningThreshold() 133 unsigned NumLoads = 0; in adjustInliningThreshold() local 138 countNumMemAccesses(CalleeArg, NumStores, NumLoads, Callee); in adjustInliningThreshold() 140 if (NumLoads > 10) in adjustInliningThreshold() 141 Bonus += NumLoads * 50; in adjustInliningThreshold()
|
| /freebsd/contrib/llvm-project/llvm/lib/Target/X86/ |
| H A D | X86InterleavedAccess.cpp | 194 unsigned int NumLoads = NumSubVectors; in decompose() local 202 NumLoads = NumSubVectors * (VecLength / 384); in decompose() 213 for (unsigned i = 0; i < NumLoads; i++) { in decompose()
|
| H A D | X86FastPreTileConfig.cpp | 39 STATISTIC(NumLoads, "Number of loads added"); 265 ++NumLoads; in reload()
|
| H A D | X86InstrInfo.h | 544 unsigned NumLoads) const override;
|
| H A D | X86InstrInfo.cpp | 8874 unsigned NumLoads) const { in shouldScheduleLoadsNear() 8901 if (NumLoads >= 3) in shouldScheduleLoadsNear() 8903 } else if (NumLoads) { in shouldScheduleLoadsNear() 8913 if (NumLoads) in shouldScheduleLoadsNear()
|
| /freebsd/contrib/llvm-project/llvm/lib/CodeGen/ |
| H A D | ExpandMemCmp.cpp | 418 const unsigned NumLoads = in getCompareLoadPairs() local 432 NumLoads == 1 ? nullptr in getCompareLoadPairs() 435 for (unsigned i = 0; i < NumLoads; ++i, ++LoadIndex) { in getCompareLoadPairs() 441 if (NumLoads != 1) { in getCompareLoadPairs()
|
| H A D | RegAllocFast.cpp | 54 STATISTIC(NumLoads, "Number of loads added"); 649 ++NumLoads; in reload()
|
| /freebsd/contrib/llvm-project/llvm/lib/CodeGen/SelectionDAG/ |
| H A D | ScheduleDAGSDNodes.cpp | 274 unsigned NumLoads = 0; in ClusterNeighboringLoads() local 281 if (!TII->shouldScheduleLoadsNear(BaseLoad, Load, BaseOff, Offset,NumLoads)) in ClusterNeighboringLoads() 284 ++NumLoads; in ClusterNeighboringLoads() 287 if (NumLoads == 0) in ClusterNeighboringLoads()
|
| /freebsd/contrib/llvm-project/llvm/include/llvm/Analysis/ |
| H A D | LoopAccessAnalysis.h | 713 unsigned getNumLoads() const { return NumLoads;} in getNumLoads() 813 unsigned NumLoads = 0; variable
|
| /freebsd/contrib/llvm-project/llvm/lib/Transforms/Scalar/ |
| H A D | LowerMatrixIntrinsics.cpp | 379 unsigned NumLoads = 0; member 389 NumLoads += RHS.NumLoads; in operator +=() 483 OpInfo.NumLoads += N; in addNumLoads() 487 void setNumLoads(unsigned N) { OpInfo.NumLoads = N; } in setNumLoads() 505 unsigned getNumLoads() const { return OpInfo.NumLoads; } in getNumLoads() 2791 << ore::NV("NumLoads", Counts.NumLoads) << " loads, " in emitRemarks() 2797 if (SharedCounts.NumStores > 0 || SharedCounts.NumLoads > 0 || in emitRemarks() 2801 << ore::NV("NumLoads", SharedCounts.NumLoads) << " loads, " in emitRemarks()
|
| /freebsd/contrib/llvm-project/llvm/lib/Target/RISCV/ |
| H A D | RISCVTargetTransformInfo.cpp | 1013 unsigned NumLoads = getEstimatedVLFor(VTy); in getInterleavedMemoryOpCost() local 1014 return NumLoads * MemOpCost; in getInterleavedMemoryOpCost() 1095 unsigned NumLoads = getEstimatedVLFor(&VTy); in getGatherScatterOpCost() local 1096 return NumLoads * MemOpCost; in getGatherScatterOpCost() 1156 unsigned NumLoads = getEstimatedVLFor(&VTy); in getStridedMemoryOpCost() local 1157 return NumLoads * MemOpCost; in getStridedMemoryOpCost()
|
| /freebsd/contrib/llvm-project/llvm/lib/Target/ARM/ |
| H A D | ARMBaseInstrInfo.h | 262 unsigned NumLoads) const override;
|
| H A D | ARMBaseInstrInfo.cpp | 1860 unsigned NumLoads) const { in shouldScheduleLoadsNear() 1882 if (NumLoads >= 3) in shouldScheduleLoadsNear()
|
| /freebsd/contrib/llvm-project/llvm/include/llvm/CodeGen/ |
| H A D | TargetInstrInfo.h | 1526 unsigned NumLoads) const { in shouldScheduleLoadsNear() argument
|
| /freebsd/contrib/llvm-project/llvm/utils/TableGen/Common/ |
| H A D | CodeGenDAGPatterns.cpp | 4196 unsigned NumLoads = 0; in VerifyInstructionFlags() local 4201 NumLoads += InstInfo.mayLoad; in VerifyInstructionFlags() 4222 if (!PatInfo.mayStore && PatInfo.mayLoad && !NumLoads) in VerifyInstructionFlags()
|
| /freebsd/contrib/llvm-project/llvm/lib/Analysis/ |
| H A D | LoopAccessAnalysis.cpp | 2544 NumLoads++; in analyzeLoop()
|
| /freebsd/contrib/llvm-project/llvm/lib/Transforms/Vectorize/ |
| H A D | LoopVectorize.cpp | 4803 unsigned NumLoads = Legal->getNumLoads(); in selectInterleaveCount() local 4805 unsigned LoadsIC = IC / (NumLoads ? NumLoads : 1); in selectInterleaveCount()
|
| /freebsd/contrib/llvm-project/llvm/lib/Target/AArch64/ |
| H A D | AArch64ISelLowering.cpp | 17177 unsigned NumLoads = getNumInterleavedAccesses(VTy, DL, UseScalable); in lowerInterleavedLoad() local 17191 FVTy->getNumElements() / NumLoads); in lowerInterleavedLoad() 17228 for (unsigned LoadCount = 0; LoadCount < NumLoads; ++LoadCount) { in lowerInterleavedLoad() 17500 unsigned NumLoads = getNumInterleavedAccesses(VTy, DL, UseScalable); in lowerDeinterleaveIntrinsicToLoad() local 17503 VTy->getElementCount().divideCoefficientBy(NumLoads)); in lowerDeinterleaveIntrinsicToLoad() 17516 if (NumLoads > 1) { in lowerDeinterleaveIntrinsicToLoad() 17519 for (unsigned I = 0; I < NumLoads; ++I) { in lowerDeinterleaveIntrinsicToLoad()
|