/freebsd/contrib/llvm-project/llvm/lib/Target/AMDGPU/ |
H A D | AMDGPULowerKernelAttributes.cpp | 131 unsigned LoadSize = DL.getTypeStoreSize(Load->getType()); in processUse() local 137 if (LoadSize == 4) in processUse() 141 if (LoadSize == 4) in processUse() 145 if (LoadSize == 4) in processUse() 149 if (LoadSize == 2) in processUse() 153 if (LoadSize == 2) in processUse() 157 if (LoadSize == 2) in processUse() 161 if (LoadSize == 2) in processUse() 165 if (LoadSize == 2) in processUse() 169 if (LoadSize == 2) in processUse() [all …]
|
H A D | AMDGPUTargetTransformInfo.h | 124 unsigned getLoadVectorFactor(unsigned VF, unsigned LoadSize,
|
H A D | AMDGPUTargetTransformInfo.cpp | 345 unsigned GCNTTIImpl::getLoadVectorFactor(unsigned VF, unsigned LoadSize, in getLoadVectorFactor() argument 348 unsigned VecRegBitWidth = VF * LoadSize; in getLoadVectorFactor() 351 return 128 / LoadSize; in getLoadVectorFactor()
|
H A D | AMDGPURegisterBankInfo.cpp | 1061 unsigned LoadSize = LoadTy.getSizeInBits(); in applyMappingLoad() local 1069 if (LoadSize != 32 && (LoadSize != 96 || Subtarget.hasScalarDwordx3Loads())) in applyMappingLoad() 1078 if (LoadSize == 32 && in applyMappingLoad() 1082 if (LoadSize == 32 && in applyMappingLoad() 1093 if (LoadSize == 32) { in applyMappingLoad() 1135 if (LoadSize <= MaxNonSmrdLoadSize) in applyMappingLoad() 1144 assert(LoadSize % MaxNonSmrdLoadSize == 0); in applyMappingLoad() 1349 unsigned LoadSize = Ty.getSizeInBits(); in applyMappingSBufferLoad() local 1351 if (LoadSize == 256 || LoadSize == 512) { in applyMappingSBufferLoad() 1352 NumLoads = LoadSize / 128; in applyMappingSBufferLoad()
|
H A D | SIInstrInfo.cpp | 579 const unsigned LoadSize = NumBytes / ClusterSize; in shouldClusterMemOps() local 580 const unsigned NumDWORDs = ((LoadSize + 3) / 4) * ClusterSize; in shouldClusterMemOps()
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/Utils/ |
H A D | VNCoercion.cpp | 189 uint64_t LoadSize = DL.getTypeSizeInBits(LoadTy).getFixedValue(); in analyzeLoadFromClobberingWrite() local 191 if ((WriteSizeInBits & 7) | (LoadSize & 7)) in analyzeLoadFromClobberingWrite() 194 LoadSize /= 8; in analyzeLoadFromClobberingWrite() 201 StoreOffset + int64_t(StoreSize) < LoadOffset + int64_t(LoadSize)) in analyzeLoadFromClobberingWrite() 309 uint64_t LoadSize = (DL.getTypeSizeInBits(LoadTy).getFixedValue() + 7) / 8; in getStoreValueForLoadHelper() local 324 ShiftAmt = (StoreSize - LoadSize - Offset) * 8; in getStoreValueForLoadHelper() 329 if (LoadSize != StoreSize) in getStoreValueForLoadHelper() 331 IntegerType::get(Ctx, LoadSize * 8)); in getStoreValueForLoadHelper() 340 unsigned LoadSize = DL.getTypeStoreSize(LoadTy).getFixedValue(); in getValueForLoad() local 341 assert(Offset + LoadSize <= SrcValSize); in getValueForLoad() [all …]
|
/freebsd/contrib/llvm-project/llvm/lib/CodeGen/ |
H A D | ExpandMemCmp.cpp | 94 LoadEntry(unsigned LoadSize, uint64_t Offset) in LoadEntry() 95 : LoadSize(LoadSize), Offset(Offset) { in LoadEntry() 99 unsigned LoadSize; member 158 const unsigned LoadSize = LoadSizes.front(); in computeGreedyLoadSequence() local 159 const uint64_t NumLoadsForThisSize = Size / LoadSize; in computeGreedyLoadSequence() 169 LoadSequence.push_back({LoadSize, Offset}); in computeGreedyLoadSequence() 170 Offset += LoadSize; in computeGreedyLoadSequence() 172 if (LoadSize > 1) in computeGreedyLoadSequence() 174 Size = Size % LoadSize; in computeGreedyLoadSequence() 237 if (PreLast.Offset + PreLast.LoadSize != Last.Offset) in optimiseLoadSequence() [all …]
|
H A D | StackSlotColoring.cpp | 477 unsigned LoadSize = 0; in RemoveDeadStores() local 479 if (!(LoadReg = TII->isLoadFromStackSlot(*I, FirstSS, LoadSize))) in RemoveDeadStores() 490 LoadSize != StoreSize || !MFI->isSpillSlotObjectIndex(FirstSS)) in RemoveDeadStores()
|
/freebsd/contrib/llvm-project/llvm/include/llvm/ExecutionEngine/Orc/Shared/ |
H A D | VTuneSharedStructs.h | 34 uint64_t LoadSize; member 67 MI.LineTable, MI.LoadAddr, MI.LoadSize, MI.MethodID, MI.NameSI, in size() 73 IB, MI.LineTable, MI.LoadAddr, MI.LoadSize, MI.MethodID, MI.NameSI, in deserialize() 79 OB, MI.LineTable, MI.LoadAddr, MI.LoadSize, MI.MethodID, MI.NameSI, in serialize()
|
/freebsd/contrib/llvm-project/llvm/lib/Analysis/ |
H A D | Loads.cpp | 376 const TypeSize LoadSize = TypeSize::getFixed(Size.getZExtValue()); in isSafeToLoadUnconditionally() local 426 TypeSize::isKnownLE(LoadSize, DL.getTypeStoreSize(AccessedTy))) in isSafeToLoadUnconditionally() 430 TypeSize::isKnownLE(LoadSize, DL.getTypeStoreSize(AccessedTy))) in isSafeToLoadUnconditionally() 545 TypeSize LoadSize = DL.getTypeSizeInBits(AccessTy); in getAvailableLoadStore() local 546 if (TypeSize::isKnownLE(LoadSize, StoreSize)) in getAvailableLoadStore() 575 uint64_t LoadSize = LoadTypeSize.getFixedValue(); in getAvailableLoadStore() local 576 if ((Len->getValue() * 8).ult(LoadSize)) in getAvailableLoadStore() 579 APInt Splat = LoadSize >= 8 ? APInt::getSplat(LoadSize, Val->getValue()) in getAvailableLoadStore() 580 : Val->getValue().trunc(LoadSize); in getAvailableLoadStore()
|
H A D | TargetTransformInfo.cpp | 1280 unsigned LoadSize, in getLoadVectorFactor() argument 1283 return TTIImpl->getLoadVectorFactor(VF, LoadSize, ChainSizeInBytes, VecTy); in getLoadVectorFactor()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/AArch64/ |
H A D | AArch64LoadStoreOptimizer.cpp | 632 int LoadSize = TII->getMemScale(LoadInst); in isLdOffsetInRangeOfSt() local 641 : AArch64InstrInfo::getLdStOffsetOp(LoadInst).getImm() * LoadSize; in isLdOffsetInRangeOfSt() 643 (UnscaledLdOffset + LoadSize <= (UnscaledStOffset + StoreSize)); in isLdOffsetInRangeOfSt() 1144 int LoadSize = TII->getMemScale(*LoadI); in promoteLoadFromStore() local 1156 if (LoadSize == StoreSize && (LoadSize == 4 || LoadSize == 8)) { in promoteLoadFromStore() 1159 if (StRt == LdRt && LoadSize == 8) { in promoteLoadFromStore() 1189 assert(LoadSize <= StoreSize && "Invalid load size"); in promoteLoadFromStore() 1193 : AArch64InstrInfo::getLdStOffsetOp(*LoadI).getImm() * LoadSize; in promoteLoadFromStore() 1198 int Width = LoadSize * 8; in promoteLoadFromStore() 1205 (UnscaledLdOffset + LoadSize) <= UnscaledStOffset + StoreSize) && in promoteLoadFromStore()
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/AggressiveInstCombine/ |
H A D | AggressiveInstCombine.cpp | 623 uint64_t LoadSize = 0; member 704 Loc = Loc.getWithNewSize(LOps.LoadSize); in foldLoadsRecursive() 745 LoadSize1 = LOps.LoadSize; in foldLoadsRecursive() 747 LoadSize2 = LOps.LoadSize; in foldLoadsRecursive() 765 LOps.LoadSize = LoadSize1 + LoadSize2; in foldLoadsRecursive() 794 IntegerType *WiderType = IntegerType::get(I.getContext(), LOps.LoadSize); in foldConsecutiveLoads() 802 Allowed = TTI.allowsMisalignedMemoryAccesses(I.getContext(), LOps.LoadSize, in foldConsecutiveLoads()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/PowerPC/ |
H A D | PPCHazardRecognizers.h | 94 bool isLoadOfStoredAddress(uint64_t LoadSize, int64_t LoadOffset,
|
H A D | PPCHazardRecognizers.cpp | 298 isLoadOfStoredAddress(uint64_t LoadSize, int64_t LoadOffset, in isLoadOfStoredAddress() argument 313 if (int64_t(LoadOffset+LoadSize) > StoreOffset[i]) return true; in isLoadOfStoredAddress()
|
/freebsd/contrib/llvm-project/llvm/lib/ExecutionEngine/Orc/TargetProcess/ |
H A D | JITLoaderVTune.cpp | 59 MethodMessage.method_size = MethodInfo.LoadSize; in registerJITLoaderVTuneRegisterImpl()
|
/freebsd/contrib/llvm-project/llvm/lib/ExecutionEngine/Orc/Debugging/ |
H A D | VTuneSupportPlugin.cpp | 58 Method.LoadSize = Sym->getSize(); in getMethodBatch()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/X86/ |
H A D | X86AvoidStoreForwardingBlocks.cpp | 619 static bool isBlockingStore(int64_t LoadDispImm, unsigned LoadSize, in isBlockingStore() argument 622 (StoreDispImm <= LoadDispImm + (LoadSize - StoreSize))); in isBlockingStore()
|
/freebsd/contrib/llvm-project/llvm/include/llvm/Analysis/ |
H A D | TargetTransformInfo.h | 1668 unsigned getLoadVectorFactor(unsigned VF, unsigned LoadSize, 2165 virtual unsigned getLoadVectorFactor(unsigned VF, unsigned LoadSize, 2900 unsigned getLoadVectorFactor(unsigned VF, unsigned LoadSize, in getLoadVectorFactor() argument 2903 return Impl.getLoadVectorFactor(VF, LoadSize, ChainSizeInBytes, VecTy); in getLoadVectorFactor()
|
H A D | TargetTransformInfoImpl.h | 913 unsigned getLoadVectorFactor(unsigned VF, unsigned LoadSize, in getLoadVectorFactor() argument
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/Scalar/ |
H A D | LoopIdiomRecognize.cpp | 1220 int64_t LoadSize = in loadAndStoreMayFormMemmove() local 1222 if (BP1 != BP2 || LoadSize != int64_t(StoreSize)) in loadAndStoreMayFormMemmove() 1225 (IsNegStride && LoadOff + LoadSize > StoreOff)) in loadAndStoreMayFormMemmove()
|
H A D | SROA.cpp | 1292 TypeSize LoadSize = DL.getTypeStoreSize(LI->getType()); in hasUnsafePHIOrSelectUse() local 1293 if (LoadSize.isScalable()) { in hasUnsafePHIOrSelectUse() 1297 Size = std::max(Size, LoadSize.getFixedValue()); in hasUnsafePHIOrSelectUse() 1570 APInt LoadSize = in isSafePHIToSpeculate() local 1594 if (isSafeToLoadUnconditionally(InVal, MaxAlign, LoadSize, DL, TI)) in isSafePHIToSpeculate()
|
/freebsd/contrib/llvm-project/clang/lib/CodeGen/ |
H A D | CGBuiltin.cpp | 471 CharUnits LoadSize = CGF.getContext().getTypeSizeInChars(ElTy); in EmitISOVolatileLoad() local 473 llvm::IntegerType::get(CGF.getLLVMContext(), LoadSize.getQuantity() * 8); in EmitISOVolatileLoad() 474 llvm::LoadInst *Load = CGF.Builder.CreateAlignedLoad(ITy, Ptr, LoadSize); in EmitISOVolatileLoad()
|