| /freebsd/contrib/llvm-project/llvm/include/llvm/Transforms/Utils/ |
| H A D | VNCoercion.h | 55 int analyzeLoadFromClobberingStore(Type *LoadTy, Value *LoadPtr, 63 int analyzeLoadFromClobberingLoad(Type *LoadTy, Value *LoadPtr, LoadInst *DepLI, 71 int analyzeLoadFromClobberingMemInst(Type *LoadTy, Value *LoadPtr,
|
| /freebsd/contrib/llvm-project/llvm/lib/Transforms/Utils/ |
| H A D | VNCoercion.cpp | 196 static int analyzeLoadFromClobberingWrite(Type *LoadTy, Value *LoadPtr, in analyzeLoadFromClobberingWrite() argument 208 Value *LoadBase = GetPointerBaseWithConstantOffset(LoadPtr, LoadOffset, DL); in analyzeLoadFromClobberingWrite() 234 int analyzeLoadFromClobberingStore(Type *LoadTy, Value *LoadPtr, in analyzeLoadFromClobberingStore() argument 248 return analyzeLoadFromClobberingWrite(LoadTy, LoadPtr, StorePtr, StoreSize, in analyzeLoadFromClobberingStore() 255 int analyzeLoadFromClobberingLoad(Type *LoadTy, Value *LoadPtr, LoadInst *DepLI, in analyzeLoadFromClobberingLoad() argument 266 return analyzeLoadFromClobberingWrite(LoadTy, LoadPtr, DepPtr, DepSize, DL); in analyzeLoadFromClobberingLoad() 269 int analyzeLoadFromClobberingMemInst(Type *LoadTy, Value *LoadPtr, in analyzeLoadFromClobberingMemInst() argument 285 return analyzeLoadFromClobberingWrite(LoadTy, LoadPtr, MI->getDest(), in analyzeLoadFromClobberingMemInst() 303 int Offset = analyzeLoadFromClobberingWrite(LoadTy, LoadPtr, MI->getDest(), in analyzeLoadFromClobberingMemInst()
|
| /freebsd/contrib/llvm-project/llvm/lib/Transforms/Scalar/ |
| H A D | LoopLoadElimination.cpp | 94 Value *LoadPtr = Load->getPointerOperand(); in isDependenceDistanceOfOne() local 99 assert(LoadPtr->getType()->getPointerAddressSpace() == in isDependenceDistanceOfOne() 105 int64_t StrideLoad = getPtrStride(PSE, LoadType, LoadPtr, L).value_or(0); in isDependenceDistanceOfOne() 122 auto *LoadPtrSCEV = cast<SCEVAddRecExpr>(PSE.getSCEV(LoadPtr)); in isDependenceDistanceOfOne()
|
| H A D | GVN.cpp | 1573 Value *LoadPtr = AvailableLoad.second; in eliminatePartiallyRedundantLoad() local 1576 Load->getType(), LoadPtr, Load->getName() + ".pre", Load->isVolatile(), in eliminatePartiallyRedundantLoad() 1613 MD->invalidateCachedPointerInfo(LoadPtr); in eliminatePartiallyRedundantLoad() 1833 Value *LoadPtr = Load->getPointerOperand(); in PerformLoadPRE() local 1836 PHITransAddr Address(LoadPtr, DL, AC); in PerformLoadPRE() 1837 LoadPtr = Address.translateWithInsertion(Cur, Cur->getSinglePredecessor(), in PerformLoadPRE() 1839 if (!LoadPtr) { in PerformLoadPRE() 1846 if (LoadPtr) { in PerformLoadPRE() 1847 PHITransAddr Address(LoadPtr, DL, AC); in PerformLoadPRE() 1848 LoadPtr = Address.translateWithInsertion(LoadBB, UnavailablePred, *DT, in PerformLoadPRE() [all …]
|
| H A D | NewGVN.cpp | 1486 NewGVN::performSymbolicLoadCoercion(Type *LoadType, Value *LoadPtr, in performSymbolicLoadCoercion() argument 1497 int Offset = analyzeLoadFromClobberingStore(LoadType, LoadPtr, DepSI, DL); in performSymbolicLoadCoercion() 1512 int Offset = analyzeLoadFromClobberingLoad(LoadType, LoadPtr, DepLI, DL); in performSymbolicLoadCoercion() 1524 int Offset = analyzeLoadFromClobberingMemInst(LoadType, LoadPtr, DepMI, DL); in performSymbolicLoadCoercion() 1538 if (LoadPtr == lookupOperandLeader(LifetimePtr) || in performSymbolicLoadCoercion() 1539 AA->isMustAlias(LoadPtr, LifetimePtr)) in performSymbolicLoadCoercion() 1547 (LoadPtr != lookupOperandLeader(DepInst) && in performSymbolicLoadCoercion() 1548 !AA->isMustAlias(LoadPtr, DepInst))) in performSymbolicLoadCoercion()
|
| H A D | LoopIdiomRecognize.cpp | 1217 Value *LoadPtr = LI->getPointerOperand(); in processLoopStoreOfLoopLoad() local 1218 const SCEVAddRecExpr *LoadEv = cast<SCEVAddRecExpr>(SE->getSCEV(LoadPtr)); in processLoopStoreOfLoopLoad() 1221 return processLoopStoreOfLoopLoad(StorePtr, LoadPtr, StoreSizeSCEV, in processLoopStoreOfLoopLoad()
|
| /freebsd/contrib/llvm-project/llvm/lib/Analysis/ |
| H A D | Loads.cpp | 554 static bool areNonOverlapSameBaseLoadAndStore(const Value *LoadPtr, in areNonOverlapSameBaseLoadAndStore() argument 559 APInt LoadOffset(DL.getIndexTypeSizeInBits(LoadPtr->getType()), 0); in areNonOverlapSameBaseLoadAndStore() 561 const Value *LoadBase = LoadPtr->stripAndAccumulateConstantOffsets( in areNonOverlapSameBaseLoadAndStore() 588 Value *LoadPtr = LI->getPointerOperand()->stripPointerCasts(); in getAvailableLoadStore() local 589 if (!AreEquivalentAddressValues(LoadPtr, Ptr)) in getAvailableLoadStore()
|
| /freebsd/contrib/llvm-project/llvm/lib/FuzzMutate/ |
| H A D | RandomIRBuilder.cpp | 145 Value *LoadPtr = buildTargetLegalPtr(M, Ptr, IP, LoadName, &NewInsts); in buildTargetLegalLoad() local 147 Instruction *Load = new LoadInst(AccessTy, LoadPtr, LoadName, IP); in buildTargetLegalLoad()
|
| /freebsd/contrib/llvm-project/llvm/lib/Target/AMDGPU/ |
| H A D | R600ISelLowering.cpp | 1045 SDValue LoadPtr = BasePtr; in lowerPrivateTruncStore() local 1047 LoadPtr = DAG.getNode(ISD::ADD, DL, MVT::i32, BasePtr, Offset); in lowerPrivateTruncStore() 1052 SDValue Ptr = DAG.getNode(ISD::AND, DL, MVT::i32, LoadPtr, in lowerPrivateTruncStore() 1063 SDValue ByteIdx = DAG.getNode(ISD::AND, DL, MVT::i32, LoadPtr, in lowerPrivateTruncStore() 1277 SDValue LoadPtr = BasePtr; in lowerPrivateExtLoad() local 1279 LoadPtr = DAG.getNode(ISD::ADD, DL, MVT::i32, BasePtr, Offset); in lowerPrivateExtLoad() 1284 SDValue Ptr = DAG.getNode(ISD::AND, DL, MVT::i32, LoadPtr, in lowerPrivateExtLoad() 1294 LoadPtr, DAG.getConstant(0x3, DL, MVT::i32)); in lowerPrivateExtLoad()
|
| /freebsd/contrib/llvm-project/compiler-rt/lib/lsan/ |
| H A D | lsan_common.cpp | 295 void *LoadPtr(uptr p) const { return *reinterpret_cast<void **>(p); } in LoadPtr() function 307 void *LoadPtr(uptr p) const { in LoadPtr() function 339 void *p = accessor.LoadPtr(pp); in ScanForPointers()
|
| /freebsd/contrib/llvm-project/llvm/lib/Transforms/InstCombine/ |
| H A D | InstCombineCalls.cpp | 292 Value *LoadPtr = II.getArgOperand(0); in simplifyMaskedLoad() local 299 LoadInst *L = Builder.CreateAlignedLoad(II.getType(), LoadPtr, Alignment, in simplifyMaskedLoad() 307 if (isDereferenceablePointer(LoadPtr, II.getType(), in simplifyMaskedLoad() 309 LoadInst *LI = Builder.CreateAlignedLoad(II.getType(), LoadPtr, Alignment, in simplifyMaskedLoad()
|
| /freebsd/contrib/llvm-project/llvm/lib/Target/Hexagon/ |
| H A D | HexagonLoopIdiomRecognition.cpp | 1947 Value *LoadPtr = LI->getPointerOperand(); in isLegalStore() local 1948 auto *LoadEv = dyn_cast<SCEVAddRecExpr>(SE->getSCEV(LoadPtr)); in isLegalStore()
|
| /freebsd/contrib/llvm-project/llvm/lib/Target/Mips/ |
| H A D | MipsISelLowering.cpp | 4613 SDValue LoadPtr = DAG.getNode(ISD::ADD, DL, PtrTy, Arg, in passByValArg() local 4615 SDValue LoadVal = DAG.getLoad(RegTy, DL, Chain, LoadPtr, in passByValArg() 4638 SDValue LoadPtr = DAG.getNode(ISD::ADD, DL, PtrTy, Arg, in passByValArg() local 4642 ISD::ZEXTLOAD, DL, RegTy, Chain, LoadPtr, MachinePointerInfo(), in passByValArg()
|
| /freebsd/contrib/llvm-project/llvm/lib/CodeGen/GlobalISel/ |
| H A D | CombinerHelper.cpp | 3986 Register LoadPtr; in findLoadOffsetsForLoadOrCombine() local 3989 m_GPtrAdd(m_Reg(LoadPtr), m_ICst(Idx)))) { in findLoadOffsetsForLoadOrCombine() 3990 LoadPtr = Load->getOperand(1).getReg(); in findLoadOffsetsForLoadOrCombine() 4002 BasePtr = LoadPtr; in findLoadOffsetsForLoadOrCombine() 4003 if (BasePtr != LoadPtr) in findLoadOffsetsForLoadOrCombine()
|
| H A D | LegalizerHelper.cpp | 9959 Register LoadPtr = Src; in lowerMemcpy() local 9965 LoadPtr = MIB.buildPtrAdd(SrcTy, Src, Offset).getReg(0); in lowerMemcpy() 9967 auto LdVal = MIB.buildLoad(CopyTy, LoadPtr, *LoadMMO); in lowerMemcpy() 10058 Register LoadPtr = Src; in lowerMemmove() local 10063 LoadPtr = MIB.buildPtrAdd(SrcTy, Src, Offset).getReg(0); in lowerMemmove() 10065 LoadVals.push_back(MIB.buildLoad(CopyTy, LoadPtr, *LoadMMO).getReg(0)); in lowerMemmove()
|