| /freebsd/contrib/llvm-project/llvm/lib/Transforms/Scalar/ |
| H A D | MergeICmps.cpp | 76 BCEAtom(GetElementPtrInst *GEP, LoadInst *LoadI, int BaseId, APInt Offset) in BCEAtom() 77 : GEP(GEP), LoadI(LoadI), BaseId(BaseId), Offset(std::move(Offset)) {} in BCEAtom() 87 LoadI = that.LoadI; in operator =() 108 LoadInst *LoadI = nullptr; member 136 auto *const LoadI = dyn_cast<LoadInst>(Val); in visitICmpLoadOperand() local 137 if (!LoadI) in visitICmpLoadOperand() 140 if (LoadI->isUsedOutsideOfBlock(LoadI->getParent())) { in visitICmpLoadOperand() 145 if (!LoadI->isSimple()) { in visitICmpLoadOperand() 149 Value *Addr = LoadI->getOperand(0); in visitICmpLoadOperand() 154 const auto &DL = LoadI->getDataLayout(); in visitICmpLoadOperand() [all …]
|
| H A D | JumpThreading.cpp | 1109 if (LoadInst *LoadI = dyn_cast<LoadInst>(SimplifyValue)) in processBlock() local 1110 if (simplifyPartiallyRedundantLoad(LoadI)) in processBlock() 1220 bool JumpThreadingPass::simplifyPartiallyRedundantLoad(LoadInst *LoadI) { in simplifyPartiallyRedundantLoad() argument 1222 if (!LoadI->isUnordered()) return false; in simplifyPartiallyRedundantLoad() 1226 BasicBlock *LoadBB = LoadI->getParent(); in simplifyPartiallyRedundantLoad() 1236 Value *LoadedPtr = LoadI->getOperand(0); in simplifyPartiallyRedundantLoad() 1245 BasicBlock::iterator BBIt(LoadI); in simplifyPartiallyRedundantLoad() 1251 LoadI, LoadBB, BBIt, DefMaxInstsToScan, &BatchAA, &IsLoadCSE)) { in simplifyPartiallyRedundantLoad() 1257 combineMetadataForCSE(NLoadI, LoadI, false); in simplifyPartiallyRedundantLoad() 1263 if (AvailableVal == LoadI) in simplifyPartiallyRedundantLoad() [all …]
|
| H A D | DeadStoreElimination.cpp | 2195 if (auto *LoadI = dyn_cast<LoadInst>(Store->getOperand(0))) { in storeIsNoop() local 2196 if (LoadI->getPointerOperand() == Store->getOperand(1)) { in storeIsNoop() 2198 auto *LoadAccess = MSSA.getMemoryAccess(LoadI)->getDefiningAccess(); in storeIsNoop()
|
| /freebsd/contrib/llvm-project/llvm/lib/Target/AArch64/ |
| H A D | AArch64LoadStoreOptimizer.cpp | 170 promoteLoadFromStore(MachineBasicBlock::iterator LoadI, 1396 AArch64LoadStoreOpt::promoteLoadFromStore(MachineBasicBlock::iterator LoadI, in promoteLoadFromStore() argument 1399 next_nodbg(LoadI, LoadI->getParent()->end()); in promoteLoadFromStore() 1401 int LoadSize = TII->getMemScale(*LoadI); in promoteLoadFromStore() 1403 Register LdRt = getLdStRegOp(*LoadI).getReg(); in promoteLoadFromStore() 1418 LoadI->getIterator())) { in promoteLoadFromStore() 1425 LLVM_DEBUG(LoadI->print(dbgs())); in promoteLoadFromStore() 1427 LoadI->eraseFromParent(); in promoteLoadFromStore() 1432 BuildMI(*LoadI->getParent(), LoadI, LoadI->getDebugLoc(), in promoteLoadFromStore() 1437 .setMIFlags(LoadI->getFlags()); in promoteLoadFromStore() [all …]
|
| H A D | AArch64FalkorHWPFFix.cpp | 150 LoadInst *LoadI = dyn_cast<LoadInst>(&I); in runOnLoop() local 151 if (!LoadI) in runOnLoop() 154 Value *PtrValue = LoadI->getPointerOperand(); in runOnLoop() 163 LoadI->setMetadata(FALKOR_STRIDED_ACCESS_MD, in runOnLoop() 164 MDNode::get(LoadI->getContext(), {})); in runOnLoop()
|
| /freebsd/contrib/llvm-project/llvm/lib/Transforms/InstCombine/ |
| H A D | InstCombinePHI.cpp | 148 if (LoadInst *LoadI = dyn_cast<LoadInst>(U)) { in foldIntegerTypedPHI() local 149 Ptr = LoadI->getPointerOperand(); in foldIntegerTypedPHI() 208 auto *LoadI = dyn_cast<LoadInst>(Arg); in foldIntegerTypedPHI() local 209 if (!LoadI) in foldIntegerTypedPHI() 212 if (!LoadI->hasOneUse()) in foldIntegerTypedPHI() 218 AvailablePtrVals.emplace_back(LoadI); in foldIntegerTypedPHI() 295 LoadInst *LoadI = dyn_cast<LoadInst>(IncomingVal); in foldIntegerTypedPHI() local 298 (LoadI && LoadI->hasOneUse())) && in foldIntegerTypedPHI()
|
| /freebsd/contrib/llvm-project/llvm/lib/Target/Hexagon/ |
| H A D | HexagonVLIWPacketizer.cpp | 1522 bool LoadI = I.mayLoad(), StoreI = I.mayStore(); in isLegalToPacketizeTogether() local 1530 if (LoadJ && LoadI && HII->isPureSlot0(J)) { in isLegalToPacketizeTogether() 1537 (StoreJ && LoadI && !NVStoreJ)) && in isLegalToPacketizeTogether() 1545 if (StoreJ && LoadI && alias(J, I)) { in isLegalToPacketizeTogether() 1551 if (!LoadJ || (!LoadI && !StoreI)) { in isLegalToPacketizeTogether()
|
| /freebsd/contrib/llvm-project/llvm/lib/CodeGen/ |
| H A D | SelectOptimize.cpp | 1130 static bool isSafeToSinkLoad(Instruction *LoadI, Instruction *SI) { in isSafeToSinkLoad() argument 1132 if (LoadI->getParent() != SI->getParent()) in isSafeToSinkLoad() 1134 auto It = LoadI->getIterator(); in isSafeToSinkLoad()
|
| /freebsd/contrib/llvm-project/llvm/lib/Transforms/IPO/ |
| H A D | OpenMPOpt.cpp | 1173 LoadInst *LoadI = new LoadInst(I.getType(), AllocaI, in mergeParallelRegions() local 1176 UsrI->replaceUsesOfWith(&I, LoadI); in mergeParallelRegions() 4039 LoadInst *LoadI = new LoadInst( in insertInstructionGuardsHelper() local 4045 A.changeUseAfterManifest(*U, *LoadI); in insertInstructionGuardsHelper()
|
| H A D | AttributorAttributes.cpp | 1754 if (auto *LoadI = dyn_cast<LoadInst>(Usr)) { in updateImpl() local 1762 if (!handleAccess(A, *LoadI, /* Content */ nullptr, AK, in updateImpl() 1764 *LoadI->getType())) in updateImpl() 1784 BasicBlock *BB = LoadI->getParent(); in updateImpl() 1790 if (IsImpactedInRange(LoadI->getNextNonDebugInstruction(), &IntrI)) in updateImpl() 1807 if (IsImpactedInRange(LoadI->getNextNonDebugInstruction(), in updateImpl() 1817 for (const Use &LoadU : LoadI->uses()) { in updateImpl() 1840 << *Assumption.second << ": " << *LoadI in updateImpl() 1850 OffsetInfoMap[CurPtr].Offsets, Changed, *LoadI->getType()); in updateImpl()
|
| /freebsd/contrib/llvm-project/llvm/lib/Target/SystemZ/ |
| H A D | SystemZISelLowering.cpp | 1355 if (auto *LoadI = dyn_cast<LoadInst>(StoreI->getValueOperand())) in supportedAddressingMode() local 1356 if (LoadI->hasOneUse() && LoadI->getParent() == I->getParent()) in supportedAddressingMode() 1358 return getLoadStoreAddrMode(HasVector, LoadI->getType()); in supportedAddressingMode()
|