/freebsd/contrib/llvm-project/llvm/lib/Transforms/Utils/ |
H A D | SimplifyIndVar.cpp | 88 Value *foldIVUser(Instruction *UseInst, Instruction *IVOperand); 90 bool eliminateIdentitySCEV(Instruction *UseInst, Instruction *IVOperand); 91 bool replaceIVUserWithLoopInvariant(Instruction *UseInst); 92 bool replaceFloatIVWithIntegerIV(Instruction *UseInst); 97 bool eliminateIVUser(Instruction *UseInst, Instruction *IVOperand); 134 Value *SimplifyIndvar::foldIVUser(Instruction *UseInst, Instruction *IVOperand) { in foldIVUser() argument 139 switch (UseInst->getOpcode()) { in foldIVUser() 146 if (IVOperand != UseInst->getOperand(OperIdx) || in foldIVUser() 147 !isa<ConstantInt>(UseInst->getOperand(1))) in foldIVUser() 160 ConstantInt *D = cast<ConstantInt>(UseInst->getOperand(1)); in foldIVUser() [all …]
|
H A D | Local.cpp | 1342 if (Instruction *UseInst = dyn_cast<Instruction>(U.getUser())) in TryToSimplifyUncondBranchFromEmptyBlock() local 1343 return UseInst->getParent() != CommonPred && in TryToSimplifyUncondBranchFromEmptyBlock() 1344 BBPreds.contains(UseInst->getParent()); in TryToSimplifyUncondBranchFromEmptyBlock()
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/Scalar/ |
H A D | DeadStoreElimination.cpp | 1151 Instruction *UseInst) { in isCompleteOverwrite() 1155 if (!UseInst->mayWriteToMemory()) in isCompleteOverwrite() 1158 if (auto *CB = dyn_cast<CallBase>(UseInst)) in isCompleteOverwrite() 1163 if (auto CC = getLocForWrite(UseInst)) in isCompleteOverwrite() 1164 return isOverwrite(UseInst, DefInst, *CC, DefLoc, InstWriteOffset, in isCompleteOverwrite() 1197 Instruction *UseInst = cast<MemoryUseOrDef>(UseAccess)->getMemoryInst(); in isWriteAtEndOfFunction() local 1198 if (isReadClobber(DefLoc, UseInst)) { in isWriteAtEndOfFunction() 1199 LLVM_DEBUG(dbgs() << " ... hit read clobber " << *UseInst << ".\n"); in isWriteAtEndOfFunction() 1264 bool isReadClobber(const MemoryLocation &DefLoc, Instruction *UseInst) { in isReadClobber() 1265 if (isNoopIntrinsic(UseInst)) in isReadClobber() [all …]
|
H A D | Sink.cpp | 131 Instruction *UseInst = cast<Instruction>(U.getUser()); in SinkInstruction() 132 BasicBlock *UseBlock = UseInst->getParent(); in SinkInstruction() local 133 if (PHINode *PN = dyn_cast<PHINode>(UseInst)) { in SinkInstruction()
|
/freebsd/contrib/llvm-project/llvm/lib/CodeGen/ |
H A D | MachineSink.cpp | 447 MachineInstr &UseInst = *MO.getParent(); in PerformSinkAndFold() local 448 if (UseInst.isCopy()) { in PerformSinkAndFold() 450 if (const MachineOperand &O = UseInst.getOperand(0); O.isReg()) in PerformSinkAndFold() 466 } else if (UseInst.mayLoadOrStore()) { in PerformSinkAndFold() 468 if (!TII->canFoldIntoAddrMode(UseInst, Reg, MI, AM)) in PerformSinkAndFold() 475 if (UseInst.getParent() != MI.getParent()) { in PerformSinkAndFold() 495 const MachineBasicBlock &MBB = *UseInst.getParent(); in PerformSinkAndFold() 506 SinkInto.emplace_back(&UseInst, MaybeAM); in PerformSinkAndFold() 625 MachineInstr *UseInst = MO.getParent(); in AllUsesDominatedByBlock() local 627 MachineBasicBlock *UseBlock = UseInst->getParent(); in AllUsesDominatedByBlock() [all …]
|
H A D | MIRCanonicalizerPass.cpp | 210 MachineInstr *UseInst = UO.getParent(); in rescheduleCanonically() local 213 const unsigned UseLoc = getInstrIdx(*UseInst); in rescheduleCanonically() 216 if (UseInst->getParent() != Def->getParent()) in rescheduleCanonically() 223 UseToBringDefCloserTo = UseInst; in rescheduleCanonically()
|
H A D | RegAllocFast.cpp | 523 for (const MachineInstr &UseInst : MRI->use_nodbg_instructions(VirtReg)) { in mayLiveOut() local 524 if (UseInst.getParent() != MBB || ++C >= Limit) { in mayLiveOut() 533 if (SelfLoopDef == &UseInst || in mayLiveOut() 534 !dominates(PosIndexes, *SelfLoopDef, UseInst)) { in mayLiveOut()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/PowerPC/ |
H A D | PPCPreEmitPeephole.cpp | 258 MachineBasicBlock::iterator UseInst; in addLinkerOpt() member 303 Pair.UseInst = BBI; in addLinkerOpt() 314 assert(Pair->UseInst.isValid() && Pair->StillValid && in addLinkerOpt() 321 for (; BBI != Pair->UseInst; ++BBI) { in addLinkerOpt() 343 Pair->UseInst->addOperand(ImplUse); in addLinkerOpt() 351 Pair->UseInst->addOperand(*MF, PCRelLabel); in addLinkerOpt()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/AMDGPU/ |
H A D | AMDGPUPromoteAlloca.cpp | 104 Instruction *UseInst, int OpIdx0, 1162 Instruction *UseInst = cast<Instruction>(User); in collectUsesWithPtrTypes() local 1163 if (UseInst->getOpcode() == Instruction::PtrToInt) in collectUsesWithPtrTypes() 1166 if (LoadInst *LI = dyn_cast<LoadInst>(UseInst)) { in collectUsesWithPtrTypes() 1173 if (StoreInst *SI = dyn_cast<StoreInst>(UseInst)) { in collectUsesWithPtrTypes() 1180 } else if (AtomicRMWInst *RMW = dyn_cast<AtomicRMWInst>(UseInst)) { in collectUsesWithPtrTypes() 1183 } else if (AtomicCmpXchgInst *CAS = dyn_cast<AtomicCmpXchgInst>(UseInst)) { in collectUsesWithPtrTypes() 1190 if (ICmpInst *ICmp = dyn_cast<ICmpInst>(UseInst)) { in collectUsesWithPtrTypes() 1198 if (UseInst->getOpcode() == Instruction::AddrSpaceCast) { in collectUsesWithPtrTypes() 1200 if (PointerMayBeCaptured(UseInst, true, true)) in collectUsesWithPtrTypes() [all …]
|
H A D | GCNNSAReassign.cpp | 220 const MachineInstr *UseInst = U.getParent(); in CheckNSA() local 221 if (UseInst->isCopy() && UseInst->getOperand(0).getReg() == PhysReg) in CheckNSA()
|
H A D | AMDGPULateCodeGenPrepare.cpp | 325 Instruction *UseInst = cast<Instruction>(V); in optimizeLiveType() local 327 if (UseInst->getParent() != II->getParent() || isa<PHINode>(II)) { in optimizeLiveType() 328 Uses.insert(UseInst); in optimizeLiveType()
|
H A D | SIPeepholeSDWA.cpp | 469 for (MachineInstr &UseInst : MRI->use_nodbg_instructions(PotentialMO->getReg())) { in potentialToConvert() 470 if (&UseInst != ParentMI) in potentialToConvert()
|
H A D | SIInstrInfo.cpp | 9415 auto &UseInst = *Use.getParent(); in execMayBeModifiedBeforeAnyUse() local 9418 if (UseInst.getParent() != DefBB || UseInst.isPHI()) in execMayBeModifiedBeforeAnyUse()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/WebAssembly/ |
H A D | WebAssemblyRegStackify.cpp | 450 const MachineInstr *UseInst = Use.getParent(); in oneUseDominatesOtherUses() local 451 VNInfo *UseVNI = LI.getVNInfoBefore(LIS.getInstructionIndex(*UseInst)); in oneUseDominatesOtherUses() 456 if (UseInst == OneUseInst) { in oneUseDominatesOtherUses() 463 while (!MDT.dominates(OneUseInst, UseInst)) { in oneUseDominatesOtherUses() 469 if (UseInst->getDesc().getNumDefs() == 0) in oneUseDominatesOtherUses() 471 const MachineOperand &MO = UseInst->getOperand(0); in oneUseDominatesOtherUses() 485 UseInst = NewUseInst; in oneUseDominatesOtherUses()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/X86/ |
H A D | X86FastPreTileConfig.cpp | 143 for (const MachineInstr &UseInst : MRI->use_nodbg_instructions(VirtReg)) { in mayLiveOut() local 144 if (UseInst.getParent() != MBB) { in mayLiveOut() 153 if (dominates(*MBB, *CfgMI, UseInst)) { in mayLiveOut()
|
/freebsd/contrib/llvm-project/llvm/lib/Analysis/ |
H A D | MemorySSA.cpp | 283 const Instruction *UseInst, AliasAnalysisType &AA) { in instructionClobbersQuery() argument 313 if (auto *CB = dyn_cast_or_null<CallBase>(UseInst)) { in instructionClobbersQuery() 319 if (auto *UseLoad = dyn_cast_or_null<LoadInst>(UseInst)) in instructionClobbersQuery()
|
H A D | ScalarEvolution.cpp | 9502 getConstantEvolvingPHIOperands(Instruction *UseInst, const Loop *L, in getConstantEvolvingPHIOperands() argument 9511 for (Value *Op : UseInst->operands()) { in getConstantEvolvingPHIOperands()
|
/freebsd/contrib/llvm-project/llvm/lib/Frontend/OpenMP/ |
H A D | OMPIRBuilder.cpp | 4688 auto *UseInst = dyn_cast<Instruction>(U.getUser()); in removeUnusedBlocksFromParent() local 4689 if (!UseInst) in removeUnusedBlocksFromParent() 4691 if (BBsToErase.count(UseInst->getParent())) in removeUnusedBlocksFromParent()
|