/freebsd/contrib/llvm-project/llvm/lib/Analysis/ |
H A D | MemoryLocation.cpp | 95 MemoryLocation MemoryLocation::getForSource(const MemTransferInst *MTI) { in getForSource() 99 MemoryLocation MemoryLocation::getForSource(const AtomicMemTransferInst *MTI) { in getForSource() 103 MemoryLocation MemoryLocation::getForSource(const AnyMemTransferInst *MTI) { in getForSource()
|
H A D | AliasSetTracker.cpp | 341 void AliasSetTracker::add(AnyMemTransferInst *MTI) { in add() 387 if (AnyMemTransferInst *MTI = dyn_cast<AnyMemTransferInst>(I)) in add() local
|
H A D | StackSafetyAnalysis.cpp | 343 if (const auto *MTI = dyn_cast<MemTransferInst>(MI)) { in getMemIntrinsicAccessRange() local 495 if (const auto *MTI = dyn_cast<MemTransferInst>(MI)) { in analyzeAllUses() local
|
H A D | LazyValueInfo.cpp | 697 if (MemTransferInst *MTI = dyn_cast<MemTransferInst>(MI)) in AddNonNullPointersByInstruction() local
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/Utils/ |
H A D | VNCoercion.cpp | 269 MemTransferInst *MTI = cast<MemTransferInst>(MI); in analyzeLoadFromClobberingMemInst() local 400 MemTransferInst *MTI = cast<MemTransferInst>(SrcInst); in getMemInstValueForLoad() local 424 MemTransferInst *MTI = cast<MemTransferInst>(SrcInst); in getConstantMemInstValueForLoad() local
|
H A D | GlobalStatus.cpp | 162 } else if (const MemTransferInst *MTI = dyn_cast<MemTransferInst>(I)) { in analyzeGlobalAux() local
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/Scalar/ |
H A D | AlignmentFromAssumptions.cpp | 250 if (MemTransferInst *MTI = dyn_cast<MemTransferInst>(MI)) { processAssumption() local
|
H A D | InferAddressSpaces.cpp | 512 if (auto *MTI = dyn_cast<MemTransferInst>(MI)) in collectFlatAddressExpressions() local 1051 } else if (auto *MTI = dyn_cast<MemTransferInst>(MI)) { in handleMemIntrinsicPtrUse() local
|
/freebsd/contrib/llvm-project/llvm/lib/CodeGen/ |
H A D | SafeStack.cpp | 258 if (auto MTI = dyn_cast<MemTransferInst>(MI)) { in IsMemIntrinsicSafe() local
|
H A D | CodeGenPrepare.cpp | 2434 if (MemTransferInst *MTI = dyn_cast<MemTransferInst>(MI)) { in optimizeCallInst() local
|
/freebsd/contrib/llvm-project/llvm/lib/Target/AMDGPU/ |
H A D | AMDGPUPromoteAlloca.cpp | 604 if (auto *MTI = dyn_cast<MemTransferInst>(Inst)) { in promoteAllocaUserToVector() local
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/InstCombine/ |
H A D | InstCombineCalls.cpp | 1574 if (AnyMemTransferInst *MTI = dyn_cast<AnyMemTransferInst>(MI)) { in visitCallInst() local 1582 if (auto *MTI = dyn_cast<AnyMemTransferInst>(MI)) { in visitCallInst() local
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/Instrumentation/ |
H A D | HWAddressSanitizer.cpp | 1060 if (MemTransferInst *MTI = dyn_cast<MemTransferInst>(MI)) { in ignoreMemIntrinsic() local
|
H A D | DataFlowSanitizer.cpp | 2948 auto *MTI = cast<MemTransferInst>( in visitMemTransferInst() local
|
/freebsd/contrib/llvm-project/llvm/lib/Target/Mips/ |
H A D | MipsFastISel.cpp | 1650 const auto *MTI = cast<MemTransferInst>(II); in fastLowerIntrinsicCall() local
|
/freebsd/contrib/llvm-project/llvm/lib/Target/ARM/ |
H A D | ARMFastISel.cpp | 2522 const MemTransferInst &MTI = cast<MemTransferInst>(I); in SelectIntrinsicCall() local
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/IPO/ |
H A D | GlobalOpt.cpp | 230 } else if (MemTransferInst *MTI = dyn_cast<MemTransferInst>(U)) { in CleanupPointerRootUsers() local
|
H A D | Attributor.cpp | 910 if (const MemTransferInst *MTI = dyn_cast<MemTransferInst>(&I)) in isPotentiallyAffectedByBarrier() local
|
/freebsd/contrib/llvm-project/llvm/lib/IR/ |
H A D | AutoUpgrade.cpp | 4785 if (auto *MTI = dyn_cast<MemTransferInst>(MemCI)) in UpgradeIntrinsicCall() local
|
/freebsd/contrib/llvm-project/llvm/lib/Target/AArch64/ |
H A D | AArch64FastISel.cpp | 3487 const auto *MTI = cast<MemTransferInst>(II); in fastLowerIntrinsicCall() local
|