/freebsd/contrib/llvm-project/llvm/lib/Transforms/Scalar/ |
H A D | InferAddressSpaces.cpp | 503 else if (auto *RMW = dyn_cast<AtomicRMWInst>(&I)) in collectFlatAddressExpressions() local 1026 if (auto *RMW = dyn_cast<AtomicRMWInst>(Inst)) in isSimplePointerUseValidToReplace() local
|
H A D | LoopStrengthReduce.cpp | 1023 } else if (AtomicRMWInst *RMW = dyn_cast<AtomicRMWInst>(Inst)) { in isAddressUse() local 1047 } else if (const AtomicRMWInst *RMW = dyn_cast<AtomicRMWInst>(Inst)) { in getAccessType() local
|
/freebsd/contrib/llvm-project/llvm/lib/Target/BPF/ |
H A D | BPFCheckAndAdjustIR.cpp | 509 else if (auto *RMW = dyn_cast<AtomicRMWInst>(&I)) in insertASpaceCasts() local
|
/freebsd/contrib/llvm-project/llvm/lib/Analysis/ |
H A D | AliasAnalysis.cpp | 586 ModRefInfo AAResults::getModRefInfo(const AtomicRMWInst *RMW, in getModRefInfo()
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/Instrumentation/ |
H A D | HWAddressSanitizer.cpp | 839 } else if (AtomicRMWInst *RMW = dyn_cast<AtomicRMWInst>(I)) { in getInterestingMemoryOperands() local 867 if (AtomicRMWInst *RMW = dyn_cast<AtomicRMWInst>(I)) in getPointerOperandIndex() local
|
H A D | MemProfiler.cpp | 347 } else if (AtomicRMWInst *RMW = dyn_cast<AtomicRMWInst>(I)) { in isInterestingMemoryAccess() local
|
H A D | AddressSanitizer.cpp | 1420 } else if (AtomicRMWInst *RMW = dyn_cast<AtomicRMWInst>(I)) { in getInterestingMemoryOperands() local
|
/freebsd/contrib/llvm-project/llvm/lib/Target/AMDGPU/ |
H A D | AMDGPUPromoteAlloca.cpp | 1180 } else if (AtomicRMWInst *RMW = dyn_cast<AtomicRMWInst>(UseInst)) { in collectUsesWithPtrTypes() local
|
H A D | AMDGPULowerBufferFatPointers.cpp | 1100 else if (auto *RMW = dyn_cast<AtomicRMWInst>(I)) { in handleMemoryInst() local
|
H A D | SIISelLowering.cpp | 16083 static OptimizationRemark emitAtomicRMWLegalRemark(const AtomicRMWInst *RMW) { in emitAtomicRMWLegalRemark()
|
/freebsd/contrib/llvm-project/llvm/lib/CodeGen/ |
H A D | TargetLoweringBase.cpp | 2267 if (const AtomicRMWInst *RMW = dyn_cast<AtomicRMWInst>(&AI)) { in getAtomicMemOperandFlags() local
|
H A D | CodeGenPrepare.cpp | 5241 if (AtomicRMWInst *RMW = dyn_cast<AtomicRMWInst>(UserI)) { in FindAllMemoryUses() local 8380 if (AtomicRMWInst *RMW = dyn_cast<AtomicRMWInst>(I)) { in optimizeInst() local
|
/freebsd/contrib/llvm-project/llvm/lib/IR/ |
H A D | AutoUpgrade.cpp | 4096 AtomicRMWInst *RMW = in upgradeAMDGCNIntrinsicCall() local
|
/freebsd/contrib/llvm-project/llvm/include/llvm/CodeGen/ |
H A D | TargetLowering.h | 2320 virtual AtomicExpansionKind shouldExpandAtomicRMWInIR(AtomicRMWInst *RMW) const { in shouldExpandAtomicRMWInIR()
|
/freebsd/contrib/llvm-project/clang/lib/CodeGen/ |
H A D | CGBuiltin.cpp | 4840 AtomicRMWInst *RMW = Builder.CreateAtomicRMW(llvm::AtomicRMWInst::Xchg, in EmitBuiltinExpr() local 19250 llvm::AtomicRMWInst *RMW = in EmitAMDGPUBuiltinExpr() local
|
/freebsd/contrib/llvm-project/llvm/lib/Target/PowerPC/ |
H A D | PPCISelLowering.cpp | 18820 Function *RMW = Intrinsic::getDeclaration( in emitMaskedAtomicRMWIntrinsic() local
|
/freebsd/contrib/llvm-project/llvm/lib/Target/AArch64/ |
H A D | AArch64ISelLowering.cpp | 26460 if (const auto *RMW = dyn_cast<AtomicRMWInst>(I)) in isOpSuitableForLSE128() local
|