/freebsd/contrib/llvm-project/llvm/lib/Transforms/Scalar/ |
H A D | LowerAtomicPass.cpp | 45 else if (AtomicCmpXchgInst *CXI = dyn_cast<AtomicCmpXchgInst>(&Inst)) in runOnBasicBlock()
|
H A D | InferAddressSpaces.cpp | 505 else if (auto *CmpX = dyn_cast<AtomicCmpXchgInst>(&I)) in collectFlatAddressExpressions() 1030 if (auto *CmpX = dyn_cast<AtomicCmpXchgInst>(Inst)) in isSimplePointerUseValidToReplace() 1031 return OpNo == AtomicCmpXchgInst::getPointerOperandIndex() && in isSimplePointerUseValidToReplace()
|
/freebsd/contrib/llvm-project/llvm/lib/CodeGen/ |
H A D | AtomicExpandPass.cpp | 92 bool expandPartwordCmpXchg(AtomicCmpXchgInst *I); 94 void expandAtomicCmpXchgToMaskedIntrinsic(AtomicCmpXchgInst *CI); 96 AtomicCmpXchgInst *convertCmpXchgToIntegerType(AtomicCmpXchgInst *CI); 102 bool tryExpandAtomicCmpXchg(AtomicCmpXchgInst *CI); 104 bool expandAtomicCmpXchg(AtomicCmpXchgInst *CI); 116 void expandAtomicCASToLibcall(AtomicCmpXchgInst *I); 190 static unsigned getAtomicOpSize(AtomicCmpXchgInst *CASI) { in getAtomicOpSize() 226 auto CASI = dyn_cast<AtomicCmpXchgInst>(I); in run() 517 AtomicCmpXchgInst::getStrongestFailureOrdering(Order)); in expandAtomicLoadToCmpXchg() 591 AtomicCmpXchgInst::getStrongestFailureOrdering(MemOpOrder), SSID); in createCmpXchgInstFun() [all …]
|
/freebsd/contrib/llvm-project/llvm/lib/IR/ |
H A D | Instruction.cpp | 831 if (const AtomicCmpXchgInst *CXI = dyn_cast<AtomicCmpXchgInst>(I1)) in hasSameSpecialState() 832 return CXI->isVolatile() == cast<AtomicCmpXchgInst>(I2)->isVolatile() && in hasSameSpecialState() 833 CXI->isWeak() == cast<AtomicCmpXchgInst>(I2)->isWeak() && in hasSameSpecialState() 835 cast<AtomicCmpXchgInst>(I2)->getSuccessOrdering() && in hasSameSpecialState() 837 cast<AtomicCmpXchgInst>(I2)->getFailureOrdering() && in hasSameSpecialState() 839 cast<AtomicCmpXchgInst>(I2)->getSyncScopeID(); in hasSameSpecialState() 1019 return cast<AtomicCmpXchgInst>(this)->isVolatile(); in isVolatile() 1046 return cast<AtomicCmpXchgInst>(this)->getNewValOperand()->getType(); in getAccessType()
|
H A D | MemoryModelRelaxationAnnotations.cpp | 168 return isa<LoadInst>(I) || isa<StoreInst>(I) || isa<AtomicCmpXchgInst>(I) || in canInstructionHaveMMRAs()
|
H A D | Core.cpp | 2122 if (AtomicCmpXchgInst *CXI = dyn_cast<AtomicCmpXchgInst>(P)) in LLVMGetAlignment() 2142 else if (AtomicCmpXchgInst *CXI = dyn_cast<AtomicCmpXchgInst>(P)) in LLVMSetAlignment() 4011 return cast<AtomicCmpXchgInst>(P)->isVolatile(); in LLVMGetVolatile() 4022 return cast<AtomicCmpXchgInst>(P)->setVolatile(isVolatile); in LLVMSetVolatile() 4026 return unwrap<AtomicCmpXchgInst>(CmpXchgInst)->isWeak(); in LLVMGetWeak() 4030 return unwrap<AtomicCmpXchgInst>(CmpXchgInst)->setWeak(isWeak); in LLVMSetWeak() 4348 return cast<AtomicCmpXchgInst>(P)->getSyncScopeID() == in LLVMIsAtomicSingleThread() 4364 return cast<AtomicCmpXchgInst>(P)->setSyncScopeID(SSID); in LLVMSetAtomicSingleThread() 4369 return mapToLLVMOrdering(cast<AtomicCmpXchgInst>(P)->getSuccessOrdering()); in LLVMGetCmpXchgSuccessOrdering() 4377 return cast<AtomicCmpXchgInst>(P)->setSuccessOrdering(O); in LLVMSetCmpXchgSuccessOrdering() [all …]
|
H A D | Instructions.cpp | 1331 void AtomicCmpXchgInst::Init(Value *Ptr, Value *Cmp, Value *NewVal, in Init() 1351 AtomicCmpXchgInst::AtomicCmpXchgInst(Value *Ptr, Value *Cmp, Value *NewVal, in AtomicCmpXchgInst() function in AtomicCmpXchgInst 1359 AtomicCmpXchg, OperandTraits<AtomicCmpXchgInst>::op_begin(this), in AtomicCmpXchgInst() 1360 OperandTraits<AtomicCmpXchgInst>::operands(this), InsertBefore) { in AtomicCmpXchgInst() 4237 AtomicCmpXchgInst *AtomicCmpXchgInst::cloneImpl() const { in cloneImpl() 4238 AtomicCmpXchgInst *Result = new AtomicCmpXchgInst( in cloneImpl()
|
/freebsd/contrib/llvm-project/llvm/include/llvm/Transforms/Utils/ |
H A D | LowerAtomic.h | 24 bool lowerAtomicCmpXchgInst(AtomicCmpXchgInst *CXI);
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/Utils/ |
H A D | FunctionComparator.cpp | 741 if (const AtomicCmpXchgInst *CXI = dyn_cast<AtomicCmpXchgInst>(L)) { in cmpOperations() 743 cast<AtomicCmpXchgInst>(R)->isVolatile())) in cmpOperations() 746 cmpNumbers(CXI->isWeak(), cast<AtomicCmpXchgInst>(R)->isWeak())) in cmpOperations() 750 cast<AtomicCmpXchgInst>(R)->getSuccessOrdering())) in cmpOperations() 754 cast<AtomicCmpXchgInst>(R)->getFailureOrdering())) in cmpOperations() 757 cast<AtomicCmpXchgInst>(R)->getSyncScopeID()); in cmpOperations()
|
H A D | LowerAtomic.cpp | 22 bool llvm::lowerAtomicCmpXchgInst(AtomicCmpXchgInst *CXI) { in lowerAtomicCmpXchgInst()
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/Instrumentation/ |
H A D | BoundsChecking.cpp | 167 } else if (AtomicCmpXchgInst *AI = dyn_cast<AtomicCmpXchgInst>(&I)) { in addBoundsChecking()
|
H A D | ThreadSanitizer.cpp | 760 } else if (AtomicCmpXchgInst *CASI = dyn_cast<AtomicCmpXchgInst>(I)) { in instrumentAtomic()
|
H A D | HWAddressSanitizer.cpp | 844 } else if (AtomicCmpXchgInst *XCHG = dyn_cast<AtomicCmpXchgInst>(I)) { in getInterestingMemoryOperands() 869 if (AtomicCmpXchgInst *XCHG = dyn_cast<AtomicCmpXchgInst>(I)) in getPointerOperandIndex()
|
/freebsd/contrib/llvm-project/llvm/include/llvm/Analysis/ |
H A D | MemoryLocation.h | 32 class AtomicCmpXchgInst; variable 258 static MemoryLocation get(const AtomicCmpXchgInst *CXI);
|
H A D | AliasAnalysis.h | 57 class AtomicCmpXchgInst; variable 597 ModRefInfo getModRefInfo(const AtomicCmpXchgInst *CX,
|
/freebsd/contrib/llvm-project/llvm/lib/Analysis/ |
H A D | MemoryLocation.cpp | 59 MemoryLocation MemoryLocation::get(const AtomicCmpXchgInst *CXI) { in get() 87 return get(cast<AtomicCmpXchgInst>(Inst)); in getOrNone()
|
H A D | AliasAnalysis.cpp | 568 ModRefInfo AAResults::getModRefInfo(const AtomicCmpXchgInst *CX, in getModRefInfo() 624 return getModRefInfo((const AtomicCmpXchgInst *)I, Loc, AAQIP); in getModRefInfo()
|
H A D | CaptureTracking.cpp | 359 auto *ACXI = cast<AtomicCmpXchgInst>(I); in DetermineUseCaptureKind()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/LoongArch/ |
H A D | LoongArchISelLowering.h | 199 shouldExpandAtomicCmpXchgInIR(AtomicCmpXchgInst *CI) const override; 201 AtomicCmpXchgInst *CI,
|
/freebsd/contrib/llvm-project/llvm/lib/Target/SPIRV/ |
H A D | SPIRVEmitIntrinsics.cpp | 166 Instruction *visitAtomicCmpXchgInst(AtomicCmpXchgInst &I); 204 isa<ExtractValueInst>(I) || isa<AtomicCmpXchgInst>(I); in isMemInstrToReplace() 439 } else if (auto *Ref = dyn_cast<AtomicCmpXchgInst>(I)) { in deduceElementTypeHelper() 662 } else if (auto *Ref = dyn_cast<AtomicCmpXchgInst>(I)) { in deduceOperandElementType() 667 AtomicCmpXchgInst::getPointerOperandIndex())); in deduceOperandElementType() 1348 Instruction *SPIRVEmitIntrinsics::visitAtomicCmpXchgInst(AtomicCmpXchgInst &I) { in visitAtomicCmpXchgInst()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/AMDGPU/Utils/ |
H A D | AMDGPUMemoryUtils.cpp | 318 if (checkNoAlias(dyn_cast<AtomicCmpXchgInst>(DefInst)) || in isReallyAClobber()
|
/freebsd/contrib/llvm-project/llvm/lib/CodeGen/SelectionDAG/ |
H A D | SelectionDAGBuilder.h | 45 class AtomicCmpXchgInst; variable 601 void visitAtomicCmpXchg(const AtomicCmpXchgInst &I);
|
/freebsd/contrib/llvm-project/llvm/include/llvm/IR/ |
H A D | InstVisitor.h | 171 RetTy visitAtomicCmpXchgInst(AtomicCmpXchgInst &I) { DELEGATE(Instruction);} in visitAtomicCmpXchgInst()
|
H A D | Instruction.def | 176 HANDLE_MEMORY_INST(36, AtomicCmpXchg , AtomicCmpXchgInst )
|
/freebsd/contrib/llvm-project/clang/lib/CodeGen/ |
H A D | CGAtomic.cpp | 388 llvm::AtomicCmpXchgInst *Pair = CGF.Builder.CreateAtomicCmpXchg( in emitAtomicCmpXchg() 1767 auto Failure = llvm::AtomicCmpXchgInst::getStrongestFailureOrdering(AO); in EmitAtomicUpdateLibcall() 1796 auto Failure = llvm::AtomicCmpXchgInst::getStrongestFailureOrdering(AO); in EmitAtomicUpdateOp() 1858 auto Failure = llvm::AtomicCmpXchgInst::getStrongestFailureOrdering(AO); in EmitAtomicUpdateLibcall() 1883 auto Failure = llvm::AtomicCmpXchgInst::getStrongestFailureOrdering(AO); in EmitAtomicUpdateOp()
|