Home
last modified time | relevance | path

Searched refs:AtomicCmpXchgInst (Results 1 – 25 of 60) sorted by relevance

123

/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/CodeGen/
DAtomicExpandPass.cpp92 void expandPartwordCmpXchg(AtomicCmpXchgInst *I);
94 void expandAtomicCmpXchgToMaskedIntrinsic(AtomicCmpXchgInst *CI);
96 AtomicCmpXchgInst *convertCmpXchgToIntegerType(AtomicCmpXchgInst *CI);
102 bool tryExpandAtomicCmpXchg(AtomicCmpXchgInst *CI);
104 bool expandAtomicCmpXchg(AtomicCmpXchgInst *CI);
116 void expandAtomicCASToLibcall(AtomicCmpXchgInst *I);
150 static unsigned getAtomicOpSize(AtomicCmpXchgInst *CASI) { in getAtomicOpSize()
182 static unsigned getAtomicOpAlign(AtomicCmpXchgInst *CASI) { in getAtomicOpAlign()
223 auto CASI = dyn_cast<AtomicCmpXchgInst>(I); in runOnFunction()
443 AtomicCmpXchgInst::getStrongestFailureOrdering(Order)); in expandAtomicLoadToCmpXchg()
[all …]
DStackProtector.cpp175 if (AI == cast<AtomicCmpXchgInst>(I)->getNewValOperand()) in HasAddressTaken()
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/Scalar/
DLowerAtomic.cpp24 static bool LowerAtomicCmpXchgInst(AtomicCmpXchgInst *CXI) { in LowerAtomicCmpXchgInst()
124 else if (AtomicCmpXchgInst *CXI = dyn_cast<AtomicCmpXchgInst>(Inst)) in runOnBasicBlock()
DInferAddressSpaces.cpp366 else if (auto *CmpX = dyn_cast<AtomicCmpXchgInst>(&I)) in collectFlatAddressExpressions()
776 if (auto *CmpX = dyn_cast<AtomicCmpXchgInst>(Inst)) in isSimplePointerUseValidToReplace()
777 return OpNo == AtomicCmpXchgInst::getPointerOperandIndex() && in isSimplePointerUseValidToReplace()
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/IR/
DInstruction.cpp423 if (const AtomicCmpXchgInst *CXI = dyn_cast<AtomicCmpXchgInst>(I1)) in haveSameSpecialState()
424 return CXI->isVolatile() == cast<AtomicCmpXchgInst>(I2)->isVolatile() && in haveSameSpecialState()
425 CXI->isWeak() == cast<AtomicCmpXchgInst>(I2)->isWeak() && in haveSameSpecialState()
427 cast<AtomicCmpXchgInst>(I2)->getSuccessOrdering() && in haveSameSpecialState()
429 cast<AtomicCmpXchgInst>(I2)->getFailureOrdering() && in haveSameSpecialState()
431 cast<AtomicCmpXchgInst>(I2)->getSyncScopeID(); in haveSameSpecialState()
DInstructions.cpp1427 void AtomicCmpXchgInst::Init(Value *Ptr, Value *Cmp, Value *NewVal, in Init()
1460 AtomicCmpXchgInst::AtomicCmpXchgInst(Value *Ptr, Value *Cmp, Value *NewVal, in AtomicCmpXchgInst() function in AtomicCmpXchgInst
1467 AtomicCmpXchg, OperandTraits<AtomicCmpXchgInst>::op_begin(this), in AtomicCmpXchgInst()
1468 OperandTraits<AtomicCmpXchgInst>::operands(this), InsertBefore) { in AtomicCmpXchgInst()
1472 AtomicCmpXchgInst::AtomicCmpXchgInst(Value *Ptr, Value *Cmp, Value *NewVal, in AtomicCmpXchgInst() function in AtomicCmpXchgInst
1479 AtomicCmpXchg, OperandTraits<AtomicCmpXchgInst>::op_begin(this), in AtomicCmpXchgInst()
1480 OperandTraits<AtomicCmpXchgInst>::operands(this), InsertAtEnd) { in AtomicCmpXchgInst()
4158 AtomicCmpXchgInst *AtomicCmpXchgInst::cloneImpl() const { in cloneImpl()
4159 AtomicCmpXchgInst *Result = in cloneImpl()
4160 new AtomicCmpXchgInst(getOperand(0), getOperand(1), getOperand(2), in cloneImpl()
DCore.cpp3652 return cast<AtomicCmpXchgInst>(P)->isVolatile(); in LLVMGetVolatile()
3663 return cast<AtomicCmpXchgInst>(P)->setVolatile(isVolatile); in LLVMSetVolatile()
3667 return unwrap<AtomicCmpXchgInst>(CmpXchgInst)->isWeak(); in LLVMGetWeak()
3671 return unwrap<AtomicCmpXchgInst>(CmpXchgInst)->setWeak(isWeak); in LLVMSetWeak()
3950 return cast<AtomicCmpXchgInst>(P)->getSyncScopeID() == in LLVMIsAtomicSingleThread()
3960 return cast<AtomicCmpXchgInst>(P)->setSyncScopeID(SSID); in LLVMSetAtomicSingleThread()
3965 return mapToLLVMOrdering(cast<AtomicCmpXchgInst>(P)->getSuccessOrdering()); in LLVMGetCmpXchgSuccessOrdering()
3973 return cast<AtomicCmpXchgInst>(P)->setSuccessOrdering(O); in LLVMSetCmpXchgSuccessOrdering()
3978 return mapToLLVMOrdering(cast<AtomicCmpXchgInst>(P)->getFailureOrdering()); in LLVMGetCmpXchgFailureOrdering()
3986 return cast<AtomicCmpXchgInst>(P)->setFailureOrdering(O); in LLVMSetCmpXchgFailureOrdering()
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/Utils/
DFunctionComparator.cpp615 if (const AtomicCmpXchgInst *CXI = dyn_cast<AtomicCmpXchgInst>(L)) { in cmpOperations()
617 cast<AtomicCmpXchgInst>(R)->isVolatile())) in cmpOperations()
620 cast<AtomicCmpXchgInst>(R)->isWeak())) in cmpOperations()
624 cast<AtomicCmpXchgInst>(R)->getSuccessOrdering())) in cmpOperations()
628 cast<AtomicCmpXchgInst>(R)->getFailureOrdering())) in cmpOperations()
631 cast<AtomicCmpXchgInst>(R)->getSyncScopeID()); in cmpOperations()
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/include/llvm/Analysis/
DMemoryLocation.h202 static MemoryLocation get(const AtomicCmpXchgInst *CXI);
216 return get(cast<AtomicCmpXchgInst>(Inst)); in getOrNone()
DAliasAnalysis.h564 ModRefInfo getModRefInfo(const AtomicCmpXchgInst *CX,
568 ModRefInfo getModRefInfo(const AtomicCmpXchgInst *CX, const Value *P, in getModRefInfo()
711 ModRefInfo getModRefInfo(const AtomicCmpXchgInst *CX,
740 return getModRefInfo((const AtomicCmpXchgInst *)I, Loc, AAQIP); in getModRefInfo()
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/Instrumentation/
DBoundsChecking.cpp162 } else if (AtomicCmpXchgInst *AI = dyn_cast<AtomicCmpXchgInst>(&I)) { in addBoundsChecking()
DThreadSanitizer.cpp422 if (isa<AtomicCmpXchgInst>(I)) in isAtomic()
677 } else if (AtomicCmpXchgInst *CASI = dyn_cast<AtomicCmpXchgInst>(I)) { in instrumentAtomic()
DHWAddressSanitizer.cpp535 } else if (AtomicCmpXchgInst *XCHG = dyn_cast<AtomicCmpXchgInst>(I)) { in isInterestingMemoryAccess()
568 if (AtomicCmpXchgInst *XCHG = dyn_cast<AtomicCmpXchgInst>(I)) in getPointerOperandIndex()
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/RISCV/
DRISCVISelLowering.h214 shouldExpandAtomicCmpXchgInIR(AtomicCmpXchgInst *CI) const override;
216 emitMaskedAtomicCmpXchgIntrinsic(IRBuilder<> &Builder, AtomicCmpXchgInst *CI,
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Analysis/
DMemoryLocation.cpp64 MemoryLocation MemoryLocation::get(const AtomicCmpXchgInst *CXI) { in get()
DCaptureTracking.cpp334 auto *ACXI = cast<AtomicCmpXchgInst>(I); in PointerMayBeCaptured()
DAliasAnalysis.cpp576 ModRefInfo AAResults::getModRefInfo(const AtomicCmpXchgInst *CX, in getModRefInfo()
582 ModRefInfo AAResults::getModRefInfo(const AtomicCmpXchgInst *CX, in getModRefInfo()
DCFLGraph.h299 void visitAtomicCmpXchgInst(AtomicCmpXchgInst &Inst) { in visitAtomicCmpXchgInst()
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/CodeGen/SelectionDAG/
DSelectionDAGBuilder.h46 class AtomicCmpXchgInst; variable
747 void visitAtomicCmpXchg(const AtomicCmpXchgInst &I);
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/AMDGPU/
DAMDGPUPromoteAlloca.cpp561 } else if (AtomicCmpXchgInst *CAS = dyn_cast<AtomicCmpXchgInst>(UseInst)) { in collectUsesWithPtrTypes()
DAMDGPUPerfHintAnalysis.cpp131 if (auto AI = dyn_cast<AtomicCmpXchgInst>(Inst)) { in getMemoryInstrPtr()
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-subzero/include/llvm/IR/
DInstruction.def153 HANDLE_MEMORY_INST(34, AtomicCmpXchg , AtomicCmpXchgInst )
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/include/llvm/IR/
DInstVisitor.h173 RetTy visitAtomicCmpXchgInst(AtomicCmpXchgInst &I) { DELEGATE(Instruction);} in visitAtomicCmpXchgInst()
DInstruction.def176 HANDLE_MEMORY_INST(36, AtomicCmpXchg , AtomicCmpXchgInst )
DInstructions.h547 class AtomicCmpXchgInst : public Instruction {
556 AtomicCmpXchgInst *cloneImpl() const;
559 AtomicCmpXchgInst(Value *Ptr, Value *Cmp, Value *NewVal,
563 AtomicCmpXchgInst(Value *Ptr, Value *Cmp, Value *NewVal,
696 struct OperandTraits<AtomicCmpXchgInst> :
697 public FixedNumOperandTraits<AtomicCmpXchgInst, 3> {
700 DEFINE_TRANSPARENT_OPERAND_ACCESSORS(AtomicCmpXchgInst, Value)

123