/external/llvm/lib/Transforms/Scalar/ |
LowerAtomic.cpp | 24 static bool LowerAtomicCmpXchgInst(AtomicCmpXchgInst *CXI) { 126 else if (AtomicCmpXchgInst *CXI = dyn_cast<AtomicCmpXchgInst>(Inst))
|
/external/llvm/include/llvm/Analysis/ |
MemoryLocation.h | 68 static MemoryLocation get(const AtomicCmpXchgInst *CXI); 77 else if (auto *I = dyn_cast<AtomicCmpXchgInst>(Inst))
|
AliasAnalysis.h | 396 ModRefInfo getModRefInfo(const AtomicCmpXchgInst *CX, 400 ModRefInfo getModRefInfo(const AtomicCmpXchgInst *CX, const Value *P, 479 return getModRefInfo((const AtomicCmpXchgInst*)I, Loc); [all...] |
/external/llvm/lib/IR/ |
Instruction.cpp | 312 if (const AtomicCmpXchgInst *CXI = dyn_cast<AtomicCmpXchgInst>(I1)) 313 return CXI->isVolatile() == cast<AtomicCmpXchgInst>(I2)->isVolatile() && 314 CXI->isWeak() == cast<AtomicCmpXchgInst>(I2)->isWeak() && 316 cast<AtomicCmpXchgInst>(I2)->getSuccessOrdering() && 318 cast<AtomicCmpXchgInst>(I2)->getFailureOrdering() && 319 CXI->getSynchScope() == cast<AtomicCmpXchgInst>(I2)->getSynchScope();
|
Instructions.cpp | [all...] |
AsmWriter.cpp | [all...] |
/external/llvm/lib/Transforms/Instrumentation/ |
BoundsChecking.cpp | 182 if (isa<LoadInst>(I) || isa<StoreInst>(I) || isa<AtomicCmpXchgInst>(I) || 198 } else if (AtomicCmpXchgInst *AI = dyn_cast<AtomicCmpXchgInst>(Inst)) {
|
ThreadSanitizer.cpp | 325 if (isa<AtomicCmpXchgInst>(I)) 554 } else if (AtomicCmpXchgInst *CASI = dyn_cast<AtomicCmpXchgInst>(I)) {
|
/external/llvm/lib/CodeGen/ |
AtomicExpandPass.cpp | 63 bool expandAtomicCmpXchg(AtomicCmpXchgInst *CI); 98 auto CASI = dyn_cast<AtomicCmpXchgInst>(I); 276 AtomicCmpXchgInst::getStrongestFailureOrdering(Order)); 337 AtomicCmpXchgInst::getStrongestFailureOrdering(MemOpOrder)); 451 bool AtomicExpand::expandAtomicCmpXchg(AtomicCmpXchgInst *CI) {
|
/external/llvm/lib/Analysis/ |
MemoryLocation.cpp | 47 MemoryLocation MemoryLocation::get(const AtomicCmpXchgInst *CXI) {
|
AliasAnalysis.cpp | 276 ModRefInfo AAResults::getModRefInfo(const AtomicCmpXchgInst *CX,
|
CFLAliasAnalysis.cpp | 190 void visitAtomicCmpXchgInst(AtomicCmpXchgInst &Inst) { 479 Value *visitAtomicCmpXchgInst(AtomicCmpXchgInst &Inst) { [all...] |
MemoryDependenceAnalysis.cpp | 367 else if (AtomicCmpXchgInst *AI = dyn_cast<AtomicCmpXchgInst>(Inst)) [all...] |
/external/llvm/unittests/Analysis/ |
AliasAnalysisTest.cpp | 184 auto *CmpXChg1 = new AtomicCmpXchgInst(Addr, ConstantInt::get(IntType, 0),
|
/external/llvm/lib/Transforms/IPO/ |
MergeFunctions.cpp | [all...] |
/external/llvm/include/llvm/IR/ |
InstVisitor.h | 180 RetTy visitAtomicCmpXchgInst(AtomicCmpXchgInst &I) { DELEGATE(Instruction);}
|
Instructions.h | 512 // AtomicCmpXchgInst Class 515 /// AtomicCmpXchgInst - an instruction that atomically checks whether a 519 class AtomicCmpXchgInst : public Instruction { 528 AtomicCmpXchgInst *cloneImpl() const; 535 AtomicCmpXchgInst(Value *Ptr, Value *Cmp, Value *NewVal, 540 AtomicCmpXchgInst(Value *Ptr, Value *Cmp, Value *NewVal, 666 struct OperandTraits<AtomicCmpXchgInst> : 667 public FixedNumOperandTraits<AtomicCmpXchgInst, 3> { 670 DEFINE_TRANSPARENT_OPERAND_ACCESSORS(AtomicCmpXchgInst, Value) [all...] |
/external/clang/lib/CodeGen/ |
CGAtomic.cpp | 386 llvm::AtomicCmpXchgInst *Pair = CGF.Builder.CreateAtomicCmpXchg( 448 llvm::AtomicCmpXchgInst::getStrongestFailureOrdering(SuccessOrder); [all...] |
/external/llvm/lib/Target/AArch64/ |
AArch64ISelLowering.h | 355 bool shouldExpandAtomicCmpXchgInIR(AtomicCmpXchgInst *AI) const override;
|
/external/llvm/lib/Target/ARM/ |
ARMISelLowering.h | 461 bool shouldExpandAtomicCmpXchgInIR(AtomicCmpXchgInst *AI) const override;
|
/external/llvm/lib/Bitcode/Writer/ |
BitcodeWriter.cpp | [all...] |
/frameworks/compile/slang/BitWriter_2_9_func/ |
BitcodeWriter.cpp | [all...] |
/frameworks/compile/slang/BitWriter_3_2/ |
BitcodeWriter.cpp | [all...] |
/external/llvm/lib/Transforms/Utils/ |
InlineFunction.cpp | 574 else if (const AtomicCmpXchgInst *CXI = dyn_cast<AtomicCmpXchgInst>(I)) [all...] |
/external/llvm/lib/CodeGen/SelectionDAG/ |
SelectionDAGBuilder.h | [all...] |