/external/llvm/lib/Transforms/Scalar/ |
LowerAtomic.cpp | 23 static bool LowerAtomicCmpXchgInst(AtomicCmpXchgInst *CXI) { 120 else if (AtomicCmpXchgInst *CXI = dyn_cast<AtomicCmpXchgInst>(Inst))
|
SCCP.cpp | 497 void visitAtomicCmpXchgInst (AtomicCmpXchgInst &I) { markOverdefined(&I); } [all...] |
/external/llvm/lib/IR/ |
Instruction.cpp | 310 if (const AtomicCmpXchgInst *CXI = dyn_cast<AtomicCmpXchgInst>(this)) 311 return CXI->isVolatile() == cast<AtomicCmpXchgInst>(I)->isVolatile() && 312 CXI->getOrdering() == cast<AtomicCmpXchgInst>(I)->getOrdering() && 313 CXI->getSynchScope() == cast<AtomicCmpXchgInst>(I)->getSynchScope(); 384 if (const AtomicCmpXchgInst *CXI = dyn_cast<AtomicCmpXchgInst>(this)) 385 return CXI->isVolatile() == cast<AtomicCmpXchgInst>(I)->isVolatile() && 386 CXI->getOrdering() == cast<AtomicCmpXchgInst>(I)->getOrdering() && 387 CXI->getSynchScope() == cast<AtomicCmpXchgInst>(I)->getSynchScope() [all...] |
Instructions.cpp | [all...] |
AsmWriter.cpp | [all...] |
Verifier.cpp | 306 void visitAtomicCmpXchgInst(AtomicCmpXchgInst &CXI); [all...] |
/external/llvm/lib/Transforms/Instrumentation/ |
BoundsChecking.cpp | 184 if (isa<LoadInst>(I) || isa<StoreInst>(I) || isa<AtomicCmpXchgInst>(I) || 199 } else if (AtomicCmpXchgInst *AI = dyn_cast<AtomicCmpXchgInst>(Inst)) {
|
ThreadSanitizer.cpp | 319 if (isa<AtomicCmpXchgInst>(I)) 545 } else if (AtomicCmpXchgInst *CASI = dyn_cast<AtomicCmpXchgInst>(I)) {
|
AddressSanitizer.cpp | 637 if (AtomicCmpXchgInst *XCHG = dyn_cast<AtomicCmpXchgInst>(I)) { [all...] |
/art/compiler/llvm/ |
ir_builder.h | 91 ::llvm::AtomicCmpXchgInst* 94 ::llvm::AtomicCmpXchgInst* inst = 164 ::llvm::AtomicCmpXchgInst* 266 ::llvm::AtomicCmpXchgInst* CompareExchangeObjectOffset(::llvm::Value* object_addr,
|
/external/llvm/include/llvm/Analysis/ |
AliasAnalysis.h | 147 Location getLocation(const AtomicCmpXchgInst *CXI); 361 return getModRefInfo((const AtomicCmpXchgInst*)I, Loc); 444 ModRefResult getModRefInfo(const AtomicCmpXchgInst *CX, const Location &Loc); 447 ModRefResult getModRefInfo(const AtomicCmpXchgInst *CX,
|
/external/llvm/lib/Transforms/IPO/ |
MergeFunctions.cpp | 339 if (const AtomicCmpXchgInst *CXI = dyn_cast<AtomicCmpXchgInst>(I1)) 340 return CXI->isVolatile() == cast<AtomicCmpXchgInst>(I2)->isVolatile() && 341 CXI->getOrdering() == cast<AtomicCmpXchgInst>(I2)->getOrdering() && 342 CXI->getSynchScope() == cast<AtomicCmpXchgInst>(I2)->getSynchScope();
|
/external/llvm/lib/Analysis/ |
AliasAnalysis.cpp | 246 AliasAnalysis::getLocation(const AtomicCmpXchgInst *CXI) { 339 AliasAnalysis::getModRefInfo(const AtomicCmpXchgInst *CX, const Location &Loc) {
|
/external/llvm/include/llvm/ |
InstVisitor.h | 177 RetTy visitAtomicCmpXchgInst(AtomicCmpXchgInst &I) { DELEGATE(Instruction);}
|
/external/llvm/include/llvm/IR/ |
Instructions.h | 444 // AtomicCmpXchgInst Class 447 /// AtomicCmpXchgInst - an instruction that atomically checks whether a 451 class AtomicCmpXchgInst : public Instruction { 456 virtual AtomicCmpXchgInst *clone_impl() const; 462 AtomicCmpXchgInst(Value *Ptr, Value *Cmp, Value *NewVal, 465 AtomicCmpXchgInst(Value *Ptr, Value *Cmp, Value *NewVal, 544 struct OperandTraits<AtomicCmpXchgInst> : 545 public FixedNumOperandTraits<AtomicCmpXchgInst, 3> { 548 DEFINE_TRANSPARENT_OPERAND_ACCESSORS(AtomicCmpXchgInst, Value) [all...] |
IRBuilder.h | [all...] |
/external/llvm/lib/CodeGen/SelectionDAG/ |
SelectionDAGBuilder.h | 520 void visitAtomicCmpXchg(const AtomicCmpXchgInst &I);
|
/external/llvm/lib/Bitcode/Writer/ |
BitcodeWriter.cpp | [all...] |
/frameworks/compile/slang/BitWriter_2_9_func/ |
BitcodeWriter.cpp | [all...] |
/frameworks/compile/slang/BitWriter_3_2/ |
BitcodeWriter.cpp | [all...] |
/external/llvm/lib/Target/CppBackend/ |
CPPBackend.cpp | [all...] |
/external/clang/lib/CodeGen/ |
CGAtomic.cpp | 206 llvm::AtomicCmpXchgInst *CXI = [all...] |
/external/llvm/lib/Bitcode/Reader/ |
BitcodeReader.cpp | [all...] |
/external/llvm/lib/Transforms/Utils/ |
SimplifyCFG.cpp | [all...] |
/frameworks/compile/libbcc/bcinfo/BitReader_3_0/ |
BitcodeReader.cpp | [all...] |