/external/clang/test/Modules/Inputs/merge-template-members/ |
a1.h | 8 typedef A<int> AI;
|
a2.h | 8 typedef A<int> AI;
|
/external/skia/src/opts/ |
SkNx_neon.h | 19 AI static float32x4_t armv7_vrndmq_f32(float32x4_t v) { 28 AI SkNx(float32x2_t vec) : fVec(vec) {} 30 AI SkNx() {} 31 AI SkNx(float val) : fVec(vdup_n_f32(val)) {} 32 AI SkNx(float a, float b) { fVec = (float32x2_t) { a, b }; } 34 AI static SkNx Load(const void* ptr) { return vld1_f32((const float*)ptr); } 35 AI void store(void* ptr) const { vst1_f32((float*)ptr, fVec); } 37 AI SkNx invert() const { 43 AI SkNx operator + (const SkNx& o) const { return vadd_f32(fVec, o.fVec); } 44 AI SkNx operator - (const SkNx& o) const { return vsub_f32(fVec, o.fVec); [all...] |
SkNx_sse.h | 21 AI SkNx(const __m128& vec) : fVec(vec) {} 23 AI SkNx() {} 24 AI SkNx(float val) : fVec(_mm_set1_ps(val)) {} 25 AI static SkNx Load(const void* ptr) { 28 AI SkNx(float a, float b) : fVec(_mm_setr_ps(a,b,0,0)) {} 30 AI void store(void* ptr) const { _mm_storel_pi((__m64*)ptr, fVec); } 32 AI SkNx operator + (const SkNx& o) const { return _mm_add_ps(fVec, o.fVec); } 33 AI SkNx operator - (const SkNx& o) const { return _mm_sub_ps(fVec, o.fVec); } 34 AI SkNx operator * (const SkNx& o) const { return _mm_mul_ps(fVec, o.fVec); } 35 AI SkNx operator / (const SkNx& o) const { return _mm_div_ps(fVec, o.fVec); [all...] |
/external/llvm/lib/Transforms/Utils/ |
InstructionNamer.cpp | 35 for (Function::arg_iterator AI = F.arg_begin(), AE = F.arg_end(); 36 AI != AE; ++AI) 37 if (!AI->hasName() && !AI->getType()->isVoidTy()) 38 AI->setName("arg");
|
MetaRenamer.cpp | 84 for (auto AI = M.alias_begin(), AE = M.alias_end(); AI != AE; ++AI) { 85 StringRef Name = AI->getName(); 89 AI->setName("alias"); 125 for (auto AI = F.arg_begin(), AE = F.arg_end(); AI != AE; ++AI) 126 if (!AI->getType()->isVoidTy()) 127 AI->setName("arg") [all...] |
/external/swiftshader/third_party/LLVM/lib/Transforms/Utils/ |
InstructionNamer.cpp | 35 for (Function::arg_iterator AI = F.arg_begin(), AE = F.arg_end(); 36 AI != AE; ++AI) 37 if (!AI->hasName() && !AI->getType()->isVoidTy()) 38 AI->setName("arg");
|
PromoteMemoryToRegister.cpp | 80 bool llvm::isAllocaPromotable(const AllocaInst *AI) { 85 for (Value::const_use_iterator UI = AI->use_begin(), UE = AI->use_end(); 94 if (SI->getOperand(0) == AI) 95 return false; // Don't allow a store OF the AI, only INTO the AI. 275 void DetermineInsertionPoint(AllocaInst *AI, unsigned AllocaNum, 277 void ComputeLiveInBlocks(AllocaInst *AI, AllocaInfo &Info, 281 void RewriteSingleStoreAlloca(AllocaInst *AI, AllocaInfo &Info, 283 void PromoteSingleBlockAlloca(AllocaInst *AI, AllocaInfo &Info 1088 DenseMap<AllocaInst *, unsigned>::iterator ai = AllocaLookup.find(Dest); local [all...] |
Mem2Reg.cpp | 72 if (AllocaInst *AI = dyn_cast<AllocaInst>(I)) // Is it an alloca? 73 if (isAllocaPromotable(AI)) 74 Allocas.push_back(AI);
|
/external/harfbuzz_ng/src/ |
hb-atomic-private.hh | 67 #define hb_atomic_int_impl_add(AI, V) InterlockedExchangeAdd (&(AI), (V)) 85 #define hb_atomic_int_impl_add(AI, V) (OSAtomicAdd32Barrier ((V), &(AI)) - (V)) 103 #define hb_atomic_int_impl_add(AI, V) __sync_fetch_and_add (&(AI), (V)) 116 #define hb_atomic_int_impl_add(AI, V) ( ({__machine_rw_barrier ();}), atomic_add_int_nv (&(AI), (V)) - (V)) 127 static inline int hb_fetch_and_add(volatile int* AI, unsigned int V) { 129 int result = __fetch_and_add(AI, V) [all...] |
/external/llvm/include/llvm/CodeGen/ |
AtomicExpandUtils.h | 56 expandAtomicRMWToCmpXchg(AtomicRMWInst *AI, CreateCmpXchgInstFun Factory);
|
/external/llvm/tools/opt/ |
AnalysisWrappers.cpp | 47 for (CallSite::arg_iterator AI = CS.arg_begin(), 48 E = CS.arg_end(); AI != E; ++AI) { 49 if (!isa<Constant>(*AI)) continue;
|
/external/skia/src/core/ |
SkNx.h | 18 #define AI SK_ALWAYS_INLINE 29 AI SkNx() = default; 30 AI SkNx(const Half& lo, const Half& hi) : fLo(lo), fHi(hi) {} 32 AI SkNx(T v) : fLo(v), fHi(v) {} 34 AI SkNx(T a, T b) : fLo(a) , fHi(b) { static_assert(N==2, ""); } 35 AI SkNx(T a, T b, T c, T d) : fLo(a,b), fHi(c,d) { static_assert(N==4, ""); } 36 AI SkNx(T a, T b, T c, T d, T e, T f, T g, T h) : fLo(a,b,c,d), fHi(e,f,g,h) { 39 AI SkNx(T a, T b, T c, T d, T e, T f, T g, T h, 45 AI T operator[](int k) const { 50 AI static SkNx Load(const void* vptr) [all...] |
/external/swiftshader/third_party/LLVM/include/llvm/Transforms/Utils/ |
PromoteMemToReg.h | 30 bool isAllocaPromotable(const AllocaInst *AI);
|
/external/llvm/lib/CodeGen/ |
StackProtector.cpp | 61 StackProtector::getSSPLayout(const AllocaInst *AI) const { 62 return AI ? Layout.lookup(AI) : SSPLK_None; 166 bool StackProtector::HasAddressTaken(const Instruction *AI) { 167 for (const User *U : AI->users()) { 169 if (AI == SI->getValueOperand()) 172 if (AI == SI->getOperand(0)) 237 if (const AllocaInst *AI = dyn_cast<AllocaInst>(&I)) { 238 if (AI->isArrayAllocation()) { 244 if (const auto *CI = dyn_cast<ConstantInt>(AI->getArraySize())) [all...] |
/toolchain/binutils/binutils-2.25/ld/testsuite/ld-elf/ |
linkinfo1a.d | 7 \[[ 0-9]+\] \.rel[a]?\.plt[ \t]+REL[A]?[ \t][ \t0-9a-f]+AI[ \t0-9a-f]+
|
linkinfo1b.d | 8 \[[ 0-9]+\] \.rel[a]?\.plt[ \t]+REL[A]?[ \t][ \t0-9a-f]+AI[ \t0-9a-f]+
|
/external/swiftshader/third_party/LLVM/lib/Transforms/Instrumentation/ |
ProfilingUtils.cpp | 67 Function::arg_iterator AI; 71 AI = MainFn->arg_begin(); ++AI; 72 if (AI->getType() != ArgVTy) { 73 Instruction::CastOps opcode = CastInst::getCastOpcode(AI, false, ArgVTy, 76 CastInst::Create(opcode, AI, ArgVTy, "argv.cast", InitCall)); 78 InitCall->setArgOperand(1, AI); 83 AI = MainFn->arg_begin(); 86 if (!AI->getType()->isIntegerTy(32)) { 88 if (!AI->use_empty()) [all...] |
/external/clang/include/clang/CodeGen/ |
CGFunctionInfo.h | 124 auto AI = ABIArgInfo(Direct); 125 AI.setCoerceToType(T); 126 AI.setPaddingType(Padding); 127 AI.setDirectOffset(Offset); 128 AI.setCanBeFlattened(CanBeFlattened); 129 return AI; 132 auto AI = getDirect(T); 133 AI.setInReg(true); 134 return AI; 137 auto AI = ABIArgInfo(Extend) [all...] |
/external/llvm/include/llvm/Transforms/Scalar/ |
SROA.h | 115 bool presplitLoadsAndStores(AllocaInst &AI, sroa::AllocaSlices &AS); 116 AllocaInst *rewritePartition(AllocaInst &AI, sroa::AllocaSlices &AS, 118 bool splitAlloca(AllocaInst &AI, sroa::AllocaSlices &AS); 119 bool runOnAlloca(AllocaInst &AI);
|
/external/llvm/include/llvm/Transforms/Utils/ |
PromoteMemToReg.h | 32 bool isAllocaPromotable(const AllocaInst *AI);
|
ASanStackFrameLayout.h | 34 AllocaInst *AI; // The actual AllocaInst.
|
/external/swiftshader/third_party/LLVM/tools/opt/ |
AnalysisWrappers.cpp | 48 for (CallSite::arg_iterator AI = CS.arg_begin(),
49 E = CS.arg_end(); AI != E; ++AI) {
50 if (!isa<Constant>(*AI)) continue;
|
/external/llvm/unittests/Transforms/Utils/ |
IntegerDivision.cpp | 37 Function::arg_iterator AI = F->arg_begin(); 38 Value *A = &*AI++; 39 Value *B = &*AI++; 67 Function::arg_iterator AI = F->arg_begin(); 68 Value *A = &*AI++; 69 Value *B = &*AI++; 97 Function::arg_iterator AI = F->arg_begin(); 98 Value *A = &*AI++; 99 Value *B = &*AI++; 127 Function::arg_iterator AI = F->arg_begin() [all...] |
/external/swiftshader/third_party/LLVM/lib/Transforms/Scalar/ |
ScalarReplAggregates.cpp | 88 AllocaInst *AI; 113 explicit AllocaInfo(AllocaInst *ai) 114 : AI(ai), isUnsafe(false), isMemCpySrc(false), isMemCpyDst(false), 125 bool isSafeAllocaToScalarRepl(AllocaInst *AI); 138 void DoScalarReplacement(AllocaInst *AI, 142 void RewriteForScalarRepl(Instruction *I, AllocaInst *AI, uint64_t Offset, 144 void RewriteBitCast(BitCastInst *BC, AllocaInst *AI, uint64_t Offset, 146 void RewriteGEP(GetElementPtrInst *GEPI, AllocaInst *AI, uint64_t Offset, 148 void RewriteLifetimeIntrinsic(IntrinsicInst *II, AllocaInst *AI, [all...] |