/external/llvm/test/Transforms/LoopStrengthReduce/ |
dont_insert_redundant_ops.ll | 13 %gep2 = getelementptr { i32, i32 }* %P, i32 %INDVAR, i32 1 ; <i32*> [#uses=1] 14 store i32 0, i32* %gep2 29 %gep2 = getelementptr [2 x i32]* %P, i32 %INDVAR, i64 1 ; <i32*> [#uses=1] 30 store i32 0, i32* %gep2
|
/external/llvm/test/Transforms/MemCpyOpt/ |
atomic.ll | 21 %gep2 = getelementptr inbounds [101 x i32]* %x, i32 0, i32 0 22 call void @otherf(i32* %gep2)
|
/external/llvm/unittests/VMCore/ |
InstructionsTest.cpp | 167 GetElementPtrInst *Gep2 = GetElementPtrInst::Create(PtrVecB, C2xi32a); 172 CastInst *BTC2 = new BitCastInst(Gep2, V2xi8PTy); 182 EXPECT_NE(S2, Gep2); 192 GetPointerBaseWithConstantOffset(Gep2, Offset, TD); 198 GetElementPtrInst *GepII2 = GetElementPtrInst::Create(Gep2, C2xi32b); 223 delete Gep2;
|
/external/llvm/lib/Transforms/IPO/ |
MergeFunctions.cpp | 181 bool isEquivalentGEP(const GEPOperator *GEP1, const GEPOperator *GEP2); 183 const GetElementPtrInst *GEP2) { 184 return isEquivalentGEP(cast<GEPOperator>(GEP1), cast<GEPOperator>(GEP2)); 346 const GEPOperator *GEP2) { 349 if (TD && GEP1->hasAllConstantIndices() && GEP2->hasAllConstantIndices()) { 351 SmallVector<Value *, 8> Indices2(GEP2->idx_begin(), GEP2->idx_end()); 354 uint64_t Offset2 = TD->getIndexedOffset(GEP2->getPointerOperandType(), 360 GEP2->getPointerOperand()->getType()) 363 if (GEP1->getNumOperands() != GEP2->getNumOperands() [all...] |
/external/llvm/test/Transforms/InstCombine/ |
icmp.ll | 596 %gep2 = getelementptr inbounds i8* %foo, i64 10 598 %cmp = icmp ult i8* %cast1, %gep2 609 %gep2 = getelementptr inbounds i8* %foo, i64 %j 611 %cmp = icmp ult i8* %cast1, %gep2 622 %gep2 = getelementptr i8* %foo, i64 %j 624 %cmp = icmp ult i8* %cast1, %gep2 628 ; CHECK: icmp ult i8* %cast1, %gep2
|
sub.ll | 307 %gep2 = getelementptr inbounds i8* %foo, i64 %j 309 %cast2 = ptrtoint i8* %gep2 to i64 320 %gep2 = getelementptr inbounds i8* %foo, i64 %j 322 %cast2 = ptrtoint i8* %gep2 to i64
|
getelementptr.ll | 477 ; would wait for gep1 and gep2 to be combined and never combine 2 and 3. 484 %gep2 = getelementptr %three_gep_t2* %gep1, i64 0, i32 0 485 %gep3 = getelementptr %three_gep_t* %gep2, i64 0, i32 0
|
/external/llvm/lib/Transforms/InstCombine/ |
InstCombineAddSub.cpp | 480 GEPOperator *GEP1 = 0, *GEP2 = 0; 493 GEP2 = RHSGEP; 509 GEP2 = LHSGEP; 516 // Avoid duplicating the arithmetic if GEP2 has non-constant indices and 519 (GEP2 != 0 && !GEP2->hasAllConstantIndices() && !GEP2->hasOneUse())) 527 if (GEP2) { 528 Value *Offset = EmitGEPOffset(GEP2);
|
/external/llvm/test/CodeGen/Thumb2/ |
machine-licm.ll | 64 %gep2 = getelementptr i8* %ptr2, i32 %tmp1 65 call void @llvm.arm.neon.vst1.v4f32(i8* %gep2, <4 x float> %tmp3, i32 1)
|
/external/llvm/lib/Transforms/Scalar/ |
MemCpyOptimizer.cpp | 79 GetElementPtrInst *GEP2 = dyn_cast<GetElementPtrInst>(Ptr2); 85 if (GEP1 && GEP2 == 0 && GEP1->getOperand(0)->stripPointerCasts() == Ptr2) { 90 if (GEP2 && GEP1 == 0 && GEP2->getOperand(0)->stripPointerCasts() == Ptr1) { 91 Offset = GetOffsetFromIndex(GEP2, 1, VariableIdxFound, TD); 100 if (!GEP1 || !GEP2 || GEP1->getOperand(0) != GEP2->getOperand(0)) 105 for (; Idx != GEP1->getNumOperands() && Idx != GEP2->getNumOperands(); ++Idx) 106 if (GEP1->getOperand(Idx) != GEP2->getOperand(Idx)) 110 int64_t Offset2 = GetOffsetFromIndex(GEP2, Idx, VariableIdxFound, TD) [all...] |
/external/freetype/src/truetype/ |
ttobjs.h | 111 FT_UShort gep2; member in struct:TT_GraphicsState_
|
ttinterp.c | [all...] |
/external/llvm/test/Transforms/InstSimplify/ |
compare.ll | 33 define i1 @gep2() { 34 ; CHECK: @gep2
|
/external/llvm/lib/Analysis/ |
BasicAliasAnalysis.cpp | 394 /// decomposed GetElementPtr instructions GEP1 and GEP2 which have common base 395 /// pointers. Subtract the GEP2 indices from GEP1 to find the symbolic [all...] |
/external/llvm/test/CodeGen/X86/ |
block-placement.ll | 32 %gep2 = getelementptr i32* %a, i32 2 33 %val2 = load i32* %gep2
|