HomeSort by relevance Sort by last modified time
    Searched full:tmp4 (Results 326 - 350 of 461) sorted by null

<<111213141516171819

  /external/llvm/test/Transforms/GVN/
crash.ll 77 %tmp4 = getelementptr inbounds [4 x %struct.attribute_spec*]* @attribute_tables, i32 0, i32 undef ; <%struct.attribute_spec**> [#uses=1]
78 %tmp3 = load %struct.attribute_spec** %tmp4, align 4 ; <%struct.attribute_spec*> [#uses=1]
85 %tmp = load %struct.attribute_spec** %tmp4, align 4 ; <%struct.attribute_spec*> [#uses=1]
117 %tmp4 = bitcast i8* %tmp40.i to i64*
118 %tmp41.i = load i64* %tmp4
  /external/chromium_org/third_party/openssl/openssl/crypto/sha/asm/
sha1-ia64.pl 58 ld1 tmp4=[tmp3],2 // LSB
65 dep tmp1=tmp0,tmp4,8,8 };;
67 and tmp4=$c,$b
73 or tmp4=tmp4,tmp1 // F_00_19(b,c,d)=(b&c)|(~b&d)
76 add $e=$e,tmp4 // e+=F_00_19(b,c,d)
78 { .mmi; ld1 tmp4=[tmp3],2 // forward Xload
90 dep tmp1=tmp0,tmp4,8,8;;
96 or tmp4=tmp3,tmp1 // F_00_19(b,c,d)=(b&c)|(~b&d)
101 { .mmi; add $e=$e,tmp4 // e+=F_00_19(b,c,d
    [all...]
  /external/openssl/crypto/sha/asm/
sha1-ia64.pl 58 ld1 tmp4=[tmp3],2 // LSB
65 dep tmp1=tmp0,tmp4,8,8 };;
67 and tmp4=$c,$b
73 or tmp4=tmp4,tmp1 // F_00_19(b,c,d)=(b&c)|(~b&d)
76 add $e=$e,tmp4 // e+=F_00_19(b,c,d)
78 { .mmi; ld1 tmp4=[tmp3],2 // forward Xload
90 dep tmp1=tmp0,tmp4,8,8;;
96 or tmp4=tmp3,tmp1 // F_00_19(b,c,d)=(b&c)|(~b&d)
101 { .mmi; add $e=$e,tmp4 // e+=F_00_19(b,c,d
    [all...]
  /external/pixman/pixman/
pixman-arm-neon-asm.S     [all...]
  /external/llvm/test/Transforms/ObjCARC/
move-and-form-retain-autorelease.ll 84 %tmp4 = bitcast %15* %arg to i8*
85 %tmp5 = tail call %18* bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to %18* (i8*, i8*)*)(i8* %tmp4, i8* %tmp)
142 %tmp49 = tail call %22* bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to %22* (i8*, i8*)*)(i8* %tmp4, i8* %tmp48)
152 tail call void bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to void (i8*, i8*)*)(i8* %tmp4, i8* %tmp56)
158 %tmp60 = tail call %22* bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to %22* (i8*, i8*)*)(i8* %tmp4, i8* %tmp59)
178 %tmp76 = tail call %22* bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to %22* (i8*, i8*)*)(i8* %tmp4, i8* %tmp75)
202 %tmp97 = tail call signext i8 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i8 (i8*, i8*)*)(i8* %tmp4, i8* %tmp96)
  /external/llvm/test/CodeGen/ARM/
reg_sequence.ll 85 %tmp4 = extractvalue %struct.__neon_int8x8x3_t %tmp1, 1 ; <<8 x i8>> [#uses=1]
86 %tmp5 = sub <8 x i8> %tmp3, %tmp4
88 %tmp7 = mul <8 x i8> %tmp4, %tmp2
90 ret <8 x i8> %tmp4
104 %tmp4 = bitcast i32* %tmp3 to i8* ; <i8*> [#uses=1]
105 %tmp5 = tail call %struct.__neon_int32x4x2_t @llvm.arm.neon.vld2.v4i32(i8* %tmp4, i32 1) ; <%struct.__neon_int32x4x2_t> [#uses=2]
150 %tmp4 = extractvalue %struct.__neon_int16x8x2_t %tmp2, 1 ; <<8 x i16>> [#uses=1]
151 %tmp5 = add <8 x i16> %tmp3, %tmp4 ; <<8 x i16>> [#uses=1]
163 %tmp4 = extractvalue %struct.__neon_int8x8x2_t %tmp2, 1 ; <<8 x i8>> [#uses=1]
164 %tmp5 = add <8 x i8> %tmp3, %tmp4 ; <<8 x i8>> [#uses=1
    [all...]
vfp.ll 118 %tmp4 = load float* %tmp3 ; <float> [#uses=2]
119 %tmp.upgrd.1 = fcmp oeq float %tmp, %tmp4 ; <i1> [#uses=1]
120 %tmp5 = fcmp uno float %tmp, %tmp4 ; <i1> [#uses=1]
2010-05-20-NEONSpillCrash.ll 42 %tmp4 = sub <8 x i8> %tmp4efgh, %tmp4abcd ; <<8 x i8>> [#uses=1]
44 ret <8 x i8> %tmp4
2012-01-24-RegSequenceLiveRange.ll 13 %tmp4 = insertelement <4 x float> %tmp3, float 0.000000e+00, i32 1
14 %tmp5 = insertelement <4 x float> %tmp4, float 0.000000e+00, i32 2
vqdmul.ll 206 %tmp4 = call <4 x i32> @llvm.arm.neon.vqdmlal.v4i32(<4 x i32> %tmp1, <4 x i16> %tmp2, <4 x i16> %tmp3)
207 ret <4 x i32> %tmp4
216 %tmp4 = call <2 x i64> @llvm.arm.neon.vqdmlal.v2i64(<2 x i64> %tmp1, <2 x i32> %tmp2, <2 x i32> %tmp3)
217 ret <2 x i64> %tmp4
247 %tmp4 = call <4 x i32> @llvm.arm.neon.vqdmlsl.v4i32(<4 x i32> %tmp1, <4 x i16> %tmp2, <4 x i16> %tmp3)
248 ret <4 x i32> %tmp4
257 %tmp4 = call <2 x i64> @llvm.arm.neon.vqdmlsl.v2i64(<2 x i64> %tmp1, <2 x i32> %tmp2, <2 x i32> %tmp3)
258 ret <2 x i64> %tmp4
  /external/libvpx/libvpx/vp9/common/x86/
vp9_idct_intrin_sse2.c 396 tmp4 = _mm_madd_epi16(lo_1, cst2); \
405 tmp4 = _mm_add_epi32(tmp4, rounding); \
414 tmp4 = _mm_srai_epi32(tmp4, DCT_CONST_BITS); \
421 res2 = _mm_packs_epi32(tmp4, tmp5); \
520 __m128i tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7; local
637 __m128i tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7; local
977 __m128i tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7; local
1284 __m128i tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7; local
2476 __m128i tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7; local
3195 __m128i tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7; local
3554 __m128i tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7; local
    [all...]
  /external/jpeg/
jsimd_arm_neon.S 393 TMP4 .req ip
398 adr TMP4, jsimd_idct_4x4_neon_consts
399 vld1.16 {d0, d1, d2, d3}, [TMP4, :128]
451 ldmia OUTPUT_BUF, {TMP1, TMP2, TMP3, TMP4}
455 add TMP4, TMP4, OUTPUT_COL
464 vst1.32 {d27[1]}, [TMP4]!
476 vst1.8 {d27[4]}, [TMP4]!
478 vst1.8 {d27[5]}, [TMP4]!
480 vst1.8 {d27[6]}, [TMP4]!
    [all...]
  /external/llvm/test/CodeGen/Thumb/
2007-05-05-InvalidPushPop.ll 26 %tmp4 = call i32 @pthread_join( i32 %tmp2, i8** %ret3 ) ; <i32> [#uses=0]
  /external/llvm/test/CodeGen/X86/
2009-08-06-branchfolder-crash.ll 51 %tmp4 = load i8* @g_3 ; <i8> [#uses=1]
52 %conv5 = sext i8 %tmp4 to i32 ; <i32> [#uses=1]
101 %tmp4 = load i8* @g_3 ; <i8> [#uses=1]
102 %conv5 = sext i8 %tmp4 to i32 ; <i32> [#uses=1]
sink-hoist.ll 83 ; %tmp4 = and i8 %b, 127 ; <i8> [#uses=1]
84 ; %b_addr.0 = select i1 %tmp2, i8 %tmp4, i8 %tmp3 ; <i8> [#uses=1]
161 %tmp4 = load i32* @cl_options_count, align 4 ; <i32> [#uses=1]
162 %tmp5 = icmp eq i32 %tmp4, 0 ; <i1> [#uses=1]
2006-11-12-CSRetCC.ll 27 %tmp4 = getelementptr { double, double }* %tmp1, i64 0, i32 0 ; <double*> [#uses=1]
28 %tmp5 = load double* %tmp4 ; <double> [#uses=1]
2007-11-30-LoadFolding-Bug.ll 42 %tmp4.i19.i = icmp slt i32 %tmp1.i18.i, %radix ; <i1> [#uses=1]
43 %x.0.i21.i = select i1 %tmp4.i19.i, i32 %tmp1.i18.i, i32 0 ; <i32> [#uses=1]
2008-02-22-LocalRegAllocBug.ll 20 %tmp4 = load i32* %dst_stride_addr, align 4 ; <i32> [#uses=1]
21 %tmp5 = getelementptr i8* %tmp3, i32 %tmp4 ; <i8*> [#uses=1]
  /external/llvm/test/Transforms/Inline/
inline_minisize.ll 19 %tmp4 = load i32** @data, align 8
20 %arrayidx2 = getelementptr inbounds i32* %tmp4, i64 %idxprom1
119 %tmp4 = load i32** @data, align 8
120 %arrayidx3 = getelementptr inbounds i32* %tmp4, i64 %idxprom2
  /external/libvpx/libvpx/vp9/common/mips/dspr2/
vp9_intrapred8_dspr2.c 19 int32_t tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7, tmp8; local
25 "lb %[tmp4], 3(%[left]) \n\t"
34 "replv.qb %[tmp4], %[tmp4] \n\t"
49 "sw %[tmp4], (%[dst]) \n\t"
50 "sw %[tmp4], 4(%[dst]) \n\t"
65 [tmp3] "=&r" (tmp3), [tmp4] "=&r" (tmp4),
    [all...]
  /external/llvm/lib/Transforms/Utils/
IntegerDivision.cpp 103 // ; %tmp4 = xor i32 %q_mag, %q_sgn
104 // ; %q = sub i32 %tmp4, %q_sgn
113 Value *Tmp4 = Builder.CreateXor(Q_Mag, Q_Sgn);
114 Value *Q = Builder.CreateSub(Tmp4, Q_Sgn);
237 // ; %tmp4 = add i32 %divisor, -1
241 Value *Tmp4 = Builder.CreateAdd(Divisor, NegOne);
254 // ; %tmp9 = sub i32 %tmp4, %tmp7
272 Value *Tmp9 = Builder.CreateSub(Tmp4, Tmp7);
  /external/llvm/test/Analysis/ScalarEvolution/
SolveQuadraticEquation.ll 21 %tmp4 = mul i32 %i.0, 2
22 %tmp5 = sub i32 %SQ, %tmp4
  /external/llvm/test/CodeGen/Mips/
alloca.ll 14 %tmp4 = alloca i8, i32 %size, align 4
17 %call7 = call i32 @foo(i8* %tmp4) nounwind
  /external/llvm/test/CodeGen/PowerPC/
vec_constants.ll 8 %tmp4 = and <4 x i32> %tmp, < i32 -2147483648, i32 -2147483648, i32 -2147483648, i32 -2147483648 > ; <<4 x i32>> [#uses=1]
9 store <4 x i32> %tmp4, <4 x i32>* %P1
  /external/llvm/test/CodeGen/Thumb2/
2010-02-11-phi-cycle.ll 20 %tmp4 = zext i32 %1 to i64 ; <i64> [#uses=1]
22 %ins = or i64 %tmp4, %mask ; <i64> [#uses=2]

Completed in 972 milliseconds

<<111213141516171819