HomeSort by relevance Sort by last modified time
    Searched full:tmp6 (Results 51 - 75 of 329) sorted by null

1 23 4 5 6 7 8 91011>>

  /external/llvm/test/Analysis/ScalarEvolution/
2007-08-06-Unsigned.ll 20 %tmp6 = icmp ult i32 %tmp2, %y ; <i1> [#uses=1]
21 br i1 %tmp6, label %bb, label %bb8.loopexit
trip-count2.ll 18 %tmp6 = trunc i32 %tmp61 to i16
19 %tmp71 = shl i16 %tmp6, 2
  /external/llvm/test/CodeGen/AArch64/
merge-store.ll 13 %tmp6 = extractelement <3 x float> %tmp5, i64 0
14 store float %tmp6, float* %tmp4
  /external/llvm/test/CodeGen/ARM/
2010-06-11-vmovdrr-bitcast.ll 10 %tmp6 = trunc i128 %srcval to i64 ; <i64> [#uses=1]
13 %tmp16.i = bitcast i64 %tmp6 to <8 x i8> ; <<8 x i8>> [#uses=1]
aliases.ll 43 %tmp6 = add i32 %tmp1, %tmp5
44 %tmp7 = add i32 %tmp6, %tmp0
vtbl.ll 35 %tmp6 = call <8 x i8> @llvm.arm.neon.vtbl3(<8 x i8> %tmp1, <8 x i8> %tmp3, <8 x i8> %tmp4, <8 x i8> %tmp5)
36 ret <8 x i8> %tmp6
47 %tmp6 = extractvalue %struct.__neon_int8x8x4_t %tmp2, 3
48 %tmp7 = call <8 x i8> @llvm.arm.neon.vtbl4(<8 x i8> %tmp1, <8 x i8> %tmp3, <8 x i8> %tmp4, <8 x i8> %tmp5, <8 x i8> %tmp6)
70 %tmp6 = call <8 x i8> @llvm.arm.neon.vtbx2(<8 x i8> %tmp1, <8 x i8> %tmp3, <8 x i8> %tmp4, <8 x i8> %tmp5)
71 ret <8 x i8> %tmp6
82 %tmp6 = load <8 x i8>, <8 x i8>* %C
83 %tmp7 = call <8 x i8> @llvm.arm.neon.vtbx3(<8 x i8> %tmp1, <8 x i8> %tmp3, <8 x i8> %tmp4, <8 x i8> %tmp5, <8 x i8> %tmp6)
95 %tmp6 = extractvalue %struct.__neon_int8x8x4_t %tmp2, 3
97 %tmp8 = call <8 x i8> @llvm.arm.neon.vtbx4(<8 x i8> %tmp1, <8 x i8> %tmp3, <8 x i8> %tmp4, <8 x i8> %tmp5, <8 x i8> %tmp6, <8 x i8> %tmp7
    [all...]
  /external/llvm/test/CodeGen/Hexagon/vect/
vect-splat.ll 11 %tmp6 = insertelement %i4 %tmp4, i32 %X, i32 3 ; <%i4> [#uses=1]
13 %R = add %i4 %q, %tmp6 ; <%i4> [#uses=1]
  /external/llvm/test/CodeGen/PowerPC/
2007-01-15-AsmDialect.ll 19 %tmp6 = load i32, i32* %temp ; <i32> [#uses=1]
20 store i32 %tmp6, i32* %retval
2007-10-18-PtrArithmetic.ll 7 %tmp6 = icmp eq i8 %.mask, 16 ; <i1> [#uses=1]
8 br i1 %tmp6, label %cond_true, label %UnifiedReturnBlock
rlwinm2.ll 17 %tmp6 = and i32 %tmp4, 127 ; <i32> [#uses=1]
18 ret i32 %tmp6
  /external/llvm/test/CodeGen/X86/
2006-05-08-CoalesceSubRegClass.ll 17 %tmp6 = lshr i32 %A, 3 ; <i32> [#uses=1]
20 %tmp9 = add i32 %tmp8, %tmp6 ; <i32> [#uses=1]
2006-07-31-SingleRegClass.ll 7 define i32 @foo(i32 %__s.i.i, i32 %tmp5.i.i, i32 %tmp6.i.i, i32 %tmp7.i.i, i32 %tmp8.i.i) {
8 %tmp9.i.i = call i32 asm sideeffect "push %ebp\0Apush %ebx\0Amovl 4($2),%ebp\0Amovl 0($2), %ebx\0Amovl $1,%eax\0Aint $$0x80\0Apop %ebx\0Apop %ebp", "={ax},i,0,{cx},{dx},{si},{di}"( i32 192, i32 %__s.i.i, i32 %tmp5.i.i, i32 %tmp6.i.i, i32 %tmp7.i.i, i32 %tmp8.i.i ) ; <i32> [#uses=1]
2007-02-04-OrAddrMode.ll 14 %tmp6 = ptrtoint i8* %ctg2 to i32 ; <i32> [#uses=1]
15 %tmp14 = or i32 %tmp6, 1 ; <i32> [#uses=1]
2007-02-19-LiveIntervalAssert.ll 11 %tmp6 = load %struct._IO_FILE*, %struct._IO_FILE** @stderr
12 %tmp7 = tail call i32 @fflush( %struct._IO_FILE* %tmp6 )
2007-05-15-maskmovq.ll 9 %tmp6 = bitcast <1 x i64> %c64 to x86_mmx ; <x86_mmx> [#uses=1]
10 tail call void @llvm.x86.mmx.maskmovq( x86_mmx %tmp4, x86_mmx %tmp6, i8* %P )
2007-07-03-GR64ToVR64.ll 12 %tmp6 = bitcast <1 x i64> %A to x86_mmx ; <<4 x i16>> [#uses=1]
13 %tmp7 = tail call x86_mmx @llvm.x86.mmx.paddus.w( x86_mmx %tmp6, x86_mmx %tmp4 ) ; <x86_mmx> [#uses=1]
and-or-fold.ll 12 %tmp6 = or i32 %tmp2, %tmp5
13 ret i32 %tmp6
commute-intrinsic.ll 10 %tmp6 = bitcast <2 x i64> %b to <8 x i16> ; <<8 x i16>> [#uses=1]
12 %tmp11 = tail call <4 x i32> @llvm.x86.sse2.pmadd.wd( <8 x i16> %tmp9, <8 x i16> %tmp6 ) nounwind readnone ; <<4 x i32>> [#uses=1]
fastcc-sret.ll 22 %tmp6 = load i32, i32* %tmp5
23 store i32 %tmp6, i32* @dst
tailcall-multiret.ll 13 %tmp6 = insertelement <3 x double> %tmp4, double %tmp5, i32 2
15 ret <3 x double> %tmp6
vec_set.ll 7 %tmp6 = insertelement <8 x i16> %tmp4, i16 %a3, i32 3 ; <<8 x i16>> [#uses=1]
8 %tmp8 = insertelement <8 x i16> %tmp6, i16 %a4, i32 4 ; <<8 x i16>> [#uses=1]
  /external/llvm/test/Transforms/InstCombine/
bswap.ll 23 %tmp6 = or i32 %tmp2, %tmp5 ; <i32> [#uses=1]
26 %tmp10 = or i32 %tmp6, %tmp9 ; <i32> [#uses=1]
55 %tmp6 = bitcast i16 %tmp.upgrd.3 to i16 ; <i16> [#uses=1]
56 %tmp6.upgrd.4 = zext i16 %tmp6 to i32 ; <i32> [#uses=1]
57 %retval = trunc i32 %tmp6.upgrd.4 to i16 ; <i16> [#uses=1]
70 %tmp6 = lshr i32 %x, 24 ; <i32> [#uses=1]
71 %tmp7 = or i32 %tmp5, %tmp6 ; <i32> [#uses=1]
known_align.ll 18 %tmp6 = load i32, i32* %tmp1, align 4 ; <i32> [#uses=1]
19 store i32 %tmp6, i32* %tmp, align 4
  /frameworks/av/media/libstagefright/codecs/on2/h264dec/source/arm11_asm/
h264bsd_interpolate_ver_half.s 56 tmp6 RN 9 label
103 ADD tmp6, y0, partH ;// (y0+partHeight)
104 ADD tmp6, tmp6, #5 ;// (y0+partH+5)
106 CMP tmp6, height
134 MLA tmp6, width, y0, x0 ;// y0*width+x0
135 ADD ref, ref, tmp6 ;// ref += y0*width+x0
155 LDR tmp6, [ref], width ;// |t4|t3|t2|t1|
166 UXTAB16 tmpa, tmpa, tmp6 ;// 16+20(G+M)+A+T
182 UXTAB16 tmpa, tmpa, tmp6, ROR #8 ;// 16+20(G+M)+A+
    [all...]
  /external/jpeg/
jfdctflt.c 61 FAST_FLOAT tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7; local
74 tmp6 = dataptr[1] - dataptr[6];
97 tmp11 = tmp5 + tmp6;
98 tmp12 = tmp6 + tmp7;
124 tmp6 = dataptr[DCTSIZE*1] - dataptr[DCTSIZE*6];
147 tmp11 = tmp5 + tmp6;
148 tmp12 = tmp6 + tmp7;

Completed in 399 milliseconds

1 23 4 5 6 7 8 91011>>