HomeSort by relevance Sort by last modified time
    Searched full:tmp4 (Results 526 - 550 of 1043) sorted by null

<<21222324252627282930>>

  /external/swiftshader/third_party/LLVM/test/CodeGen/PowerPC/
2007-11-16-landingpad-split.ll 10 %tmp4 = call i8* @llvm.stacksave() ; <i8*> [#uses=1]
23 call void @llvm.stackrestore(i8* %tmp4)
vec_constants.ll 5 %tmp4 = and <4 x i32> %tmp, < i32 -2147483648, i32 -2147483648, i32 -2147483648, i32 -2147483648 > ; <<4 x i32>> [#uses=1]
6 store <4 x i32> %tmp4, <4 x i32>* %P1
vec_misaligned.ll 23 %tmp4 = load i8** %ap.0, align 4 ; <i8*> [#uses=1]
24 %tmp45 = bitcast i8* %tmp4 to %struct.S2203* ; <%struct.S2203*> [#uses=1]
  /external/swiftshader/third_party/LLVM/test/CodeGen/X86/
2008-01-16-InvalidDAGCombineXform.ll 10 %tmp4 = load %struct.node_t** %cur_node, align 4 ; <%struct.node_t*> [#uses=1]
11 %tmp5 = getelementptr %struct.node_t* %tmp4, i32 0, i32 4 ; <double**> [#uses=1]
2010-06-25-CoalescerSubRegDefDead.ll 12 %tmp4 = getelementptr inbounds %struct.type* %s, i32 0, i32 40
28 %tmp5 = bitcast i32* %tmp4 to i8*
byval3.ll 40 %tmp4 = getelementptr %struct.s* %d, i32 0, i32 2
41 store i32 %a3, i32* %tmp4, align 16
byval4.ll 46 %tmp4 = getelementptr %struct.s* %a, i32 0, i32 2
47 store i16 %a3, i16* %tmp4, align 16
coalescer-commute4.ll 17 %tmp4 = load i32* %tmp3, align 4 ; <i32> [#uses=1]
18 %tmp45 = sitofp i32 %tmp4 to float ; <float> [#uses=1]
vec_shift.ll 17 %tmp4 = insertelement <4 x i32> undef, i32 %c, i32 0 ; <<4 x i32>> [#uses=1]
18 %tmp8 = bitcast <4 x i32> %tmp4 to <8 x i16> ; <<8 x i16>> [#uses=1]
vec_shuffle-14.ll 23 %tmp4 = load <2 x i64>* %a, align 16 ; <<2 x i64>> [#uses=1]
24 %tmp6 = bitcast <2 x i64> %tmp4 to <4 x i32> ; <<4 x i32>> [#uses=1]
widen_arith-1.ll 30 %tmp4 = load i32* %i ; <i32> [#uses=1]
32 %arrayidx6 = getelementptr <3 x i8>* %tmp5, i32 %tmp4 ; <<3 x i8>*> [#uses=1]
widen_cast-1.ll 28 %tmp4 = load <4 x i16>** %src.addr ; <<4 x i16>*> [#uses=1]
29 %arrayidx5 = getelementptr <4 x i16>* %tmp4, i32 %tmp3 ; <<4 x i16>*> [#uses=1]
widen_cast-2.ll 30 %tmp4 = load <14 x i16>** %src.addr ; <<4 x i16>*> [#uses=1]
31 %arrayidx5 = getelementptr <14 x i16>* %tmp4, i32 %tmp3 ; <<4 x i16>*> [#uses=1]
mmx-arith.ll 12 %tmp4 = add <8 x i8> %tmp1a, %tmp3a ; <<8 x i8>> [#uses=2]
13 %tmp4a = bitcast <8 x i8> %tmp4 to x86_mmx
65 %tmp4 = add <2 x i32> %tmp1a, %tmp3a ; <<2 x i32>> [#uses=2]
66 %tmp4a = bitcast <2 x i32> %tmp4 to x86_mmx
70 %tmp10 = sub <2 x i32> %tmp4, %tmp9a ; <<2 x i32>> [#uses=2]
71 %tmp10a = bitcast <2 x i32> %tmp4 to x86_mmx
107 %tmp4 = add <4 x i16> %tmp1a, %tmp3a ; <<4 x i16>> [#uses=2]
108 %tmp4a = bitcast <4 x i16> %tmp4 to x86_mmx
169 %tmp4 = tail call x86_mmx @llvm.x86.mmx.padd.b( x86_mmx %tmp1, x86_mmx %tmp3 ) ; <x86_mmx> [#uses=2]
170 store x86_mmx %tmp4, x86_mmx* %A
    [all...]
  /external/swiftshader/third_party/LLVM/test/Transforms/GVN/
nonescaping-malloc.ll 47 %tmp4.i = getelementptr inbounds %"struct.llvm::StringMap<void*,llvm::MallocAllocator>"* %this, i64 0, i32 0, i32 0
48 %tmp5.i = load %"struct.llvm::StringMapImpl::ItemBucket"** %tmp4.i, align 8
59 %tmp4.i35.i = getelementptr inbounds i8* %tmp41.i, i64 %val2
60 %tmp.i.i = ptrtoint i8* %tmp4.i35.i to i64
63 %tmp4.i.i = sub i32 %tmp1.i.i, %tmp3.i.i
64 %tmp5.i.i = add i32 %tmp4.i.i, 17
73 store i32 %tmp4.i.i, i32* %tmp.i.i.i.i, align 4
81 %tmp15.i.i = zext i32 %tmp4.i.i to i64
  /external/swiftshader/third_party/LLVM/test/Transforms/IndVarSimplify/
polynomial-expand.ll 7 define void @ctpmv_(float* noalias nocapture %tmp4, i32 %tmp21) nounwind {
23 %tmp35 = getelementptr float* %tmp4, i64 %tmp33 ; <%0*> [#uses=1]
  /external/swiftshader/third_party/LLVM/test/Transforms/LoopStrengthReduce/
count-to-zero.ll 19 %tmp4 = add i32 %c_addr.1, -1 ; <i32> [#uses=1]
20 %c_addr.1.be = select i1 %tmp2, i32 %tmp3, i32 %tmp4 ; <i32> [#uses=1]
  /external/swiftshader/third_party/LLVM/test/Transforms/Mem2Reg/
2007-08-27-VolatileLoadsStores.ll 25 %tmp4 = volatile load i32* %v, align 4 ; <i32> [#uses=1]
26 store i32 %tmp4, i32* %tmp, align 4
  /external/swiftshader/third_party/LLVM/test/Transforms/MemCpyOpt/
2008-02-24-MultipleUseofSRet.ll 30 %tmp4 = bitcast %0* %tmp to i8*
31 call void @llvm.memcpy.p0i8.p0i8.i32(i8* %z3, i8* %tmp4, i32 24, i32 8, i1 false)
sret.ll 15 %tmp4 = getelementptr %0* %iz, i32 0, i32 1
20 store x86_fp80 %tmp8, x86_fp80* %tmp4, align 16
  /external/swiftshader/third_party/LLVM/test/Transforms/ScalarRepl/
union-pointer.ll 29 %tmp4 = getelementptr %struct.Val* %tmp1, i32 0, i32 1 ; <i32*> [#uses=1]
30 %tmp.upgrd.6 = load i32* %tmp4 ; <i32> [#uses=1]
  /external/llvm/test/CodeGen/AArch64/
arm64-vadd.ll 132 %tmp4 = sext <8 x i8> %tmp2 to <8 x i16>
133 %tmp5 = add <8 x i16> %tmp3, %tmp4
143 %tmp4 = sext <4 x i16> %tmp2 to <4 x i32>
144 %tmp5 = add <4 x i32> %tmp3, %tmp4
154 %tmp4 = sext <2 x i32> %tmp2 to <2 x i64>
155 %tmp5 = add <2 x i64> %tmp3, %tmp4
213 %tmp4 = zext <8 x i8> %tmp2 to <8 x i16>
214 %tmp5 = add <8 x i16> %tmp3, %tmp4
224 %tmp4 = zext <4 x i16> %tmp2 to <4 x i32>
225 %tmp5 = add <4 x i32> %tmp3, %tmp4
    [all...]
  /external/webrtc/webrtc/modules/audio_processing/aecm/
aecm_core_mips.c 81 int32_t tmp1, tmp2, tmp3, tmp4; local
106 "lh %[tmp4], 0(%[hann1]) \n\t"
109 "mul %[tmp3], %[tmp3], %[tmp4] \n\t"
129 "lh %[tmp4], 0(%[hann1]) \n\t"
132 "mul %[tmp3], %[tmp3], %[tmp4] \n\t"
151 [tmp4] "=&r" (tmp4), [i] "=&r" (i), [f_coef] "=&r" (f_coef),
172 "lh %[tmp4], 6(%[pfrfi]) \n\t"
176 "subu %[tmp4], $zero, %[tmp4] \n\t
208 int32_t tmp1, tmp2, tmp3, tmp4, tmp_re, tmp_im; local
    [all...]
  /frameworks/av/media/libstagefright/codecs/mp3dec/src/
pvmp3_mdct_18.cpp 140 int32 tmp4; local
168 tmp4 = vec[14]; //
182 vec[ 9] = vec[13] - tmp4;
183 vec[11] = vec[15] - tmp4;
196 tmp4 = vec[i+10];
197 vec[i+10] = tmp3 + tmp4;
200 tmp3 = tmp4;
206 tmp4 = vec[16];
207 vec[16] = tmp3 + tmp4;
215 tmp4 = vec[17] + tmp4
    [all...]
  /external/libjpeg-turbo/simd/
jfdctfst-altivec.c 66 tmp10 = vec_add(tmp4, tmp5); \
96 tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7, tmp10, tmp11, tmp12, tmp13, local
129 tmp4 = vec_sub(col3, col4);
144 tmp4 = vec_sub(row3, row4);

Completed in 1183 milliseconds

<<21222324252627282930>>