| /external/swiftshader/third_party/LLVM/test/CodeGen/PowerPC/ |
| 2007-11-16-landingpad-split.ll | 10 %tmp4 = call i8* @llvm.stacksave() ; <i8*> [#uses=1] 23 call void @llvm.stackrestore(i8* %tmp4)
|
| vec_constants.ll | 5 %tmp4 = and <4 x i32> %tmp, < i32 -2147483648, i32 -2147483648, i32 -2147483648, i32 -2147483648 > ; <<4 x i32>> [#uses=1]
6 store <4 x i32> %tmp4, <4 x i32>* %P1
|
| vec_misaligned.ll | 23 %tmp4 = load i8** %ap.0, align 4 ; <i8*> [#uses=1] 24 %tmp45 = bitcast i8* %tmp4 to %struct.S2203* ; <%struct.S2203*> [#uses=1]
|
| /external/swiftshader/third_party/LLVM/test/CodeGen/X86/ |
| 2008-01-16-InvalidDAGCombineXform.ll | 10 %tmp4 = load %struct.node_t** %cur_node, align 4 ; <%struct.node_t*> [#uses=1] 11 %tmp5 = getelementptr %struct.node_t* %tmp4, i32 0, i32 4 ; <double**> [#uses=1]
|
| 2010-06-25-CoalescerSubRegDefDead.ll | 12 %tmp4 = getelementptr inbounds %struct.type* %s, i32 0, i32 40 28 %tmp5 = bitcast i32* %tmp4 to i8*
|
| byval3.ll | 40 %tmp4 = getelementptr %struct.s* %d, i32 0, i32 2 41 store i32 %a3, i32* %tmp4, align 16
|
| byval4.ll | 46 %tmp4 = getelementptr %struct.s* %a, i32 0, i32 2 47 store i16 %a3, i16* %tmp4, align 16
|
| coalescer-commute4.ll | 17 %tmp4 = load i32* %tmp3, align 4 ; <i32> [#uses=1] 18 %tmp45 = sitofp i32 %tmp4 to float ; <float> [#uses=1]
|
| vec_shift.ll | 17 %tmp4 = insertelement <4 x i32> undef, i32 %c, i32 0 ; <<4 x i32>> [#uses=1] 18 %tmp8 = bitcast <4 x i32> %tmp4 to <8 x i16> ; <<8 x i16>> [#uses=1]
|
| vec_shuffle-14.ll | 23 %tmp4 = load <2 x i64>* %a, align 16 ; <<2 x i64>> [#uses=1] 24 %tmp6 = bitcast <2 x i64> %tmp4 to <4 x i32> ; <<4 x i32>> [#uses=1]
|
| widen_arith-1.ll | 30 %tmp4 = load i32* %i ; <i32> [#uses=1] 32 %arrayidx6 = getelementptr <3 x i8>* %tmp5, i32 %tmp4 ; <<3 x i8>*> [#uses=1]
|
| widen_cast-1.ll | 28 %tmp4 = load <4 x i16>** %src.addr ; <<4 x i16>*> [#uses=1] 29 %arrayidx5 = getelementptr <4 x i16>* %tmp4, i32 %tmp3 ; <<4 x i16>*> [#uses=1]
|
| widen_cast-2.ll | 30 %tmp4 = load <14 x i16>** %src.addr ; <<4 x i16>*> [#uses=1] 31 %arrayidx5 = getelementptr <14 x i16>* %tmp4, i32 %tmp3 ; <<4 x i16>*> [#uses=1]
|
| mmx-arith.ll | 12 %tmp4 = add <8 x i8> %tmp1a, %tmp3a ; <<8 x i8>> [#uses=2]
13 %tmp4a = bitcast <8 x i8> %tmp4 to x86_mmx
65 %tmp4 = add <2 x i32> %tmp1a, %tmp3a ; <<2 x i32>> [#uses=2]
66 %tmp4a = bitcast <2 x i32> %tmp4 to x86_mmx
70 %tmp10 = sub <2 x i32> %tmp4, %tmp9a ; <<2 x i32>> [#uses=2]
71 %tmp10a = bitcast <2 x i32> %tmp4 to x86_mmx
107 %tmp4 = add <4 x i16> %tmp1a, %tmp3a ; <<4 x i16>> [#uses=2]
108 %tmp4a = bitcast <4 x i16> %tmp4 to x86_mmx
169 %tmp4 = tail call x86_mmx @llvm.x86.mmx.padd.b( x86_mmx %tmp1, x86_mmx %tmp3 ) ; <x86_mmx> [#uses=2]
170 store x86_mmx %tmp4, x86_mmx* %A [all...] |
| /external/swiftshader/third_party/LLVM/test/Transforms/GVN/ |
| nonescaping-malloc.ll | 47 %tmp4.i = getelementptr inbounds %"struct.llvm::StringMap<void*,llvm::MallocAllocator>"* %this, i64 0, i32 0, i32 0 48 %tmp5.i = load %"struct.llvm::StringMapImpl::ItemBucket"** %tmp4.i, align 8 59 %tmp4.i35.i = getelementptr inbounds i8* %tmp41.i, i64 %val2 60 %tmp.i.i = ptrtoint i8* %tmp4.i35.i to i64 63 %tmp4.i.i = sub i32 %tmp1.i.i, %tmp3.i.i 64 %tmp5.i.i = add i32 %tmp4.i.i, 17 73 store i32 %tmp4.i.i, i32* %tmp.i.i.i.i, align 4 81 %tmp15.i.i = zext i32 %tmp4.i.i to i64
|
| /external/swiftshader/third_party/LLVM/test/Transforms/IndVarSimplify/ |
| polynomial-expand.ll | 7 define void @ctpmv_(float* noalias nocapture %tmp4, i32 %tmp21) nounwind { 23 %tmp35 = getelementptr float* %tmp4, i64 %tmp33 ; <%0*> [#uses=1]
|
| /external/swiftshader/third_party/LLVM/test/Transforms/LoopStrengthReduce/ |
| count-to-zero.ll | 19 %tmp4 = add i32 %c_addr.1, -1 ; <i32> [#uses=1] 20 %c_addr.1.be = select i1 %tmp2, i32 %tmp3, i32 %tmp4 ; <i32> [#uses=1]
|
| /external/swiftshader/third_party/LLVM/test/Transforms/Mem2Reg/ |
| 2007-08-27-VolatileLoadsStores.ll | 25 %tmp4 = volatile load i32* %v, align 4 ; <i32> [#uses=1] 26 store i32 %tmp4, i32* %tmp, align 4
|
| /external/swiftshader/third_party/LLVM/test/Transforms/MemCpyOpt/ |
| 2008-02-24-MultipleUseofSRet.ll | 30 %tmp4 = bitcast %0* %tmp to i8* 31 call void @llvm.memcpy.p0i8.p0i8.i32(i8* %z3, i8* %tmp4, i32 24, i32 8, i1 false)
|
| sret.ll | 15 %tmp4 = getelementptr %0* %iz, i32 0, i32 1 20 store x86_fp80 %tmp8, x86_fp80* %tmp4, align 16
|
| /external/swiftshader/third_party/LLVM/test/Transforms/ScalarRepl/ |
| union-pointer.ll | 29 %tmp4 = getelementptr %struct.Val* %tmp1, i32 0, i32 1 ; <i32*> [#uses=1]
30 %tmp.upgrd.6 = load i32* %tmp4 ; <i32> [#uses=1]
|
| /external/llvm/test/CodeGen/AArch64/ |
| arm64-vadd.ll | 132 %tmp4 = sext <8 x i8> %tmp2 to <8 x i16> 133 %tmp5 = add <8 x i16> %tmp3, %tmp4 143 %tmp4 = sext <4 x i16> %tmp2 to <4 x i32> 144 %tmp5 = add <4 x i32> %tmp3, %tmp4 154 %tmp4 = sext <2 x i32> %tmp2 to <2 x i64> 155 %tmp5 = add <2 x i64> %tmp3, %tmp4 213 %tmp4 = zext <8 x i8> %tmp2 to <8 x i16> 214 %tmp5 = add <8 x i16> %tmp3, %tmp4 224 %tmp4 = zext <4 x i16> %tmp2 to <4 x i32> 225 %tmp5 = add <4 x i32> %tmp3, %tmp4 [all...] |
| /external/webrtc/webrtc/modules/audio_processing/aecm/ |
| aecm_core_mips.c | 81 int32_t tmp1, tmp2, tmp3, tmp4; local 106 "lh %[tmp4], 0(%[hann1]) \n\t" 109 "mul %[tmp3], %[tmp3], %[tmp4] \n\t" 129 "lh %[tmp4], 0(%[hann1]) \n\t" 132 "mul %[tmp3], %[tmp3], %[tmp4] \n\t" 151 [tmp4] "=&r" (tmp4), [i] "=&r" (i), [f_coef] "=&r" (f_coef), 172 "lh %[tmp4], 6(%[pfrfi]) \n\t" 176 "subu %[tmp4], $zero, %[tmp4] \n\t 208 int32_t tmp1, tmp2, tmp3, tmp4, tmp_re, tmp_im; local [all...] |
| /frameworks/av/media/libstagefright/codecs/mp3dec/src/ |
| pvmp3_mdct_18.cpp | 140 int32 tmp4; local 168 tmp4 = vec[14]; // 182 vec[ 9] = vec[13] - tmp4; 183 vec[11] = vec[15] - tmp4; 196 tmp4 = vec[i+10]; 197 vec[i+10] = tmp3 + tmp4; 200 tmp3 = tmp4; 206 tmp4 = vec[16]; 207 vec[16] = tmp3 + tmp4; 215 tmp4 = vec[17] + tmp4 [all...] |
| /external/libjpeg-turbo/simd/ |
| jfdctfst-altivec.c | 66 tmp10 = vec_add(tmp4, tmp5); \ 96 tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7, tmp10, tmp11, tmp12, tmp13, local 129 tmp4 = vec_sub(col3, col4); 144 tmp4 = vec_sub(row3, row4);
|