/external/llvm/test/CodeGen/X86/ |
misched-new.ll | 94 %tmp6 = sub i8 0, %tmp 96 %tmp8 = udiv i8 %tmp6, %tmp7 101 %tmp13 = urem i8 %tmp6, %tmp7
|
2009-02-11-codegenprepare-reuse.ll | 22 %tmp6.i569 = load i8*, i8** %tmp5.i424 ; <i8*> [#uses=0]
|
inline-asm-mrv.ll | 14 %tmp6 = extractvalue {i32, i8*, i8*} %mrv, 0
|
sse3.ll | 19 %tmp6 = shufflevector <8 x i16> %tmp3, 22 store <8 x i16> %tmp6, <8 x i16>* %dest 124 %tmp6 = extractelement <8 x i16> %tmp.upgrd.1, i32 6 130 %tmp12 = insertelement <8 x i16> %tmp11, i16 %tmp6, i32 4 151 %tmp6 = bitcast <2 x double> %tmp5 to <4 x float> 154 %tmp8 = extractelement <4 x float> %tmp6, i32 0 155 %tmp9 = extractelement <4 x float> %tmp6, i32 1 284 %tmp6 = shufflevector <4 x i32> %tmp5, <4 x i32> undef, <4 x i32> <i32 undef, i32 undef, i32 0, i32 1> 285 %tmp7 = and <4 x i32> %tmp6, <i32 undef, i32 undef, i32 -1, i32 0>
|
/external/llvm/test/Transforms/IndVarSimplify/ |
2008-09-02-IVType.ll | 33 %tmp6 = getelementptr %struct.JPEGGlobals, %struct.JPEGGlobals* %globp, i32 0, i32 77, i32 %tmp12 ; <i16**> [#uses=2] 34 %tmp7 = load i16*, i16** %tmp6, align 4 ; <i16*> [#uses=2] 54 store i16* %theDCTBufferIter.0.lcssa, i16** %tmp6, align 4
|
/external/llvm/test/Transforms/LoopStrengthReduce/ |
2013-01-14-ReuseCast.ll | 18 ; CHECK: sub i64 [[TMP]], %tmp6 25 %tmp6 = ptrtoint i8* %tmp3 to i64 26 %tmp7 = sub i64 %tmp5, %tmp6
|
/external/llvm/test/Transforms/ObjCARC/ |
escape.ll | 13 ; CHECK: %tmp7 = call i8* @objc_retainBlock(i8* %tmp6) [[NUW:#[0-9]+]], !clang.arc.copy_on_escape !0 48 %tmp6 = bitcast <{ i8*, i32, i32, i8*, %struct.__block_descriptor*, i8* }>* %block to i8* 49 %tmp7 = call i8* @objc_retainBlock(i8* %tmp6) nounwind, !clang.arc.copy_on_escape !0 103 %tmp6 = bitcast <{ i8*, i32, i32, i8*, %struct.__block_descriptor*, i8* }>* %block to i8* 104 %tmp7 = call i8* @objc_retainBlock(i8* %tmp6) nounwind, !clang.arc.copy_on_escape !0
|
/frameworks/av/media/libstagefright/codecs/on2/h264dec/source/arm11_asm/ |
h264bsd_interpolate_chroma_ver.s | 64 tmp6 RN 12 label 181 LDRB tmp6, [ptrA,width, LSL #1] ;// 3 row, 2 col 189 PKHBT tmp5, tmp5, tmp6, LSL #16 ;// |C|B| 248 LDRB tmp6, [ptrA,width, LSL #1] ;// 3 row, 2 col 256 PKHBT tmp5, tmp5, tmp6, LSL #16 ;// |C|B|
|
/external/libjpeg-turbo/simd/ |
jfdctflt-sse-64.asm | 128 subps xmm7,xmm2 ; xmm7=data1-data6=tmp6 135 movaps XMMWORD [wk(0)], xmm7 ; wk(0)=tmp6 178 movaps xmm6, XMMWORD [wk(0)] ; xmm6=tmp6 263 subps xmm7,xmm2 ; xmm7=data1-data6=tmp6 270 movaps XMMWORD [wk(0)], xmm7 ; wk(0)=tmp6 313 movaps xmm6, XMMWORD [wk(0)] ; xmm6=tmp6
|
jfdctflt-sse.asm | 135 subps xmm7,xmm2 ; xmm7=data1-data6=tmp6 142 movaps XMMWORD [wk(0)], xmm7 ; wk(0)=tmp6 185 movaps xmm6, XMMWORD [wk(0)] ; xmm6=tmp6 271 subps xmm7,xmm2 ; xmm7=data1-data6=tmp6 278 movaps XMMWORD [wk(0)], xmm7 ; wk(0)=tmp6 321 movaps xmm6, XMMWORD [wk(0)] ; xmm6=tmp6
|
jfdctfst-mmx.asm | 150 psubw mm7,mm2 ; mm7=data1-data6=tmp6 157 movq MMWORD [wk(0)], mm7 ; wk(0)=tmp6 201 movq mm6, MMWORD [wk(0)] ; mm6=tmp6 291 psubw mm7,mm2 ; mm7=data1-data6=tmp6 298 movq MMWORD [wk(0)], mm7 ; wk(0)=tmp6 342 movq mm6, MMWORD [wk(0)] ; mm6=tmp6
|
/external/libvpx/libvpx/vpx_dsp/mips/ |
idct16x16_msa.c | 17 v8i16 tmp5, tmp6, tmp7; local 61 tmp6 = loc0; 92 reg14 = tmp6; 111 v8i16 tmp5, tmp6, tmp7; local 158 tmp6 = loc0; 188 reg14 = tmp6;
|
intrapred8_dspr2.c | 16 int32_t tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7, tmp8; local 24 "lb %[tmp6], 5(%[left]) \n\t" 33 "replv.qb %[tmp6], %[tmp6] \n\t" 52 "sw %[tmp6], (%[dst]) \n\t" 53 "sw %[tmp6], 4(%[dst]) \n\t" 64 [tmp6] "=&r" (tmp6), [tmp8] "=&r" (tmp8) [all...] |
/external/llvm/test/CodeGen/AMDGPU/ |
valu-i1.ll | 158 %tmp6 = load i32, i32 addrspace(1)* %tmp5, align 4 159 %tmp7 = icmp sgt i32 %tmp6, 0 160 %tmp8 = sext i32 %tmp6 to i64
|
/frameworks/av/media/libstagefright/codecs/mp3dec/src/asm/ |
pvmp3_dct_9_gcc.s | 146 @ vec[3] = fxp_mul32_Q32((tmp5 + tmp6 - tmp8)<<1, cos_pi_6)@ 156 @ vec[5] = fxp_mac32_Q32( vec[5], tmp6<<1, cos_7pi_18)@ 167 @ vec[7] = fxp_mac32_Q32( vec[7], tmp6<<1, cos_17pi_18)@
|
/external/llvm/test/Analysis/ScalarEvolution/ |
nsw.ll | 25 %tmp6 = sext i32 %i.01 to i64 ; <i64> [#uses=1] 26 %tmp7 = getelementptr double, double* %p, i64 %tmp6 ; <double*> [#uses=1] 138 %tmp6 = icmp ult i32* %tmp5, %tmp 139 br i1 %tmp6, label %bb1, label %bb7
|
/external/llvm/test/CodeGen/Generic/ |
2006-03-01-dagcombineinfloop.ll | 82 %t.0.0 = phi %struct.tree_node* [ %parms, %cond_true92.preheader ], [ %tmp6, %cond_true34 ], [ %tmp6, %cond_next84 ] ; <%struct.tree_node*> [#uses=2] 88 %tmp6 = load %struct.tree_node*, %struct.tree_node** %tmp5 ; <%struct.tree_node*> [#uses=3] 89 %tmp.upgrd.6 = icmp eq %struct.tree_node* %tmp6, null ; <i1> [#uses=3]
|
vector.ll | 137 %tmp6 = insertelement %f4 %tmp4, float %X, i32 3 ; <%f4> [#uses=1] 139 %R = fadd %f4 %q, %tmp6 ; <%f4> [#uses=1] 148 %tmp6 = insertelement %i4 %tmp4, i32 %X, i32 3 ; <%i4> [#uses=1] 150 %R = add %i4 %q, %tmp6 ; <%i4> [#uses=1]
|
/external/llvm/test/Transforms/InstCombine/ |
2008-06-08-ICmpPHI.ll | 17 %tmp6 = tail call i32 (...) @bork( ) nounwind ; <i32> [#uses=0]
|
/external/llvm/test/Transforms/JumpThreading/ |
2010-08-26-and.ll | 33 %tmp6 = load i8*, i8** %arrayidx, align 8 ; <i8*> [#uses=8] 34 %call.i.i = call i64 @strlen(i8* %tmp6) nounwind ; <i64> [#uses=1] 45 %call.i = call i32 @memcmp(i8* getelementptr inbounds ([4 x i8], [4 x i8]* @.str, i64 0, i64 0), i8* %tmp6, i64 4) nounwind ; <i32> [#uses=1] 53 %call.i35 = call i32 @memcmp(i8* getelementptr inbounds ([7 x i8], [7 x i8]* @.str1, i64 0, i64 0), i8* %tmp6, i64 7) nounwind ; <i32> [#uses=1] 69 %call.i53 = call i32 @memcmp(i8* getelementptr inbounds ([7 x i8], [7 x i8]* @.str2, i64 0, i64 0), i8* %tmp6, i64 7) nounwind ; <i32> [#uses=1] 86 %call.i72 = call i32 @memcmp(i8* getelementptr inbounds ([6 x i8], [6 x i8]* @.str3, i64 0, i64 0), i8* %tmp6, i64 6) nounwind ; <i32> [#uses=1] 103 %call.i90 = call i32 @memcmp(i8* getelementptr inbounds ([5 x i8], [5 x i8]* @.str4, i64 0, i64 0), i8* %tmp6, i64 5) nounwind ; <i32> [#uses=1] 120 %call.i108 = call i32 @memcmp(i8* getelementptr inbounds ([7 x i8], [7 x i8]* @.str5, i64 0, i64 0), i8* %tmp6, i64 7) nounwind ; <i32> [#uses=1] 137 %call.i127 = call i32 @memcmp(i8* getelementptr inbounds ([7 x i8], [7 x i8]* @.str6, i64 0, i64 0), i8* %tmp6, i64 7) nounwind ; <i32> [#uses=1]
|
/external/llvm/test/Transforms/ScalarRepl/ |
vector_promote.ll | 14 %tmp6 = fadd <4 x float> %tmp4, %tmp4 ; <<4 x float>> [#uses=1] 15 store <4 x float> %tmp6, <4 x float>* %F 33 %tmp6 = fadd <4 x float> %tmp4, %tmp4 ; <<4 x float>> [#uses=1] 34 store <4 x float> %tmp6, <4 x float>* %F
|
/external/llvm/test/Transforms/SimplifyCFG/ |
2006-12-08-Ptr-ICmp-Branch.ll | 64 %tmp6 = call i32 @tolower( i32 %tmp.upgrd.9 ) ; <i32> [#uses=1] 65 %tmp6.upgrd.10 = trunc i32 %tmp6 to i8 ; <i8> [#uses=1] 66 store i8 %tmp6.upgrd.10, i8* %c
|
/external/llvm/test/CodeGen/AArch64/ |
arm64-prefetch.ll | 53 %tmp6 = load i32*, i32** @a, align 8, !tbaa !3 54 %arrayidx6 = getelementptr inbounds i32, i32* %tmp6, i64 %idxprom
|
/external/llvm/test/CodeGen/ARM/ |
2012-01-24-RegSequenceLiveRange.ll | 15 %tmp6 = insertelement <4 x float> %tmp5, float 0.000000e+00, i32 3 20 %tmp11 = bitcast <4 x float> %tmp6 to <2 x i64>
|
intrinsics-crypto.ll | 12 %tmp6 = call <16 x i8> @llvm.arm.neon.aesmc(<16 x i8> %tmp5) 14 ret <16 x i8> %tmp6
|