/external/chromium_org/third_party/WebKit/Source/platform/graphics/cpu/arm/filters/ |
FEBlendNEON.h | 62 uint16x8_t tmp4 = vmulq_u16(tmp3, pixelA); local 63 uint16x8_t tmp5 = vaddq_u16(tmp2, tmp4); 82 uint16x8_t tmp4 = vaddq_u16(tmp3, pixelA); local 89 return vminq_u16(tmp4, tmp8); 98 uint16x8_t tmp4 = vaddq_u16(tmp3, pixelA); local 105 return vmaxq_u16(tmp4, tmp8);
|
/external/llvm/test/CodeGen/X86/ |
2007-10-31-extractelement-i64.ll | 17 %tmp4 = load <1 x i64>* %retval, align 8 ; <<1 x i64>> [#uses=0] 36 %tmp4 = load <1 x i64>* %retval, align 8 ; <<1 x i64>> [#uses=0] 55 %tmp4 = load i64* %tmp, align 8 ; <i64> [#uses=1] 56 store i64 %tmp4, i64* %retval, align 8 75 %tmp4 = load i64* %tmp, align 8 ; <i64> [#uses=1] 76 store i64 %tmp4, i64* %retval, align 8
|
vararg_tailcall.ll | 57 %tmp4 = load i8** @sel5, align 8 59 %call = tail call i8* (i8*, i8*, i8*, ...)* @x3(i8* %arg1, i8* %arg2, i8* %tmp2, i8* %tmp3, i8* %tmp4, i8* %tmp5) nounwind optsize noredzone 73 %tmp4 = load i8** @sel5, align 8 76 %call = tail call i8* (i8*, i8*, i8*, i8*, i8*, i8*, i8*, ...)* @x7(i8* %arg1, i8* %arg2, i8* %tmp2, i8* %tmp3, i8* %tmp4, i8* %tmp5, i8* %tmp6) nounwind optsize noredzone 90 %tmp4 = load i8** @sel5, align 8 92 %call = tail call i8* (i8*, i8*, i8*, ...)* @x3(i8* %arg1, i8* %arg2, i8* %tmp2, i8* %tmp3, i8* %tmp4, i8* %tmp5, i32 48879, i32 48879) nounwind optsize noredzone
|
2008-02-18-TailMergingBug.ll | 10 %tmp4 = getelementptr float* %result, i32 2 ; <float*> [#uses=5] 11 %tmp5 = load float* %tmp4, align 4 ; <float> [#uses=10] 48 %tmp4.mux787 = select i1 %tmp82475, float* %tmp4, float* %tmp7 ; <float*> [#uses=1] 54 %tmp4.mux = select i1 %tmp82, float* %tmp4, float* %tmp7 ; <float*> [#uses=1] 61 %iftmp.0.0.in = phi float* [ %tmp10, %bb103 ], [ %result, %bb26 ], [ %result, %bb40 ], [ %result, %bb50 ], [ %tmp4.mux, %bb80 ], [ %tmp4.mux787, %bb72 ] ; <float*> [#uses=1] 95 %tmp4.mux791 = select i1 %tmp197483, float* %tmp4, float* %tmp7 ; <float*> [#uses=1 [all...] |
vec_ins_extract.ll | 26 %tmp4 = load <4 x float>* %G ; <<4 x float>> [#uses=2] 27 %tmp6 = fadd <4 x float> %tmp4, %tmp4 ; <<4 x float>> [#uses=1]
|
vector.ll | 120 %tmp4 = add %i4 %tmp3, < i32 1, i32 2, i32 3, i32 4 > ; <%i4> [#uses=1] 121 store %i4 %tmp4, %i4* %a 138 %tmp4 = insertelement %f4 %tmp2, float %X, i32 2 ; <%f4> [#uses=1] 139 %tmp6 = insertelement %f4 %tmp4, float %X, i32 3 ; <%f4> [#uses=1] 149 %tmp4 = insertelement %i4 %tmp2, i32 %X, i32 2 ; <%i4> [#uses=1] 150 %tmp6 = insertelement %i4 %tmp4, i32 %X, i32 3 ; <%i4> [#uses=1]
|
/external/llvm/test/Transforms/IndVarSimplify/ |
ada-loops.ll | 30 %tmp4 = getelementptr [256 x i32]* %a, i32 0, i32 %tmp3 ; <i32*> [#uses=1] 31 store i32 0, i32* %tmp4 65 %tmp4 = add i32 %tmp12, 10 ; <i32> [#uses=1] 66 %tmp5 = getelementptr [21 x i32]* %a, i32 0, i32 %tmp4 ; <i32*> [#uses=1] 83 %tmp4 = add i32 %tmp12, -10 ; <i32> [#uses=1] 84 %tmp5 = getelementptr [21 x i32]* %a, i32 0, i32 %tmp4 ; <i32*> [#uses=1]
|
/external/llvm/test/Transforms/Inline/ |
devirtualize-3.ll | 42 %tmp4 = getelementptr inbounds %struct.cont_t* %next, i32 0, i32 0 ; <void (i8*, i32)**> [#uses=1] 43 store void (i8*, i32)* bitcast (void (%struct.foo_sf_t*, i32)* @foo2 to void (i8*, i32)*), void (i8*, i32)** %tmp4 56 %tmp4 = load void (i8*, i32)** %tmp3 ; <void (i8*, i32)*> [#uses=1] 63 call void %tmp4(i8* %conv, i32 %mul) 71 %tmp4 = getelementptr inbounds %struct.cont_t* %c, i32 0, i32 1 ; <i8**> [#uses=1] 72 %tmp5 = load i8** %tmp4 ; <i8*> [#uses=1]
|
/external/llvm/test/Transforms/ScalarRepl/ |
memset-aggregate.ll | 18 %tmp4 = getelementptr %struct.foo* %L, i32 0, i32 0 ; <i32*> [#uses=1] 19 %tmp5 = load i32* %tmp4 ; <i32> [#uses=1] 29 %tmp4 = getelementptr [4 x %struct.foo]* %L, i32 0, i32 0, i32 0 ; <i32*> [#uses=1] 30 %tmp5 = load i32* %tmp4 ; <i32> [#uses=1] 42 %tmp4 = getelementptr %struct.bar* %B, i32 0, i32 2 ; <double*> [#uses=1] 43 store double 1.000000e+01, double* %tmp4
|
/external/jpeg/ |
jfdctint.c | 142 INT32 tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7; local 162 tmp4 = dataptr[3] - dataptr[4]; 184 * i0..i3 in the paper are tmp4..tmp7 here. 187 z1 = tmp4 + tmp7; 189 z3 = tmp4 + tmp6; 193 tmp4 = MULTIPLY(tmp4, FIX_0_298631336); /* sqrt(2) * (-c1+c3+c5-c7) */ 205 dataptr[7] = (DCTELEM) DESCALE(tmp4 + z1 + z3, CONST_BITS-PASS1_BITS); 227 tmp4 = dataptr[DCTSIZE*3] - dataptr[DCTSIZE*4]; 249 * i0..i3 in the paper are tmp4..tmp7 here [all...] |
jidctflt.c | 72 FAST_FLOAT tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7; local 141 tmp4 = DEQUANTIZE(inptr[DCTSIZE*1], quantptr[DCTSIZE*1]); 148 z11 = tmp4 + tmp7; 149 z12 = tmp4 - tmp7; 160 tmp4 = tmp10 + tmp5; 168 wsptr[DCTSIZE*4] = tmp3 + tmp4; 169 wsptr[DCTSIZE*3] = tmp3 - tmp4; 217 tmp4 = tmp10 + tmp5; 233 outptr[4] = range_limit[(int) DESCALE((INT32) (tmp3 + tmp4), 3) 235 outptr[3] = range_limit[(int) DESCALE((INT32) (tmp3 - tmp4), 3 [all...] |
jfdctflt.c | 61 FAST_FLOAT tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7; local 78 tmp4 = dataptr[3] - dataptr[4]; 96 tmp10 = tmp4 + tmp5; /* phase 2 */ 128 tmp4 = dataptr[DCTSIZE*3] - dataptr[DCTSIZE*4]; 146 tmp10 = tmp4 + tmp5; /* phase 2 */
|
/external/qemu/distrib/jpeg-6b/ |
jfdctint.c | 142 INT32 tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7; local 162 tmp4 = dataptr[3] - dataptr[4]; 184 * i0..i3 in the paper are tmp4..tmp7 here. 187 z1 = tmp4 + tmp7; 189 z3 = tmp4 + tmp6; 193 tmp4 = MULTIPLY(tmp4, FIX_0_298631336); /* sqrt(2) * (-c1+c3+c5-c7) */ 205 dataptr[7] = (DCTELEM) DESCALE(tmp4 + z1 + z3, CONST_BITS-PASS1_BITS); 227 tmp4 = dataptr[DCTSIZE*3] - dataptr[DCTSIZE*4]; 249 * i0..i3 in the paper are tmp4..tmp7 here [all...] |
jidctflt.c | 72 FAST_FLOAT tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7; local 141 tmp4 = DEQUANTIZE(inptr[DCTSIZE*1], quantptr[DCTSIZE*1]); 148 z11 = tmp4 + tmp7; 149 z12 = tmp4 - tmp7; 160 tmp4 = tmp10 + tmp5; 168 wsptr[DCTSIZE*4] = tmp3 + tmp4; 169 wsptr[DCTSIZE*3] = tmp3 - tmp4; 217 tmp4 = tmp10 + tmp5; 233 outptr[4] = range_limit[(int) DESCALE((INT32) (tmp3 + tmp4), 3) 235 outptr[3] = range_limit[(int) DESCALE((INT32) (tmp3 - tmp4), 3 [all...] |
jfdctflt.c | 61 FAST_FLOAT tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7; local 78 tmp4 = dataptr[3] - dataptr[4]; 96 tmp10 = tmp4 + tmp5; /* phase 2 */ 128 tmp4 = dataptr[DCTSIZE*3] - dataptr[DCTSIZE*4]; 146 tmp10 = tmp4 + tmp5; /* phase 2 */
|
/frameworks/av/media/libstagefright/codecs/on2/h264dec/source/arm11_asm/ |
h264bsd_interpolate_ver_half.s | 52 tmp4 RN 7 label 153 LDR tmp4, [ref], width ;// |m4|m3|m2|m1| 159 UXTAB16 tmpa, tmpa, tmp4 ;// |g3+m3|g1+m1| 174 UXTAB16 tmpa, tmpa, tmp4, ROR #8 ;// |g4+m4|g2+m2| 197 ;// tmp4 = |g4|g3|g2|g1| 203 UXTB16 tmpa, tmp4 ;// |g3|g1| 217 UXTB16 tmpa, tmp4, ROR #8 ;// |g4|g2| 240 ;// tmp4 = |c4|c3|c2|c1| 249 UXTB16 tmpb, tmp4 ;// |c3|c1| 266 UXTB16 tmpb, tmp4, ROR #8 ;// |c4|c2 [all...] |
h264bsd_interpolate_chroma_hor.s | 61 tmp4 RN 10 label 181 LDRB tmp4, [ptrA], #1 184 PKHBT tmp6, tmp2, tmp4, LSL #16 193 PKHBT tmp8, tmp4, tmp2, LSL #16 252 LDRB tmp4, [ptrA], #1 255 PKHBT tmp6, tmp2, tmp4, LSL #16 264 PKHBT tmp8, tmp4, tmp2, LSL #16
|
h264bsd_interpolate_chroma_ver.s | 59 tmp4 RN 10 label 182 LDRB tmp4, [ptrA],#1 ;// 1 row, 2 col 186 PKHBT tmp4, tmp4, tmp5, LSL #16 ;// |B|A| 192 SMLAD tmp5, tmp4, valY, c32 ;// multiply 249 LDRB tmp4, [ptrA],#1 ;// 1 row, 2 col 253 PKHBT tmp4, tmp4, tmp5, LSL #16 ;// |B|A| 259 SMLAD tmp5, tmp4, valY, c32 ;// multiply
|
/frameworks/av/media/libstagefright/codecs/mp3dec/src/ |
pvmp3_dct_16.cpp | 158 int32 tmp4; local 209 tmp4 = vec[ 4] + vec[11]; 211 tmp1 = (tmp3 + tmp4); 212 tmp4 = fxp_mul32_Q32((tmp3 - tmp4) << 2, Qfmt_31(0.64072886193538F)); 230 tmp1 = fxp_mul32_Q32((itmp_e0 - tmp4) << 1, Qfmt_31(0.54119610014620F)); 231 tmp7 = itmp_e0 + tmp4; 236 tmp4 = fxp_mul32_Q32((tmp7 - tmp6) << 1, Qfmt_31(0.70710678118655F)); 242 vec[ 6] = tmp1 + tmp4; 243 vec[10] = tmp7 + tmp4; [all...] |
/external/llvm/test/CodeGen/ARM/ |
rev.ll | 8 %tmp4 = shl i32 %X15, 8 10 %tmp5 = and i32 %tmp4, -16777216 12 %tmp13 = and i32 %tmp4, 65280 26 %tmp4 = shl i16 %tmp3, 8 27 %tmp5 = or i16 %tmp2, %tmp4
|
vabd.ll | 155 %tmp4 = zext <8 x i8> %tmp3 to <8 x i16> 156 ret <8 x i16> %tmp4 165 %tmp4 = zext <4 x i16> %tmp3 to <4 x i32> 166 ret <4 x i32> %tmp4 175 %tmp4 = zext <2 x i32> %tmp3 to <2 x i64> 176 ret <2 x i64> %tmp4 185 %tmp4 = zext <8 x i8> %tmp3 to <8 x i16> 186 ret <8 x i16> %tmp4 195 %tmp4 = zext <4 x i16> %tmp3 to <4 x i32> 196 ret <4 x i32> %tmp4 [all...] |
/external/llvm/test/Analysis/ScalarEvolution/ |
sext-iv-2.ll | 5 ; CHECK: %tmp4 = mul i32 %tmp3, %i.02 32 %tmp4 = mul i32 %tmp3, %i.02 ; <i32> [#uses=1] 36 store i32 %tmp4, i32* %tmp7, align 4
|
/external/llvm/test/Transforms/InstCombine/ |
shift-sra.ll | 16 %tmp4 = add i32 %tmp3, 7 ; <i32> [#uses=1] 17 %tmp5 = ashr i32 %tmp4, 3 ; <i32> [#uses=1] 20 ; CHECK: lshr i32 %tmp4, 3
|
/external/llvm/test/CodeGen/Generic/ |
vector.ll | 120 %tmp4 = add %i4 %tmp3, < i32 1, i32 2, i32 3, i32 4 > ; <%i4> [#uses=1] 121 store %i4 %tmp4, %i4* %a 137 %tmp4 = insertelement %f4 %tmp2, float %X, i32 2 ; <%f4> [#uses=1] 138 %tmp6 = insertelement %f4 %tmp4, float %X, i32 3 ; <%f4> [#uses=1] 148 %tmp4 = insertelement %i4 %tmp2, i32 %X, i32 2 ; <%i4> [#uses=1] 149 %tmp6 = insertelement %i4 %tmp4, i32 %X, i32 3 ; <%i4> [#uses=1]
|
/external/llvm/test/CodeGen/PowerPC/ |
vector.ll | 130 %tmp4 = add %i4 %tmp3, < i32 1, i32 2, i32 3, i32 4 > 131 store %i4 %tmp4, %i4* %a 148 %tmp4 = insertelement %f4 %tmp2, float %X, i32 2 149 %tmp6 = insertelement %f4 %tmp4, float %X, i32 3 159 %tmp4 = insertelement %i4 %tmp2, i32 %X, i32 2 160 %tmp6 = insertelement %i4 %tmp4, i32 %X, i32 3
|