HomeSort by relevance Sort by last modified time
    Searched full:tmp4 (Results 201 - 225 of 461) sorted by null

1 2 3 4 5 6 7 891011>>

  /external/chromium_org/third_party/WebKit/Source/platform/graphics/cpu/arm/filters/
FEBlendNEON.h 62 uint16x8_t tmp4 = vmulq_u16(tmp3, pixelA); local
63 uint16x8_t tmp5 = vaddq_u16(tmp2, tmp4);
82 uint16x8_t tmp4 = vaddq_u16(tmp3, pixelA); local
89 return vminq_u16(tmp4, tmp8);
98 uint16x8_t tmp4 = vaddq_u16(tmp3, pixelA); local
105 return vmaxq_u16(tmp4, tmp8);
  /external/llvm/test/CodeGen/X86/
2007-10-31-extractelement-i64.ll 17 %tmp4 = load <1 x i64>* %retval, align 8 ; <<1 x i64>> [#uses=0]
36 %tmp4 = load <1 x i64>* %retval, align 8 ; <<1 x i64>> [#uses=0]
55 %tmp4 = load i64* %tmp, align 8 ; <i64> [#uses=1]
56 store i64 %tmp4, i64* %retval, align 8
75 %tmp4 = load i64* %tmp, align 8 ; <i64> [#uses=1]
76 store i64 %tmp4, i64* %retval, align 8
vararg_tailcall.ll 57 %tmp4 = load i8** @sel5, align 8
59 %call = tail call i8* (i8*, i8*, i8*, ...)* @x3(i8* %arg1, i8* %arg2, i8* %tmp2, i8* %tmp3, i8* %tmp4, i8* %tmp5) nounwind optsize noredzone
73 %tmp4 = load i8** @sel5, align 8
76 %call = tail call i8* (i8*, i8*, i8*, i8*, i8*, i8*, i8*, ...)* @x7(i8* %arg1, i8* %arg2, i8* %tmp2, i8* %tmp3, i8* %tmp4, i8* %tmp5, i8* %tmp6) nounwind optsize noredzone
90 %tmp4 = load i8** @sel5, align 8
92 %call = tail call i8* (i8*, i8*, i8*, ...)* @x3(i8* %arg1, i8* %arg2, i8* %tmp2, i8* %tmp3, i8* %tmp4, i8* %tmp5, i32 48879, i32 48879) nounwind optsize noredzone
2008-02-18-TailMergingBug.ll 10 %tmp4 = getelementptr float* %result, i32 2 ; <float*> [#uses=5]
11 %tmp5 = load float* %tmp4, align 4 ; <float> [#uses=10]
48 %tmp4.mux787 = select i1 %tmp82475, float* %tmp4, float* %tmp7 ; <float*> [#uses=1]
54 %tmp4.mux = select i1 %tmp82, float* %tmp4, float* %tmp7 ; <float*> [#uses=1]
61 %iftmp.0.0.in = phi float* [ %tmp10, %bb103 ], [ %result, %bb26 ], [ %result, %bb40 ], [ %result, %bb50 ], [ %tmp4.mux, %bb80 ], [ %tmp4.mux787, %bb72 ] ; <float*> [#uses=1]
95 %tmp4.mux791 = select i1 %tmp197483, float* %tmp4, float* %tmp7 ; <float*> [#uses=1
    [all...]
vec_ins_extract.ll 26 %tmp4 = load <4 x float>* %G ; <<4 x float>> [#uses=2]
27 %tmp6 = fadd <4 x float> %tmp4, %tmp4 ; <<4 x float>> [#uses=1]
vector.ll 120 %tmp4 = add %i4 %tmp3, < i32 1, i32 2, i32 3, i32 4 > ; <%i4> [#uses=1]
121 store %i4 %tmp4, %i4* %a
138 %tmp4 = insertelement %f4 %tmp2, float %X, i32 2 ; <%f4> [#uses=1]
139 %tmp6 = insertelement %f4 %tmp4, float %X, i32 3 ; <%f4> [#uses=1]
149 %tmp4 = insertelement %i4 %tmp2, i32 %X, i32 2 ; <%i4> [#uses=1]
150 %tmp6 = insertelement %i4 %tmp4, i32 %X, i32 3 ; <%i4> [#uses=1]
  /external/llvm/test/Transforms/IndVarSimplify/
ada-loops.ll 30 %tmp4 = getelementptr [256 x i32]* %a, i32 0, i32 %tmp3 ; <i32*> [#uses=1]
31 store i32 0, i32* %tmp4
65 %tmp4 = add i32 %tmp12, 10 ; <i32> [#uses=1]
66 %tmp5 = getelementptr [21 x i32]* %a, i32 0, i32 %tmp4 ; <i32*> [#uses=1]
83 %tmp4 = add i32 %tmp12, -10 ; <i32> [#uses=1]
84 %tmp5 = getelementptr [21 x i32]* %a, i32 0, i32 %tmp4 ; <i32*> [#uses=1]
  /external/llvm/test/Transforms/Inline/
devirtualize-3.ll 42 %tmp4 = getelementptr inbounds %struct.cont_t* %next, i32 0, i32 0 ; <void (i8*, i32)**> [#uses=1]
43 store void (i8*, i32)* bitcast (void (%struct.foo_sf_t*, i32)* @foo2 to void (i8*, i32)*), void (i8*, i32)** %tmp4
56 %tmp4 = load void (i8*, i32)** %tmp3 ; <void (i8*, i32)*> [#uses=1]
63 call void %tmp4(i8* %conv, i32 %mul)
71 %tmp4 = getelementptr inbounds %struct.cont_t* %c, i32 0, i32 1 ; <i8**> [#uses=1]
72 %tmp5 = load i8** %tmp4 ; <i8*> [#uses=1]
  /external/llvm/test/Transforms/ScalarRepl/
memset-aggregate.ll 18 %tmp4 = getelementptr %struct.foo* %L, i32 0, i32 0 ; <i32*> [#uses=1]
19 %tmp5 = load i32* %tmp4 ; <i32> [#uses=1]
29 %tmp4 = getelementptr [4 x %struct.foo]* %L, i32 0, i32 0, i32 0 ; <i32*> [#uses=1]
30 %tmp5 = load i32* %tmp4 ; <i32> [#uses=1]
42 %tmp4 = getelementptr %struct.bar* %B, i32 0, i32 2 ; <double*> [#uses=1]
43 store double 1.000000e+01, double* %tmp4
  /external/jpeg/
jfdctint.c 142 INT32 tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7; local
162 tmp4 = dataptr[3] - dataptr[4];
184 * i0..i3 in the paper are tmp4..tmp7 here.
187 z1 = tmp4 + tmp7;
189 z3 = tmp4 + tmp6;
193 tmp4 = MULTIPLY(tmp4, FIX_0_298631336); /* sqrt(2) * (-c1+c3+c5-c7) */
205 dataptr[7] = (DCTELEM) DESCALE(tmp4 + z1 + z3, CONST_BITS-PASS1_BITS);
227 tmp4 = dataptr[DCTSIZE*3] - dataptr[DCTSIZE*4];
249 * i0..i3 in the paper are tmp4..tmp7 here
    [all...]
jidctflt.c 72 FAST_FLOAT tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7; local
141 tmp4 = DEQUANTIZE(inptr[DCTSIZE*1], quantptr[DCTSIZE*1]);
148 z11 = tmp4 + tmp7;
149 z12 = tmp4 - tmp7;
160 tmp4 = tmp10 + tmp5;
168 wsptr[DCTSIZE*4] = tmp3 + tmp4;
169 wsptr[DCTSIZE*3] = tmp3 - tmp4;
217 tmp4 = tmp10 + tmp5;
233 outptr[4] = range_limit[(int) DESCALE((INT32) (tmp3 + tmp4), 3)
235 outptr[3] = range_limit[(int) DESCALE((INT32) (tmp3 - tmp4), 3
    [all...]
jfdctflt.c 61 FAST_FLOAT tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7; local
78 tmp4 = dataptr[3] - dataptr[4];
96 tmp10 = tmp4 + tmp5; /* phase 2 */
128 tmp4 = dataptr[DCTSIZE*3] - dataptr[DCTSIZE*4];
146 tmp10 = tmp4 + tmp5; /* phase 2 */
  /external/qemu/distrib/jpeg-6b/
jfdctint.c 142 INT32 tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7; local
162 tmp4 = dataptr[3] - dataptr[4];
184 * i0..i3 in the paper are tmp4..tmp7 here.
187 z1 = tmp4 + tmp7;
189 z3 = tmp4 + tmp6;
193 tmp4 = MULTIPLY(tmp4, FIX_0_298631336); /* sqrt(2) * (-c1+c3+c5-c7) */
205 dataptr[7] = (DCTELEM) DESCALE(tmp4 + z1 + z3, CONST_BITS-PASS1_BITS);
227 tmp4 = dataptr[DCTSIZE*3] - dataptr[DCTSIZE*4];
249 * i0..i3 in the paper are tmp4..tmp7 here
    [all...]
jidctflt.c 72 FAST_FLOAT tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7; local
141 tmp4 = DEQUANTIZE(inptr[DCTSIZE*1], quantptr[DCTSIZE*1]);
148 z11 = tmp4 + tmp7;
149 z12 = tmp4 - tmp7;
160 tmp4 = tmp10 + tmp5;
168 wsptr[DCTSIZE*4] = tmp3 + tmp4;
169 wsptr[DCTSIZE*3] = tmp3 - tmp4;
217 tmp4 = tmp10 + tmp5;
233 outptr[4] = range_limit[(int) DESCALE((INT32) (tmp3 + tmp4), 3)
235 outptr[3] = range_limit[(int) DESCALE((INT32) (tmp3 - tmp4), 3
    [all...]
jfdctflt.c 61 FAST_FLOAT tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7; local
78 tmp4 = dataptr[3] - dataptr[4];
96 tmp10 = tmp4 + tmp5; /* phase 2 */
128 tmp4 = dataptr[DCTSIZE*3] - dataptr[DCTSIZE*4];
146 tmp10 = tmp4 + tmp5; /* phase 2 */
  /frameworks/av/media/libstagefright/codecs/on2/h264dec/source/arm11_asm/
h264bsd_interpolate_ver_half.s 52 tmp4 RN 7 label
153 LDR tmp4, [ref], width ;// |m4|m3|m2|m1|
159 UXTAB16 tmpa, tmpa, tmp4 ;// |g3+m3|g1+m1|
174 UXTAB16 tmpa, tmpa, tmp4, ROR #8 ;// |g4+m4|g2+m2|
197 ;// tmp4 = |g4|g3|g2|g1|
203 UXTB16 tmpa, tmp4 ;// |g3|g1|
217 UXTB16 tmpa, tmp4, ROR #8 ;// |g4|g2|
240 ;// tmp4 = |c4|c3|c2|c1|
249 UXTB16 tmpb, tmp4 ;// |c3|c1|
266 UXTB16 tmpb, tmp4, ROR #8 ;// |c4|c2
    [all...]
h264bsd_interpolate_chroma_hor.s 61 tmp4 RN 10 label
181 LDRB tmp4, [ptrA], #1
184 PKHBT tmp6, tmp2, tmp4, LSL #16
193 PKHBT tmp8, tmp4, tmp2, LSL #16
252 LDRB tmp4, [ptrA], #1
255 PKHBT tmp6, tmp2, tmp4, LSL #16
264 PKHBT tmp8, tmp4, tmp2, LSL #16
h264bsd_interpolate_chroma_ver.s 59 tmp4 RN 10 label
182 LDRB tmp4, [ptrA],#1 ;// 1 row, 2 col
186 PKHBT tmp4, tmp4, tmp5, LSL #16 ;// |B|A|
192 SMLAD tmp5, tmp4, valY, c32 ;// multiply
249 LDRB tmp4, [ptrA],#1 ;// 1 row, 2 col
253 PKHBT tmp4, tmp4, tmp5, LSL #16 ;// |B|A|
259 SMLAD tmp5, tmp4, valY, c32 ;// multiply
  /frameworks/av/media/libstagefright/codecs/mp3dec/src/
pvmp3_dct_16.cpp 158 int32 tmp4; local
209 tmp4 = vec[ 4] + vec[11];
211 tmp1 = (tmp3 + tmp4);
212 tmp4 = fxp_mul32_Q32((tmp3 - tmp4) << 2, Qfmt_31(0.64072886193538F));
230 tmp1 = fxp_mul32_Q32((itmp_e0 - tmp4) << 1, Qfmt_31(0.54119610014620F));
231 tmp7 = itmp_e0 + tmp4;
236 tmp4 = fxp_mul32_Q32((tmp7 - tmp6) << 1, Qfmt_31(0.70710678118655F));
242 vec[ 6] = tmp1 + tmp4;
243 vec[10] = tmp7 + tmp4;
    [all...]
  /external/llvm/test/CodeGen/ARM/
rev.ll 8 %tmp4 = shl i32 %X15, 8
10 %tmp5 = and i32 %tmp4, -16777216
12 %tmp13 = and i32 %tmp4, 65280
26 %tmp4 = shl i16 %tmp3, 8
27 %tmp5 = or i16 %tmp2, %tmp4
vabd.ll 155 %tmp4 = zext <8 x i8> %tmp3 to <8 x i16>
156 ret <8 x i16> %tmp4
165 %tmp4 = zext <4 x i16> %tmp3 to <4 x i32>
166 ret <4 x i32> %tmp4
175 %tmp4 = zext <2 x i32> %tmp3 to <2 x i64>
176 ret <2 x i64> %tmp4
185 %tmp4 = zext <8 x i8> %tmp3 to <8 x i16>
186 ret <8 x i16> %tmp4
195 %tmp4 = zext <4 x i16> %tmp3 to <4 x i32>
196 ret <4 x i32> %tmp4
    [all...]
  /external/llvm/test/Analysis/ScalarEvolution/
sext-iv-2.ll 5 ; CHECK: %tmp4 = mul i32 %tmp3, %i.02
32 %tmp4 = mul i32 %tmp3, %i.02 ; <i32> [#uses=1]
36 store i32 %tmp4, i32* %tmp7, align 4
  /external/llvm/test/Transforms/InstCombine/
shift-sra.ll 16 %tmp4 = add i32 %tmp3, 7 ; <i32> [#uses=1]
17 %tmp5 = ashr i32 %tmp4, 3 ; <i32> [#uses=1]
20 ; CHECK: lshr i32 %tmp4, 3
  /external/llvm/test/CodeGen/Generic/
vector.ll 120 %tmp4 = add %i4 %tmp3, < i32 1, i32 2, i32 3, i32 4 > ; <%i4> [#uses=1]
121 store %i4 %tmp4, %i4* %a
137 %tmp4 = insertelement %f4 %tmp2, float %X, i32 2 ; <%f4> [#uses=1]
138 %tmp6 = insertelement %f4 %tmp4, float %X, i32 3 ; <%f4> [#uses=1]
148 %tmp4 = insertelement %i4 %tmp2, i32 %X, i32 2 ; <%i4> [#uses=1]
149 %tmp6 = insertelement %i4 %tmp4, i32 %X, i32 3 ; <%i4> [#uses=1]
  /external/llvm/test/CodeGen/PowerPC/
vector.ll 130 %tmp4 = add %i4 %tmp3, < i32 1, i32 2, i32 3, i32 4 >
131 store %i4 %tmp4, %i4* %a
148 %tmp4 = insertelement %f4 %tmp2, float %X, i32 2
149 %tmp6 = insertelement %f4 %tmp4, float %X, i32 3
159 %tmp4 = insertelement %i4 %tmp2, i32 %X, i32 2
160 %tmp6 = insertelement %i4 %tmp4, i32 %X, i32 3

Completed in 516 milliseconds

1 2 3 4 5 6 7 891011>>