/external/libvpx/vp8/encoder/x86/ |
variance_mmx.c | 137 unsigned int sse0, sse1, sse2, sse3, var; local 141 vp8_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0) ; 146 var = sse0 + sse1 + sse2 + sse3; 199 unsigned int sse0, sse1, sse2, sse3, var; local 203 vp8_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0) ; 208 var = sse0 + sse1 + sse2 + sse3; 221 unsigned int sse0, sse1, sse2, sse3, var; local 225 vp8_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0) ; 230 var = sse0 + sse1 + sse2 + sse3; 243 unsigned int sse0, sse1, var local 264 unsigned int sse0, sse1, var; local 466 unsigned int sse0, sse1, sse2, sse3, var; local 489 unsigned int sse0, sse1, var; local [all...] |
variance_sse2.c | 158 unsigned int sse0; local 162 vp8_get16x16var_sse2(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0) ; 163 *sse = sse0; 164 return (sse0 - ((sum0 * sum0) >> 8)); 174 unsigned int sse0; local 176 vp8_get16x16var_sse2(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0) ; 177 *sse = sse0; 178 return sse0; 191 unsigned int sse0, sse1, var; local 194 vp8_get8x8var_sse2(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0) 212 unsigned int sse0, sse1, var; local 449 unsigned int sse0, sse1, sse2, sse3, var; local 473 unsigned int sse0, sse1, var; local [all...] |
/external/libffi/src/x86/ |
ffi64.c | 299 _Bool sse0 = SSE_CLASS_P (classes[0]); local 301 if (sse0 && !sse1) 303 else if (!sse0 && sse1) 305 else if (sse0 && sse1) 497 _Bool sse0 = SSE_CLASS_P (classes[0]); local 499 if (!sse0 && sse1) 501 else if (sse0 && !sse1)
|
/external/libvpx/vp8/encoder/ |
rdopt.c | 833 unsigned int sse0, sse1; local 835 VARIANCE_INVOKE(rtcd, get8x8var)(x->src.u_buffer, x->src.uv_stride, x->e_mbd.dst.u_buffer, x->e_mbd.dst.uv_stride, &sse0, &sum0); 837 return (sse0 + sse1); [all...] |