/external/libvpx/libvpx/vp8/common/ |
sad_c.c | 17 static unsigned int sad_mx_n_c(const unsigned char *src_ptr, int src_stride, 28 sad += abs(src_ptr[c] - ref_ptr[c]); 34 src_ptr += src_stride; 45 unsigned int vp8_sad16x16_c(const unsigned char *src_ptr, int src_stride, 49 return sad_mx_n_c(src_ptr, src_stride, ref_ptr, ref_stride, max_sad, 16, 16); 52 unsigned int vp8_sad8x8_c(const unsigned char *src_ptr, int src_stride, 56 return sad_mx_n_c(src_ptr, src_stride, ref_ptr, ref_stride, max_sad, 8, 8); 59 unsigned int vp8_sad16x8_c(const unsigned char *src_ptr, int src_stride, 63 return sad_mx_n_c(src_ptr, src_stride, ref_ptr, ref_stride, max_sad, 16, 8); 67 unsigned int vp8_sad8x16_c(const unsigned char *src_ptr, int src_stride [all...] |
reconintra4x4.h | 22 unsigned int *src_ptr = (unsigned int *)above_right_src; local 27 *dst_ptr0 = *src_ptr; 28 *dst_ptr1 = *src_ptr; 29 *dst_ptr2 = *src_ptr;
|
filter.c | 41 unsigned char *src_ptr, 57 Temp = ((int)src_ptr[-2 * (int)pixel_step] * vp8_filter[0]) + 58 ((int)src_ptr[-1 * (int)pixel_step] * vp8_filter[1]) + 59 ((int)src_ptr[0] * vp8_filter[2]) + 60 ((int)src_ptr[pixel_step] * vp8_filter[3]) + 61 ((int)src_ptr[2*pixel_step] * vp8_filter[4]) + 62 ((int)src_ptr[3*pixel_step] * vp8_filter[5]) + 74 src_ptr++; 78 src_ptr += src_pixels_per_line - output_width; 85 int *src_ptr, [all...] |
variance_c.c | 18 const short *src_ptr 25 sum += (src_ptr[i] * src_ptr[i]); 35 const unsigned char *src_ptr, 54 diff = src_ptr[j] - ref_ptr[j]; 59 src_ptr += source_stride; 66 const unsigned char *src_ptr, 76 variance(src_ptr, source_stride, ref_ptr, recon_stride, 16, 16, &var, &avg); 82 const unsigned char *src_ptr, 92 variance(src_ptr, source_stride, ref_ptr, recon_stride, 8, 16, &var, &avg) [all...] |
variance.h | 18 const unsigned char *src_ptr, 25 const unsigned char *src_ptr, 32 const unsigned char *src_ptr, 40 const unsigned char *src_ptr, 49 const unsigned char *src_ptr, 58 const unsigned char *src_ptr, 67 const unsigned char *src_ptr, 93 const unsigned char *src_ptr,
|
/external/libyuv/files/source/ |
scale_neon.cc | 29 void ScaleRowDown2_NEON(const uint8* src_ptr, ptrdiff_t /* src_stride */, 38 : "+r"(src_ptr), // %0 46 void ScaleRowDown2Int_NEON(const uint8* src_ptr, ptrdiff_t src_stride, 63 : "+r"(src_ptr), // %0 72 void ScaleRowDown4_NEON(const uint8* src_ptr, ptrdiff_t /* src_stride */, 82 : "+r"(src_ptr), // %0 90 void ScaleRowDown4Int_NEON(const uint8* src_ptr, ptrdiff_t src_stride, 111 : "+r"(src_ptr), // %0 122 void ScaleRowDown34_NEON(const uint8* src_ptr, 132 : "+r"(src_ptr), // % [all...] |
scale.cc | 56 void ScaleRowDown2_NEON(const uint8* src_ptr, ptrdiff_t /* src_stride */, 59 void ScaleRowDown2Int_NEON(const uint8* src_ptr, ptrdiff_t src_stride, 63 void ScaleRowDown4_NEON(const uint8* src_ptr, ptrdiff_t /* src_stride */, 65 void ScaleRowDown4Int_NEON(const uint8* src_ptr, ptrdiff_t src_stride, 72 void ScaleRowDown34_NEON(const uint8* src_ptr, 75 void ScaleRowDown34_0_Int_NEON(const uint8* src_ptr, 78 void ScaleRowDown34_1_Int_NEON(const uint8* src_ptr, 84 void ScaleRowDown38_NEON(const uint8* src_ptr, 88 void ScaleRowDown38_3_Int_NEON(const uint8* src_ptr, 92 void ScaleRowDown38_2_Int_NEON(const uint8* src_ptr, [all...] |
scale_argb.cc | 38 // Alignment requirement: src_ptr 16 byte aligned, dst_ptr 16 byte aligned. 40 static void ScaleARGBRowDown2_SSE2(const uint8* src_ptr, 44 mov eax, [esp + 4] // src_ptr 65 // Alignment requirement: src_ptr 16 byte aligned, dst_ptr 16 byte aligned. 67 static void ScaleARGBRowDown2Int_SSE2(const uint8* src_ptr, 72 mov eax, [esp + 4 + 4] // src_ptr 104 void ScaleARGBRowDownEven_SSE2(const uint8* src_ptr, ptrdiff_t src_stride, 110 mov eax, [esp + 8 + 4] // src_ptr 142 static void ScaleARGBRowDownEvenInt_SSE2(const uint8* src_ptr, 150 mov eax, [esp + 12 + 4] // src_ptr [all...] |
/external/libvpx/libvpx/vp8/common/x86/ |
vp8_asm_stubs.c | 21 unsigned char *src_ptr, 31 unsigned short *src_ptr, 42 unsigned char *src_ptr, 52 unsigned char *src_ptr, 62 unsigned short *src_ptr, 73 unsigned short *src_ptr, 84 unsigned char *src_ptr, 92 unsigned char *src_ptr, 101 unsigned char *src_ptr, 110 unsigned char *src_ptr, [all...] |
variance_sse2.c | 17 extern void filter_block1d_h6_mmx(const unsigned char *src_ptr, unsigned short *output_ptr, unsigned int src_pixels_per_line, unsigned int pixel_step, unsigned int output_height, unsigned int output_width, short *filter); 18 extern void filter_block1d_v6_mmx(const short *src_ptr, unsigned char *output_ptr, unsigned int pixels_per_line, unsigned int pixel_step, unsigned int output_height, unsigned int output_width, short *filter); 19 extern void filter_block1d8_h6_sse2(const unsigned char *src_ptr, unsigned short *output_ptr, unsigned int src_pixels_per_line, unsigned int pixel_step, unsigned int output_height, unsigned int output_width, short *filter); 20 extern void filter_block1d8_v6_sse2(const short *src_ptr, unsigned char *output_ptr, unsigned int pixels_per_line, unsigned int pixel_step, unsigned int output_height, unsigned int output_width, short *filter); 26 const unsigned char *src_ptr, 36 const unsigned char *src_ptr, 46 const short *src_ptr 50 const unsigned char *src_ptr, 59 const unsigned char *src_ptr, 70 const unsigned char *src_ptr, [all...] |
variance_mmx.c | 19 const unsigned char *src_ptr, 29 const short *src_ptr, 38 extern unsigned int vp8_get_mb_ss_mmx(const short *src_ptr); 41 const unsigned char *src_ptr, 50 const unsigned char *src_ptr, 61 const unsigned char *src_ptr, 72 const unsigned char *src_ptr, 83 const unsigned char *src_ptr, 92 vp8_get4x4var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &var, &avg) ; 99 const unsigned char *src_ptr, [all...] |
variance_ssse3.c | 18 const unsigned char *src_ptr, 29 const unsigned char *src_ptr, 39 const unsigned char *src_ptr, 49 const unsigned char *src_ptr, 59 const unsigned char *src_ptr, 70 const unsigned char *src_ptr, 88 src_ptr, src_pixels_per_line, 95 src_ptr, src_pixels_per_line, 102 src_ptr, src_pixels_per_line, 109 src_ptr, src_pixels_per_line [all...] |
/external/libvpx/libvpx/vp8/common/arm/ |
bilinearfilter_arm.h | 17 const unsigned char *src_ptr, 27 const unsigned short *src_ptr,
|
filter_arm.c | 20 unsigned char *src_ptr, 31 unsigned char *src_ptr, 42 unsigned char *src_ptr, 52 short *src_ptr, 61 short *src_ptr, 70 unsigned char *src_ptr, 81 unsigned char *src_ptr, 92 unsigned char *src_ptr, 111 /*vp8_filter_block2d_first_pass_armv6 ( src_ptr, FData+2, src_pixels_per_line, 4, 4, HFilter ); 114 vp8_filter_block2d_first_pass_only_armv6(src_ptr, dst_ptr, src_pixels_per_line, 4, dst_pitch, HFilter) [all...] |
bilinearfilter_arm.c | 19 unsigned char *src_ptr, 32 vp8_filter_block2d_bil_first_pass_armv6(src_ptr, FData, src_pitch, Height + 1, Width, HFilter); 41 unsigned char *src_ptr, 55 vp8_filter_block2d_bil_armv6(src_ptr, dst_ptr, src_pixels_per_line, dst_pitch, HFilter, VFilter, 4, 4); 60 unsigned char *src_ptr, 74 vp8_filter_block2d_bil_armv6(src_ptr, dst_ptr, src_pixels_per_line, dst_pitch, HFilter, VFilter, 8, 8); 79 unsigned char *src_ptr, 93 vp8_filter_block2d_bil_armv6(src_ptr, dst_ptr, src_pixels_per_line, dst_pitch, HFilter, VFilter, 8, 4); 98 unsigned char *src_ptr, 112 vp8_filter_block2d_bil_armv6(src_ptr, dst_ptr, src_pixels_per_line, dst_pitch, HFilter, VFilter, 16, 16) [all...] |
variance_arm.c | 21 const unsigned char *src_ptr, 37 vp8_filter_block2d_bil_first_pass_armv6(src_ptr, first_pass, 49 const unsigned char *src_ptr, 65 var = vp8_variance_halfpixvar16x16_h_armv6(src_ptr, src_pixels_per_line, 70 var = vp8_variance_halfpixvar16x16_v_armv6(src_ptr, src_pixels_per_line, 75 var = vp8_variance_halfpixvar16x16_hv_armv6(src_ptr, src_pixels_per_line, 83 vp8_filter_block2d_bil_first_pass_armv6(src_ptr, first_pass, 102 const unsigned char *src_ptr, 113 const unsigned char *src_ptr, 123 return vp8_variance_halfpixvar16x16_h_neon(src_ptr, src_pixels_per_line, dst_ptr, dst_pixels_per_line, sse) [all...] |
/external/libvpx/libvpx/vp9/encoder/ |
vp9_sad_c.c | 20 unsigned int vp9_sad##m##x##n##_c(const uint8_t *src_ptr, \ 25 return sad_mx_n_c(src_ptr, src_stride, ref_ptr, ref_stride, m, n); \ 27 unsigned int vp9_sad##m##x##n##_avg_c(const uint8_t *src_ptr, \ 35 return sad_mx_n_c(src_ptr, src_stride, comp_pred, m, m, n); \ 52 void vp9_sad64x32x4d_c(const uint8_t *src_ptr, 57 sad_array[0] = vp9_sad64x32(src_ptr, src_stride, 59 sad_array[1] = vp9_sad64x32(src_ptr, src_stride, 61 sad_array[2] = vp9_sad64x32(src_ptr, src_stride, 63 sad_array[3] = vp9_sad64x32(src_ptr, src_stride, 67 void vp9_sad32x64x4d_c(const uint8_t *src_ptr, [all...] |
vp9_variance_c.c | 21 void variance(const uint8_t *src_ptr, 37 diff = src_ptr[j] - ref_ptr[j]; 42 src_ptr += source_stride; 51 * INPUTS : uint8_t *src_ptr : Pointer to source block. 78 static void var_filter_block2d_bil_first_pass(const uint8_t *src_ptr, 89 output_ptr[j] = ROUND_POWER_OF_TWO((int)src_ptr[0] * vp9_filter[0] + 90 (int)src_ptr[pixel_step] * vp9_filter[1], 93 src_ptr++; 97 src_ptr += src_pixels_per_line - output_width; 106 * INPUTS : int32_t *src_ptr : Pointer to source block [all...] |
vp9_variance.h | 17 void variance(const uint8_t *src_ptr, 26 typedef unsigned int(*vp9_sad_fn_t)(const uint8_t *src_ptr, 32 typedef unsigned int(*vp9_sad_avg_fn_t)(const uint8_t *src_ptr, 39 typedef void (*vp9_sad_multi_fn_t)(const uint8_t *src_ptr, 45 typedef void (*vp9_sad_multi1_fn_t)(const uint8_t *src_ptr, 51 typedef void (*vp9_sad_multi_d_fn_t)(const uint8_t *src_ptr, 56 typedef unsigned int (*vp9_variance_fn_t)(const uint8_t *src_ptr, 62 typedef unsigned int (*vp9_subpixvariance_fn_t)(const uint8_t *src_ptr, 70 typedef unsigned int (*vp9_subp_avg_variance_fn_t)(const uint8_t *src_ptr, 81 typedef unsigned int (*vp9_get16x16prederror_fn_t)(const uint8_t *src_ptr, [all...] |
/external/libvpx/libvpx/vp9/common/ |
vp9_sadmxn.h | 17 static INLINE unsigned int sad_mx_n_c(const uint8_t *src_ptr, 28 sad += abs(src_ptr[c] - ref_ptr[c]); 31 src_ptr += src_stride;
|
/external/libvpx/libvpx/vp9/encoder/x86/ |
vp9_variance_mmx.c | 16 extern unsigned int vp9_get_mb_ss_mmx(const int16_t *src_ptr); 19 const unsigned char *src_ptr, 28 const unsigned char *src_ptr, 37 const unsigned char *src_ptr, 45 vp9_get4x4var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &var, &avg); 51 const unsigned char *src_ptr, 59 vp9_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &var, &avg); 66 const unsigned char *src_ptr, 75 vp9_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, 77 vp9_get8x8var_mmx(src_ptr + 8, source_stride, ref_ptr + 8, recon_stride [all...] |
vp9_variance_sse2.c | 19 const unsigned char *src_ptr, 29 const int16_t *src_ptr 33 const unsigned char *src_ptr, 42 const unsigned char *src_ptr, 53 const unsigned char *src_ptr, 63 const unsigned char *src_ptr, 73 const unsigned char *src_ptr, 83 const unsigned char *src_ptr, 93 const unsigned char *src_ptr, 103 const unsigned char *src_ptr, [all...] |
/external/libvpx/libvpx/vp8/common/mips/dspr2/ |
filter_dspr2.c | 118 unsigned char *RESTRICT src_ptr, 143 /* prefetch src_ptr data to cache memory */ 144 prefetch_load(src_ptr + src_pixels_per_line); 145 dst_ptr[0] = src_ptr[0]; 146 dst_ptr[1] = src_ptr[1]; 147 dst_ptr[2] = src_ptr[2]; 148 dst_ptr[3] = src_ptr[3]; 151 src_ptr += src_pixels_per_line; 164 /* prefetch src_ptr data to cache memory */ 165 prefetch_load(src_ptr + src_pixels_per_line) [all...] |
/external/libvpx/libvpx/third_party/libyuv/source/ |
scale.c | 63 void ScaleRowDown2_NEON(const uint8* src_ptr, int src_stride, 71 : "+r"(src_ptr), // %0 79 void ScaleRowDown2Int_NEON(const uint8* src_ptr, int src_stride, 95 : "+r"(src_ptr), // %0 105 static void ScaleRowDown4_NEON(const uint8* src_ptr, int src_stride, 116 : "+r"(src_ptr), // %0 124 static void ScaleRowDown4Int_NEON(const uint8* src_ptr, int src_stride, 151 : "+r"(src_ptr), // %0 163 static void ScaleRowDown34_NEON(const uint8* src_ptr, int src_stride, 172 : "+r"(src_ptr), // % [all...] |
/external/libvpx/libvpx/vp8/common/arm/neon/ |
sixtappredict4x4_neon.asm | 29 ; r0 unsigned char *src_ptr, 75 vext.8 d18, d6, d7, #5 ;construct src_ptr[3] 80 vswp d7, d8 ;discard 2nd half data after src_ptr[3] is done 83 vzip.32 d18, d19 ;put 2-line data in 1 register (src_ptr[3]) 85 vmull.u8 q7, d18, d5 ;(src_ptr[3] * vp8_filter[5]) 91 vzip.32 d6, d7 ;construct src_ptr[-2], and put 2-line data together 93 vshr.u64 q9, q4, #8 ;construct src_ptr[-1] 95 vmlal.u8 q7, d6, d0 ;+(src_ptr[-2] * vp8_filter[0]) 98 vzip.32 d18, d19 ;put 2-line data in 1 register (src_ptr[-1]) 100 vshr.u64 q3, q4, #32 ;construct src_ptr[2 [all...] |