/external/libvpx/libvpx/vpx_dsp/x86/ |
avg_pred_sse2.c | 18 void vpx_comp_avg_pred_sse2(uint8_t *comp_pred, const uint8_t *pred, int width, 20 /* comp_pred and pred must be 16 byte aligned. */ 21 assert(((intptr_t)comp_pred & 0xf) == 0); 30 _mm_store_si128((__m128i *)(comp_pred + x), avg); 32 comp_pred += width; 38 // Process 16 elements at a time. comp_pred and pred have width == stride 63 _mm_store_si128((__m128i *)comp_pred, avg); 66 comp_pred += 16;
|
/external/libaom/libaom/aom_dsp/ |
sad.c | 53 uint8_t comp_pred[m * n]; \ 54 aom_comp_avg_pred(comp_pred, second_pred, m, n, ref, ref_stride); \ 55 return sad(src, src_stride, comp_pred, m, m, n); \ 60 uint8_t comp_pred[m * n]; \ 61 aom_dist_wtd_comp_avg_pred_c(comp_pred, second_pred, m, n, ref, \ 63 return sad(src, src_stride, comp_pred, m, m, n); \ 206 uint16_t comp_pred[m * n]; \ 207 aom_highbd_comp_avg_pred(CONVERT_TO_BYTEPTR(comp_pred), second_pred, m, n, \ 209 return highbd_sadb(src, src_stride, comp_pred, m, m, n); \ 214 uint16_t comp_pred[m * n]; [all...] |
variance.c | 267 void aom_comp_avg_pred_c(uint8_t *comp_pred, const uint8_t *pred, int width, 274 comp_pred[j] = ROUND_POWER_OF_TWO(tmp, 1); 276 comp_pred += width; 285 uint8_t *comp_pred, int width, int height, 361 av1_make_inter_predictor(pre, pre_buf->stride, comp_pred, width, 375 memcpy(comp_pred, ref, width * sizeof(*comp_pred)); 376 comp_pred += width; 382 aom_convolve8_horiz_c(ref, ref_stride, comp_pred, width, kernel, 16, NULL, 387 aom_convolve8_vert_c(ref, ref_stride, comp_pred, width, NULL, -1, kernel 862 uint16_t *comp_pred = CONVERT_TO_SHORTPTR(comp_pred8); local 967 uint16_t *comp_pred = CONVERT_TO_SHORTPTR(comp_pred8); local 1012 uint16_t *comp_pred = CONVERT_TO_SHORTPTR(comp_pred8); local 1034 uint16_t *comp_pred = CONVERT_TO_SHORTPTR(comp_pred8); local 1058 uint16_t *comp_pred = CONVERT_TO_SHORTPTR(comp_pred8); local 1163 uint16_t *comp_pred = CONVERT_TO_SHORTPTR(comp_pred8); local [all...] |
/external/libaom/libaom/aom_dsp/x86/ |
jnt_variance_ssse3.c | 48 void aom_dist_wtd_comp_avg_pred_ssse3(uint8_t *comp_pred, const uint8_t *pred, 70 compute_dist_wtd_avg(&p0, &p1, &w, &r, comp_pred); 72 comp_pred += 16; 88 compute_dist_wtd_avg(&p0, &p1, &w, &r, comp_pred); 90 comp_pred += 16; 110 compute_dist_wtd_avg(&p0, &p1, &w, &r, comp_pred); 112 comp_pred += 16; 121 const MV *const mv, uint8_t *comp_pred, const uint8_t *pred, int width, 126 aom_upsampled_pred(xd, cm, mi_row, mi_col, mv, comp_pred, width, height, 141 __m128i p0 = xx_loadu_128(comp_pred); [all...] |
variance_sse2.c | 487 uint8_t *comp_pred, int width, int height, 564 av1_make_inter_predictor(pre, pre_buf->stride, comp_pred, width, 587 xx_storeu_128(comp_pred, xx_loadu_128(ref)); 588 comp_pred += 16; 601 xx_storeu_128(comp_pred, _mm_unpacklo_epi64(s0, s1)); 602 comp_pred += 16; 617 xx_storeu_128(comp_pred, reg); 618 comp_pred += 16; 625 aom_convolve8_horiz(ref, ref_stride, comp_pred, width, kernel, 16, NULL, -1, 630 aom_convolve8_vert(ref, ref_stride, comp_pred, width, NULL, -1, kernel, 16 722 uint16_t *comp_pred = CONVERT_TO_SHORTPTR(comp_pred8); local [all...] |
masked_variance_intrin_ssse3.h | 50 static INLINE void comp_mask_pred_8_ssse3(uint8_t *comp_pred, int height, 82 // comp_pred's stride == width == 8 83 _mm_store_si128((__m128i *)(comp_pred), round); 84 comp_pred += (8 << 1);
|
jnt_sad_ssse3.c | 199 uint8_t comp_pred[m * n]; \ 200 aom_dist_wtd_comp_avg_pred(comp_pred, second_pred, m, n, ref, ref_stride, \ 202 return aom_sad##m##xh_sse2(src, src_stride, comp_pred, m, m, n); \ 209 uint8_t comp_pred[m * n]; \ 210 aom_dist_wtd_comp_avg_pred(comp_pred, second_pred, m, n, ref, ref_stride, \ 212 return aom_sad##m##xh_avx2(src, src_stride, comp_pred, m, m, n); \
|
variance_avx2.c | 335 uint8_t *comp_pred) { 353 _mm256_storeu_si256((__m256i *)(comp_pred), roundA); 356 void aom_comp_mask_pred_avx2(uint8_t *comp_pred, const uint8_t *pred, int width, 366 comp_mask_pred_8_ssse3(comp_pred, height, src0, stride0, src1, stride1, 382 // comp_pred's stride == width == 16 383 comp_mask_pred_line_avx2(sA0, sA1, aA, comp_pred); 384 comp_mask_pred_line_avx2(sB0, sB1, aB, comp_pred + 32); 385 comp_pred += (16 << 2); 399 comp_mask_pred_line_avx2(sA0, sA1, aA, comp_pred); 400 comp_mask_pred_line_avx2(sB0, sB1, aB, comp_pred + 32) 443 uint16_t *comp_pred = CONVERT_TO_SHORTPTR(comp_pred8); local [all...] |
highbd_variance_sse2.c | 707 uint16_t *comp_pred = CONVERT_TO_SHORTPTR(comp_pred8); local 820 uint16_t *comp_pred = CONVERT_TO_SHORTPTR(comp_pred8); local [all...] |
masked_variance_intrin_ssse3.c | [all...] |
/external/libvpx/libvpx/vpx_dsp/ppc/ |
sad_vsx.c | 118 DECLARE_ALIGNED(16, uint8_t, comp_pred[16 * height]); \ 119 vpx_comp_avg_pred_vsx(comp_pred, second_pred, 16, height, ref, \ 122 return vpx_sad16x##height##_vsx(src, src_stride, comp_pred, 16); \ 129 DECLARE_ALIGNED(32, uint8_t, comp_pred[32 * height]); \ 130 vpx_comp_avg_pred_vsx(comp_pred, second_pred, 32, height, ref, \ 133 return vpx_sad32x##height##_vsx(src, src_stride, comp_pred, 32); \ 140 DECLARE_ALIGNED(64, uint8_t, comp_pred[64 * height]); \ 141 vpx_comp_avg_pred_vsx(comp_pred, second_pred, 64, height, ref, \ 143 return vpx_sad64x##height##_vsx(src, src_stride, comp_pred, 64); \
|
variance_vsx.c | 54 void vpx_comp_avg_pred_vsx(uint8_t *comp_pred, const uint8_t *pred, int width, 57 /* comp_pred and pred must be 16 byte aligned. */ 58 assert(((intptr_t)comp_pred & 0xf) == 0); 64 vec_vsx_st(v, j, comp_pred); 66 comp_pred += width; 77 vec_vsx_st(v, 0, comp_pred); 78 comp_pred += 16; // width * 2; 93 vec_vsx_st(v, 0, comp_pred); 94 comp_pred += 16; // width * 4;
|
/external/libvpx/libvpx/vpx_dsp/ |
sad.c | 43 DECLARE_ALIGNED(16, uint8_t, comp_pred[m * n]); \ 44 vpx_comp_avg_pred_c(comp_pred, second_pred, m, n, ref_ptr, ref_stride); \ 45 return sad(src_ptr, src_stride, comp_pred, m, m, n); \ 177 DECLARE_ALIGNED(16, uint16_t, comp_pred[m * n]); \ 178 vpx_highbd_comp_avg_pred_c(comp_pred, CONVERT_TO_SHORTPTR(second_pred), m, \ 180 return highbd_sadb(src_ptr, src_stride, comp_pred, m, m, n); \
|
variance.c | 223 void vpx_comp_avg_pred_c(uint8_t *comp_pred, const uint8_t *pred, int width, 230 comp_pred[j] = ROUND_POWER_OF_TWO(tmp, 1); 232 comp_pred += width; 552 void vpx_highbd_comp_avg_pred(uint16_t *comp_pred, const uint16_t *pred, 559 comp_pred[j] = ROUND_POWER_OF_TWO(tmp, 1); 561 comp_pred += width;
|
/external/libaom/libaom/test/ |
sad_test.cc | 38 typedef void (*DistWtdCompAvgFunc)(uint8_t *comp_pred, const uint8_t *pred, 196 const uint8_t comp_pred = ROUND_POWER_OF_TWO(tmp, 1); local 197 sad += abs(source8[h * source_stride_ + w] - comp_pred); 201 const uint16_t comp_pred = ROUND_POWER_OF_TWO(tmp, 1); local 202 sad += abs(source16[h * source_stride_ + w] - comp_pred); 249 const uint8_t comp_pred = ROUND_POWER_OF_TWO(tmp, 4); local 250 sad += abs(source8[h * source_stride_ + w] - comp_pred); 255 const uint16_t comp_pred = ROUND_POWER_OF_TWO(tmp, 4); local 256 sad += abs(source16[h * source_stride_ + w] - comp_pred); [all...] |
comp_avg_pred_test.h | 30 typedef void (*distwtdcompavg_func)(uint8_t *comp_pred, const uint8_t *pred, 37 const MV *const mv, uint8_t *comp_pred, const uint8_t *pred, int width,
|
comp_mask_variance_test.cc | 32 typedef void (*comp_mask_pred_func)(uint8_t *comp_pred, const uint8_t *pred,
|
/external/libvpx/libvpx/test/ |
sad_test.cc | 173 const uint8_t comp_pred = ROUND_POWER_OF_TWO(tmp, 1); local 174 sad += abs(source8[h * source_stride_ + w] - comp_pred); 179 const uint16_t comp_pred = ROUND_POWER_OF_TWO(tmp, 1); local 180 sad += abs(source16[h * source_stride_ + w] - comp_pred); [all...] |
/external/libvpx/libvpx/vp9/encoder/ |
vp9_mcomp.c | 316 DECLARE_ALIGNED(16, uint8_t, comp_pred[64 * 64]); 317 vpx_comp_avg_pred(comp_pred, second_pred, w, h, y + offset, y_stride); 318 besterr = vfp->vf(comp_pred, w, src, src_stride, sse1); 331 DECLARE_ALIGNED(16, uint8_t, comp_pred[64 * 64]); 332 vpx_comp_avg_pred(comp_pred, second_pred, w, h, y + offset, y_stride); 333 besterr = vfp->vf(comp_pred, w, src, src_stride, sse1); 654 DECLARE_ALIGNED(16, uint8_t, comp_pred[64 * 64]); 655 vpx_comp_avg_pred(comp_pred, second_pred, w, h, pred, w); 656 besterr = vfp->vf(comp_pred, w, src_address, src_stride, sse); 673 DECLARE_ALIGNED(16, uint8_t, comp_pred[64 * 64]) [all...] |
vp9_rdopt.c | 3162 int comp_pred, i, k; local 3845 const int comp_pred = 0; local 3945 int comp_pred, i; local [all...] |
vp9_pickmode.c | 1954 int comp_pred = 0; local [all...] |
/external/libaom/libaom/av1/encoder/ |
mcomp.c | 343 uint8_t *comp_pred = CONVERT_TO_BYTEPTR(comp_pred16); local 345 aom_highbd_comp_mask_pred(comp_pred, second_pred, w, h, y + offset, 348 aom_highbd_comp_avg_pred(comp_pred, second_pred, w, h, y + offset, 351 besterr = vfp->vf(comp_pred, w, src, src_stride, sse1); 353 DECLARE_ALIGNED(16, uint8_t, comp_pred[MAX_SB_SQUARE]); 355 aom_comp_mask_pred(comp_pred, second_pred, w, h, y + offset, y_stride, 358 aom_comp_avg_pred(comp_pred, second_pred, w, h, y + offset, y_stride); 360 besterr = vfp->vf(comp_pred, w, src, src_stride, sse1); [all...] |
rdopt.c | 11793 const int comp_pred = ref_frame[1] > INTRA_FRAME; local 11852 const int comp_pred = ref_frame[1] > INTRA_FRAME; local 12740 const int comp_pred = second_ref_frame > INTRA_FRAME; local 13300 const int comp_pred = second_ref_frame > INTRA_FRAME; local 13710 const int comp_pred = 0; local [all...] |
/external/libvpx/config/arm-neon/ |
vpx_dsp_rtcd.h | 32 void vpx_comp_avg_pred_c(uint8_t *comp_pred, const uint8_t *pred, int width, int height, const uint8_t *ref, int ref_stride); 33 void vpx_comp_avg_pred_neon(uint8_t *comp_pred, const uint8_t *pred, int width, int height, const uint8_t *ref, int ref_stride); [all...] |
/external/libvpx/config/arm64/ |
vpx_dsp_rtcd.h | 32 void vpx_comp_avg_pred_c(uint8_t *comp_pred, const uint8_t *pred, int width, int height, const uint8_t *ref, int ref_stride); 33 void vpx_comp_avg_pred_neon(uint8_t *comp_pred, const uint8_t *pred, int width, int height, const uint8_t *ref, int ref_stride); [all...] |