HomeSort by relevance Sort by last modified time
    Searched refs:comp_pred (Results 1 - 25 of 28) sorted by null

1 2

  /external/libvpx/libvpx/vpx_dsp/x86/
avg_pred_sse2.c 18 void vpx_comp_avg_pred_sse2(uint8_t *comp_pred, const uint8_t *pred, int width,
20 /* comp_pred and pred must be 16 byte aligned. */
21 assert(((intptr_t)comp_pred & 0xf) == 0);
30 _mm_store_si128((__m128i *)(comp_pred + x), avg);
32 comp_pred += width;
38 // Process 16 elements at a time. comp_pred and pred have width == stride
63 _mm_store_si128((__m128i *)comp_pred, avg);
66 comp_pred += 16;
  /external/libaom/libaom/aom_dsp/
sad.c 53 uint8_t comp_pred[m * n]; \
54 aom_comp_avg_pred(comp_pred, second_pred, m, n, ref, ref_stride); \
55 return sad(src, src_stride, comp_pred, m, m, n); \
60 uint8_t comp_pred[m * n]; \
61 aom_dist_wtd_comp_avg_pred_c(comp_pred, second_pred, m, n, ref, \
63 return sad(src, src_stride, comp_pred, m, m, n); \
206 uint16_t comp_pred[m * n]; \
207 aom_highbd_comp_avg_pred(CONVERT_TO_BYTEPTR(comp_pred), second_pred, m, n, \
209 return highbd_sadb(src, src_stride, comp_pred, m, m, n); \
214 uint16_t comp_pred[m * n];
    [all...]
variance.c 267 void aom_comp_avg_pred_c(uint8_t *comp_pred, const uint8_t *pred, int width,
274 comp_pred[j] = ROUND_POWER_OF_TWO(tmp, 1);
276 comp_pred += width;
285 uint8_t *comp_pred, int width, int height,
361 av1_make_inter_predictor(pre, pre_buf->stride, comp_pred, width,
375 memcpy(comp_pred, ref, width * sizeof(*comp_pred));
376 comp_pred += width;
382 aom_convolve8_horiz_c(ref, ref_stride, comp_pred, width, kernel, 16, NULL,
387 aom_convolve8_vert_c(ref, ref_stride, comp_pred, width, NULL, -1, kernel
862 uint16_t *comp_pred = CONVERT_TO_SHORTPTR(comp_pred8); local
967 uint16_t *comp_pred = CONVERT_TO_SHORTPTR(comp_pred8); local
1012 uint16_t *comp_pred = CONVERT_TO_SHORTPTR(comp_pred8); local
1034 uint16_t *comp_pred = CONVERT_TO_SHORTPTR(comp_pred8); local
1058 uint16_t *comp_pred = CONVERT_TO_SHORTPTR(comp_pred8); local
1163 uint16_t *comp_pred = CONVERT_TO_SHORTPTR(comp_pred8); local
    [all...]
  /external/libaom/libaom/aom_dsp/x86/
jnt_variance_ssse3.c 48 void aom_dist_wtd_comp_avg_pred_ssse3(uint8_t *comp_pred, const uint8_t *pred,
70 compute_dist_wtd_avg(&p0, &p1, &w, &r, comp_pred);
72 comp_pred += 16;
88 compute_dist_wtd_avg(&p0, &p1, &w, &r, comp_pred);
90 comp_pred += 16;
110 compute_dist_wtd_avg(&p0, &p1, &w, &r, comp_pred);
112 comp_pred += 16;
121 const MV *const mv, uint8_t *comp_pred, const uint8_t *pred, int width,
126 aom_upsampled_pred(xd, cm, mi_row, mi_col, mv, comp_pred, width, height,
141 __m128i p0 = xx_loadu_128(comp_pred);
    [all...]
variance_sse2.c 487 uint8_t *comp_pred, int width, int height,
564 av1_make_inter_predictor(pre, pre_buf->stride, comp_pred, width,
587 xx_storeu_128(comp_pred, xx_loadu_128(ref));
588 comp_pred += 16;
601 xx_storeu_128(comp_pred, _mm_unpacklo_epi64(s0, s1));
602 comp_pred += 16;
617 xx_storeu_128(comp_pred, reg);
618 comp_pred += 16;
625 aom_convolve8_horiz(ref, ref_stride, comp_pred, width, kernel, 16, NULL, -1,
630 aom_convolve8_vert(ref, ref_stride, comp_pred, width, NULL, -1, kernel, 16
722 uint16_t *comp_pred = CONVERT_TO_SHORTPTR(comp_pred8); local
    [all...]
masked_variance_intrin_ssse3.h 50 static INLINE void comp_mask_pred_8_ssse3(uint8_t *comp_pred, int height,
82 // comp_pred's stride == width == 8
83 _mm_store_si128((__m128i *)(comp_pred), round);
84 comp_pred += (8 << 1);
jnt_sad_ssse3.c 199 uint8_t comp_pred[m * n]; \
200 aom_dist_wtd_comp_avg_pred(comp_pred, second_pred, m, n, ref, ref_stride, \
202 return aom_sad##m##xh_sse2(src, src_stride, comp_pred, m, m, n); \
209 uint8_t comp_pred[m * n]; \
210 aom_dist_wtd_comp_avg_pred(comp_pred, second_pred, m, n, ref, ref_stride, \
212 return aom_sad##m##xh_avx2(src, src_stride, comp_pred, m, m, n); \
variance_avx2.c 335 uint8_t *comp_pred) {
353 _mm256_storeu_si256((__m256i *)(comp_pred), roundA);
356 void aom_comp_mask_pred_avx2(uint8_t *comp_pred, const uint8_t *pred, int width,
366 comp_mask_pred_8_ssse3(comp_pred, height, src0, stride0, src1, stride1,
382 // comp_pred's stride == width == 16
383 comp_mask_pred_line_avx2(sA0, sA1, aA, comp_pred);
384 comp_mask_pred_line_avx2(sB0, sB1, aB, comp_pred + 32);
385 comp_pred += (16 << 2);
399 comp_mask_pred_line_avx2(sA0, sA1, aA, comp_pred);
400 comp_mask_pred_line_avx2(sB0, sB1, aB, comp_pred + 32)
443 uint16_t *comp_pred = CONVERT_TO_SHORTPTR(comp_pred8); local
    [all...]
highbd_variance_sse2.c 707 uint16_t *comp_pred = CONVERT_TO_SHORTPTR(comp_pred8); local
820 uint16_t *comp_pred = CONVERT_TO_SHORTPTR(comp_pred8); local
    [all...]
masked_variance_intrin_ssse3.c     [all...]
  /external/libvpx/libvpx/vpx_dsp/ppc/
sad_vsx.c 118 DECLARE_ALIGNED(16, uint8_t, comp_pred[16 * height]); \
119 vpx_comp_avg_pred_vsx(comp_pred, second_pred, 16, height, ref, \
122 return vpx_sad16x##height##_vsx(src, src_stride, comp_pred, 16); \
129 DECLARE_ALIGNED(32, uint8_t, comp_pred[32 * height]); \
130 vpx_comp_avg_pred_vsx(comp_pred, second_pred, 32, height, ref, \
133 return vpx_sad32x##height##_vsx(src, src_stride, comp_pred, 32); \
140 DECLARE_ALIGNED(64, uint8_t, comp_pred[64 * height]); \
141 vpx_comp_avg_pred_vsx(comp_pred, second_pred, 64, height, ref, \
143 return vpx_sad64x##height##_vsx(src, src_stride, comp_pred, 64); \
variance_vsx.c 54 void vpx_comp_avg_pred_vsx(uint8_t *comp_pred, const uint8_t *pred, int width,
57 /* comp_pred and pred must be 16 byte aligned. */
58 assert(((intptr_t)comp_pred & 0xf) == 0);
64 vec_vsx_st(v, j, comp_pred);
66 comp_pred += width;
77 vec_vsx_st(v, 0, comp_pred);
78 comp_pred += 16; // width * 2;
93 vec_vsx_st(v, 0, comp_pred);
94 comp_pred += 16; // width * 4;
  /external/libvpx/libvpx/vpx_dsp/
sad.c 43 DECLARE_ALIGNED(16, uint8_t, comp_pred[m * n]); \
44 vpx_comp_avg_pred_c(comp_pred, second_pred, m, n, ref_ptr, ref_stride); \
45 return sad(src_ptr, src_stride, comp_pred, m, m, n); \
177 DECLARE_ALIGNED(16, uint16_t, comp_pred[m * n]); \
178 vpx_highbd_comp_avg_pred_c(comp_pred, CONVERT_TO_SHORTPTR(second_pred), m, \
180 return highbd_sadb(src_ptr, src_stride, comp_pred, m, m, n); \
variance.c 223 void vpx_comp_avg_pred_c(uint8_t *comp_pred, const uint8_t *pred, int width,
230 comp_pred[j] = ROUND_POWER_OF_TWO(tmp, 1);
232 comp_pred += width;
552 void vpx_highbd_comp_avg_pred(uint16_t *comp_pred, const uint16_t *pred,
559 comp_pred[j] = ROUND_POWER_OF_TWO(tmp, 1);
561 comp_pred += width;
  /external/libaom/libaom/test/
sad_test.cc 38 typedef void (*DistWtdCompAvgFunc)(uint8_t *comp_pred, const uint8_t *pred,
196 const uint8_t comp_pred = ROUND_POWER_OF_TWO(tmp, 1); local
197 sad += abs(source8[h * source_stride_ + w] - comp_pred);
201 const uint16_t comp_pred = ROUND_POWER_OF_TWO(tmp, 1); local
202 sad += abs(source16[h * source_stride_ + w] - comp_pred);
249 const uint8_t comp_pred = ROUND_POWER_OF_TWO(tmp, 4); local
250 sad += abs(source8[h * source_stride_ + w] - comp_pred);
255 const uint16_t comp_pred = ROUND_POWER_OF_TWO(tmp, 4); local
256 sad += abs(source16[h * source_stride_ + w] - comp_pred);
    [all...]
comp_avg_pred_test.h 30 typedef void (*distwtdcompavg_func)(uint8_t *comp_pred, const uint8_t *pred,
37 const MV *const mv, uint8_t *comp_pred, const uint8_t *pred, int width,
comp_mask_variance_test.cc 32 typedef void (*comp_mask_pred_func)(uint8_t *comp_pred, const uint8_t *pred,
  /external/libvpx/libvpx/test/
sad_test.cc 173 const uint8_t comp_pred = ROUND_POWER_OF_TWO(tmp, 1); local
174 sad += abs(source8[h * source_stride_ + w] - comp_pred);
179 const uint16_t comp_pred = ROUND_POWER_OF_TWO(tmp, 1); local
180 sad += abs(source16[h * source_stride_ + w] - comp_pred);
    [all...]
  /external/libvpx/libvpx/vp9/encoder/
vp9_mcomp.c 316 DECLARE_ALIGNED(16, uint8_t, comp_pred[64 * 64]);
317 vpx_comp_avg_pred(comp_pred, second_pred, w, h, y + offset, y_stride);
318 besterr = vfp->vf(comp_pred, w, src, src_stride, sse1);
331 DECLARE_ALIGNED(16, uint8_t, comp_pred[64 * 64]);
332 vpx_comp_avg_pred(comp_pred, second_pred, w, h, y + offset, y_stride);
333 besterr = vfp->vf(comp_pred, w, src, src_stride, sse1);
654 DECLARE_ALIGNED(16, uint8_t, comp_pred[64 * 64]);
655 vpx_comp_avg_pred(comp_pred, second_pred, w, h, pred, w);
656 besterr = vfp->vf(comp_pred, w, src_address, src_stride, sse);
673 DECLARE_ALIGNED(16, uint8_t, comp_pred[64 * 64])
    [all...]
vp9_rdopt.c 3162 int comp_pred, i, k; local
3845 const int comp_pred = 0; local
3945 int comp_pred, i; local
    [all...]
vp9_pickmode.c 1954 int comp_pred = 0; local
    [all...]
  /external/libaom/libaom/av1/encoder/
mcomp.c 343 uint8_t *comp_pred = CONVERT_TO_BYTEPTR(comp_pred16); local
345 aom_highbd_comp_mask_pred(comp_pred, second_pred, w, h, y + offset,
348 aom_highbd_comp_avg_pred(comp_pred, second_pred, w, h, y + offset,
351 besterr = vfp->vf(comp_pred, w, src, src_stride, sse1);
353 DECLARE_ALIGNED(16, uint8_t, comp_pred[MAX_SB_SQUARE]);
355 aom_comp_mask_pred(comp_pred, second_pred, w, h, y + offset, y_stride,
358 aom_comp_avg_pred(comp_pred, second_pred, w, h, y + offset, y_stride);
360 besterr = vfp->vf(comp_pred, w, src, src_stride, sse1);
    [all...]
rdopt.c 11793 const int comp_pred = ref_frame[1] > INTRA_FRAME; local
11852 const int comp_pred = ref_frame[1] > INTRA_FRAME; local
12740 const int comp_pred = second_ref_frame > INTRA_FRAME; local
13300 const int comp_pred = second_ref_frame > INTRA_FRAME; local
13710 const int comp_pred = 0; local
    [all...]
  /external/libvpx/config/arm-neon/
vpx_dsp_rtcd.h 32 void vpx_comp_avg_pred_c(uint8_t *comp_pred, const uint8_t *pred, int width, int height, const uint8_t *ref, int ref_stride);
33 void vpx_comp_avg_pred_neon(uint8_t *comp_pred, const uint8_t *pred, int width, int height, const uint8_t *ref, int ref_stride);
    [all...]
  /external/libvpx/config/arm64/
vpx_dsp_rtcd.h 32 void vpx_comp_avg_pred_c(uint8_t *comp_pred, const uint8_t *pred, int width, int height, const uint8_t *ref, int ref_stride);
33 void vpx_comp_avg_pred_neon(uint8_t *comp_pred, const uint8_t *pred, int width, int height, const uint8_t *ref, int ref_stride);
    [all...]

Completed in 704 milliseconds

1 2