/external/libhevc/common/arm/ |
ihevc_weighted_pred_uni.s | 48 @* dst = ( (src + lvl_shift) * wgt0 + (1 << (shift - 1)) ) >> shift + 63 @* @param[in] wgt0 96 @ word32 wgt0, 108 @ r4 => wgt0 129 ldr r4,[sp,#40] @load wgt0 133 mul r10,r7,r4 @lvl_shift * wgt0 136 add r10,r10,r5,lsl r6 @lvl_shift * wgt0 + (off0 << shift) 159 vmull.s16 q2,d1,d0[0] @vmull_n_s16(pi2_src_val1, (int16_t) wgt0) 164 vmull.s16 q3,d2,d0[0] @vmull_n_s16(pi2_src_val2, (int16_t) wgt0) ii iteration 170 vmull.s16 q5,d8,d0[0] @vmull_n_s16(pi2_src_val1, (int16_t) wgt0) iii iteratio [all...] |
ihevc_weighted_pred_neon_intr.c | 60 * dst = ( (src + lvl_shift) * wgt0 + (1 << (shift - 1)) ) >> shift + 75 * @param[in] wgt0 108 WORD32 wgt0, 130 WORD32 tmp_lvl_shift = lvl_shift * wgt0 + (off0 << shift); 153 i4_tmp1_t = vmull_n_s16(pi2_src_val1, (int16_t)wgt0); 156 i4_tmp2_t = vmull_n_s16(pi2_src_val2, (int16_t)wgt0); 191 * dst = ( (src + lvl_shift) * wgt0 + (1 << (shift - 1)) ) >> shift + 206 * @param[in] wgt0 262 int16x4x2_t wgt0; local 280 wgt0 = vzip_s16(tmp_wgt0_u, tmp_wgt0_v) 598 int16x4x2_t wgt0, wgt1; local [all...] |
ihevc_weighted_pred_bi.s | 48 @* dst = ( (src1 + lvl_shift1)*wgt0 + (src2 + lvl_shift2)*wgt1 + (off0 + 69 @* @param[in] wgt0 110 @ word32 wgt0, 127 @ r6 => wgt0 151 ldr r6,[sp,#48] @load wgt0 155 mul r4,r11,r6 @lvl_shift1 * wgt0 159 mla r4,r12,r8,r4 @(lvl_shift1 * wgt0) + (lvl_shift2 * wgt1) 191 vmull.s16 q2,d0,d7[0] @vmull_n_s16(pi2_src1_val1, (int16_t) wgt0) 198 vmull.s16 q5,d2,d7[0] @vmull_n_s16(pi2_src1_val2, (int16_t) wgt0) ii iteration 202 vmull.s16 q7,d0,d7[0] @vmull_n_s16(pi2_src1_val1, (int16_t) wgt0) iii iteratio [all...] |
/external/libhevc/common/ |
ihevc_weighted_pred.c | 62 * dst = ( (src + lvl_shift) * wgt0 + (1 << (shift - 1)) ) >> shift + 77 * @param[in] wgt0 110 WORD32 wgt0, 124 i4_tmp = (pi2_src[col] + lvl_shift) * wgt0; 145 * dst = ( (src + lvl_shift) * wgt0 + (1 << (shift - 1)) ) >> shift + 160 * @param[in] wgt0 236 * dst = ( (src1 + lvl_shift1)*wgt0 + (src2 + lvl_shift2)*wgt1 + (off0 + 257 * @param[in] wgt0 298 WORD32 wgt0, 315 i4_tmp = (pi2_src1[col] + lvl_shift1) * wgt0; [all...] |
ihevc_weighted_pred.h | 42 WORD32 wgt0, 68 WORD32 wgt0,
|
ihevc_inter_pred.h | 137 WORD32 wgt0, 153 WORD32 wgt0,
|
/external/libhevc/common/x86/ |
ihevc_weighted_pred_sse42_intr.c | 67 * dst = ( (src + lvl_shift) * wgt0 + (1 << (shift - 1)) ) >> shift + 82 * @param[in] wgt0 115 WORD32 wgt0, 138 wgt0_4x32b = _mm_set1_epi32(wgt0); 181 /*i4_tmp = (pi2_src[col] + lvl_shift) * wgt0*/ /* First 4 pixels */ 199 /*i4_tmp = (pi2_src[col] + lvl_shift) * wgt0*/ /* Last 4 pixels */ 297 /*i4_tmp = (pi2_src[col] + lvl_shift) * wgt0*/ 366 * dst = ( (src + lvl_shift) * wgt0 + (1 << (shift - 1)) ) >> shift + 381 * @param[in] wgt0 479 /*i4_tmp = (pi2_src[col] + lvl_shift) * wgt0*/ [all...] |
ihevc_weighted_pred_ssse3_intr.c | 69 * dst = ( (src + lvl_shift) * wgt0 + (1 << (shift - 1)) ) >> shift + 84 * @param[in] wgt0 117 WORD32 wgt0, 139 wgt0_8x16b = _mm_set1_epi16(wgt0); 141 /* lvl_shift * wgt0 */ 149 /* lvl_shift * wgt0 */ 151 /* lvl_shift * wgt0 + 1 << (shift - 1) */ 173 /*i4_tmp = (pi2_src[col]) * wgt0*/ /* Lower 16 bit */ 179 /*i4_tmp = (pi2_src[col] ) * wgt0*/ /* Higher 16 bit */ 196 /* i4_tmp = (pi2_src[col] + lvl_shift) * wgt0 + 1 << (shift - 1) * [all...] |