/external/libhevc/common/arm/ |
ihevc_weighted_pred_bi.s | 48 @* dst = ( (src1 + lvl_shift1)*wgt0 + (src2 + lvl_shift2)*wgt1 + (off0 + 75 @* @param[in] wgt1 112 @ word32 wgt1, 129 @ r8 => wgt1 169 ldr r8,[sp,#wgt1_offset] @load wgt1 172 mla r4,r12,r8,r4 @(lvl_shift1 * wgt0) + (lvl_shift2 * wgt1) 206 vmull.s16 q4,d1,d7[1] @vmull_n_s16(pi2_src2_val1, (int16_t) wgt1) 218 vmull.s16 q6,d3,d7[1] @vmull_n_s16(pi2_src2_val2, (int16_t) wgt1) ii iteration 225 vmull.s16 q8,d1,d7[1] @vmull_n_s16(pi2_src2_val1, (int16_t) wgt1) iii iteration 237 vmull.s16 q10,d3,d7[1] @vmull_n_s16(pi2_src2_val2, (int16_t) wgt1) iv iteratio [all...] |
ihevc_weighted_pred_neon_intr.c | 339 * dst = ( (src1 + lvl_shift1)*wgt0 + (src2 + lvl_shift2)*wgt1 + (off0 + 366 * @param[in] wgt1 403 WORD32 wgt1, 431 WORD32 tmp_lvl_shift = (lvl_shift1 * wgt0) + (lvl_shift2 * wgt1); 458 i4_tmp1_t2 = vmull_n_s16(pi2_src2_val1, (int16_t)wgt1); 466 i4_tmp2_t2 = vmull_n_s16(pi2_src2_val2, (int16_t)wgt1); 504 * dst = ( (src1 + lvl_shift1)*wgt0 + (src2 + lvl_shift2)*wgt1 + (off0 + 531 * @param[in] wgt1 598 int16x4x2_t wgt0, wgt1; local 620 wgt1 = vzip_s16(tmp_wgt1_u, tmp_wgt1_v) [all...] |
/external/libhevc/common/ |
ihevc_weighted_pred.c | 236 * dst = ( (src1 + lvl_shift1)*wgt0 + (src2 + lvl_shift2)*wgt1 + (off0 + 263 * @param[in] wgt1 300 WORD32 wgt1, 316 i4_tmp += (pi2_src2[col] + lvl_shift2) * wgt1; 337 * dst = ( (src1 + lvl_shift1)*wgt0 + (src2 + lvl_shift2)*wgt1 + (off0 + 364 * @param[in] wgt1
|
ihevc_weighted_pred.h | 70 WORD32 wgt1,
|
ihevc_inter_pred.h | 155 WORD32 wgt1,
|
/external/libhevc/encoder/ |
ihevce_inter_pred.c | 409 WORD32 wgt0, wgt1, off0, off1, shift, lvl_shift0, lvl_shift1; local 723 wgt1 = ps_ref_pic_l1->s_weight_offset.i2_luma_weight; 737 wgt1, [all...] |
/external/libhevc/common/x86/ |
ihevc_weighted_pred_sse42_intr.c | [all...] |
ihevc_weighted_pred_ssse3_intr.c | 679 * dst = ( (src1 + lvl_shift1)*wgt0 + (src2 + lvl_shift2)*wgt1 + (off0 + 706 * @param[in] wgt1 [all...] |