/external/libhevc/common/arm/ |
ihevc_weighted_pred_neon_intr.c | 418 int32x4_t i4_tmp2_t1; local 463 i4_tmp2_t1 = vmull_n_s16(pi2_src1_val2, (int16_t)wgt0); 469 i4_tmp2_t1 = vaddq_s32(i4_tmp2_t1, i4_tmp2_t2); 472 i4_tmp2_t1 = vaddq_s32(i4_tmp2_t1, tmp_lvl_shift_t); 475 sto_res_tmp1 = vshlq_s32(i4_tmp2_t1, tmp_shift_t); 587 int32x4_t i4_tmp2_t1; local 649 i4_tmp2_t1 = vmull_s16(pi2_src1_val2, wgt0.val[0]); 655 i4_tmp2_t1 = vaddq_s32(i4_tmp2_t1, i4_tmp2_t2) 749 int32x4_t i4_tmp2_t1; local 898 int32x4_t i4_tmp2_t1; local [all...] |
ihevc_weighted_pred_bi.s | 209 vadd.s32 q5,q5,q6 @vaddq_s32(i4_tmp2_t1, i4_tmp2_t2) ii iteration 214 vadd.s32 q5,q5,q15 @vaddq_s32(i4_tmp2_t1, tmp_lvl_shift_t) ii iteration 218 vshl.s32 q5,q5,q14 @vshlq_s32(i4_tmp2_t1, tmp_shift_t) ii iteration 229 vadd.s32 q9,q9,q10 @vaddq_s32(i4_tmp2_t1, i4_tmp2_t2) iv iteration 232 vadd.s32 q9,q9,q15 @vaddq_s32(i4_tmp2_t1, tmp_lvl_shift_t) iv iteration 236 vshl.s32 q9,q9,q14 @vshlq_s32(i4_tmp2_t1, tmp_shift_t) iv iteration
|
ihevc_weighted_pred_bi_default.s | 188 vqadd.s16 d20,d8,d9 @vaddq_s32(i4_tmp2_t1, i4_tmp2_t2) 189 vqadd.s16 d19,d20,d0 @vaddq_s32(i4_tmp2_t1, tmp_lvl_shift_t) 197 vqadd.s16 d18,d24,d25 @vaddq_s32(i4_tmp2_t1, i4_tmp2_t2) iv iteration 202 vst1.32 {d30[0]},[r14],r5 @store pu1_dst iii iteration @vaddq_s32(i4_tmp2_t1, tmp_lvl_shift_t) iv iteratio 238 vqadd.s16 d20,d8,d9 @vaddq_s32(i4_tmp2_t1, i4_tmp2_t2) 239 vqadd.s16 d19,d20,d0 @vaddq_s32(i4_tmp2_t1, tmp_lvl_shift_t) 279 vqadd.s16 q11,q14,q15 @vaddq_s32(i4_tmp2_t1, i4_tmp2_t2) 281 vqadd.s16 q11,q11,q0 @vaddq_s32(i4_tmp2_t1, tmp_lvl_shift_t) 289 vqadd.s16 q4,q6,q7 @vaddq_s32(i4_tmp2_t1, i4_tmp2_t2) iv iteration 296 vst1.32 {d30},[r14],r5 @store pu1_dst iii iteration @vaddq_s32(i4_tmp2_t1, tmp_lvl_shift_t) iv iterati [all...] |