/frameworks/av/media/libstagefright/codecs/amrwbenc/src/asm/ARMV7/ |
Filt_6k_7k_neon.s | 66 VLD1.S16 {Q6, Q7}, [r7]! @ signal[48] ~ signal[63] 75 VSHR.S16 Q10, Q6, #2 96 VLD1.S16 {Q6, Q7}, [r4]! @x[16] ~ X[31] 111 VEXT.8 Q5,Q5,Q6,#2 116 VEXT.8 Q6,Q6,Q7,#2 133 VEXT.8 Q5,Q5,Q6,#2 138 VEXT.8 Q6,Q6,Q7,#2 155 VEXT.8 Q5,Q5,Q6,# [all...] |
Dot_p_neon.s | 42 VLD1.S16 {Q6, Q7}, [r0]! @load 16 Word16 x[] 79 VLD1.S16 {Q6, Q7}, [r0]!
|
Norm_Corr_neon.s | 76 VLD1.S16 {Q6, Q7}, [r14]! 119 VLD1.S16 {Q6, Q7}, [r12]! @ load 16 x[] 140 VLD1.S16 {Q6, Q7}, [r12]! @ load 16 x[]
|
pred_lt4_1_neon.s | 57 VLD1.S16 {Q6, Q7}, [r4]! @load 16 x[]
|
/external/libhevc/decoder/arm/ |
ihevcd_fmt_conv_420sp_to_rgba8888.s | 198 VMULL.S16 Q6,D4,D0[1] @//(U-128)*C2 FOR G 199 VMLAL.S16 Q6,D6,D0[2] @//Q6 = (U-128)*C2 + (V-128)*C3 214 VQSHRN.S32 D12,Q6,#13 @//D12 = [(U-128)*C2 + (V-128)*C3]>>13 4 16-BIT VALUES 216 @//Q6 - WEIGHT FOR G 220 VADDW.U8 Q9,Q6,D30 @//Q9 - HAS Y + G 224 VADDW.U8 Q12,Q6,D31 @//Q12 - HAS Y + G 261 VADDW.U8 Q9,Q6,D28 @//Q3 - HAS Y + G 265 VADDW.U8 Q12,Q6,D29 @//Q12 - HAS Y + G 329 VMULL.S16 Q6,D4,D0[1] @//(U-128)*C2 FOR [all...] |
/frameworks/av/media/libstagefright/codecs/aacenc/src/asm/ARMV7/ |
Radix4FFT_v7.s | 96 VQDMULH.S32 Q10, Q6, Q14 @ MULHIGH(cosx, t0) 99 VQDMULH.S32 Q13, Q6, Q15 @ MULHIGH(sinx, t0) 101 VADD.S32 Q6, Q10, Q11 @ MULHIGH(cosx, t0) + MULHIGH(sinx, t1) 104 VADD.S32 Q4, Q8, Q6 @ r4 = t0 + r6@ 106 VSUB.S32 Q6, Q8, Q6 @ r6 = t0 - r6@ 110 VADD.S32 Q9, Q1, Q6 @ xptr[1] = r1 + r6@ 120 VSUB.S32 Q9, Q1, Q6 @ xptr[1] = r1 - r6@
|
PrePostMDCT_v7.s | 48 VREV64.32 Q9, Q6 109 VREV64.32 Q8, Q6
|
/external/libhevc/common/arm/ |
ihevc_sao_edge_offset_class0_chroma.s | 159 VEXT.8 Q7,Q7,Q6,#14 @pu1_cur_row_tmp = vextq_u8(pu1_cur_row_tmp, pu1_cur_row, 14) 163 VCGT.U8 Q8,Q6,Q7 @vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 167 VCLT.U8 Q9,Q6,Q7 @vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 187 VEXT.8 Q7,Q6,Q7,#2 @pu1_cur_row_tmp = vextq_u8(pu1_cur_row, pu1_cur_row_tmp, 2) 191 VCGT.U8 Q8,Q6,Q7 @vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 194 VCLT.U8 Q9,Q6,Q7 @vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 223 VMOVL.U8 Q6,D13 @pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8(vget_high_u8(pu1_cur_row))) 233 VADDW.S8 Q6,Q6,D17 @pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[1], offset) 235 VMAX.S16 Q6,Q6,Q2 @pi2_tmp_cur_row.val[1] = vmaxq_s16(pi2_tmp_cur_row.val[1], const_min_cli (…) [all...] |
ihevc_sao_edge_offset_class1.s | 132 VCGT.U8 Q6,Q5,Q4 @vcgtq_u8(pu1_cur_row, pu1_top_row) 137 VSUB.U8 Q8,Q7,Q6 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 147 VCGT.U8 Q6,Q5,Q9 @vcgtq_u8(pu1_cur_row, pu1_top_row) 155 VSUB.U8 Q10,Q7,Q6 @sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 158 VADD.I8 Q6,Q0,Q8 @edge_idx = vaddq_s8(const_2, sign_up) 161 VADD.I8 Q6,Q6,Q10 @edge_idx = vaddq_s8(edge_idx, sign_down) 226 VCGT.U8 Q6,Q5,Q9 @vcgtq_u8(pu1_cur_row, pu1_top_row) 228 VSUB.U8 Q10,Q7,Q6 @sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 277 VCGT.U8 Q6,Q5,Q4 @vcgtq_u8(pu1_cur_row, pu1_top_row [all...] |
ihevc_sao_edge_offset_class1_chroma.s | 137 VCGT.U8 Q6,Q5,Q14 @vcgtq_u8(pu1_cur_row, pu1_top_row) 142 VSUB.U8 Q8,Q7,Q6 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 152 VCGT.U8 Q6,Q5,Q9 @vcgtq_u8(pu1_cur_row, pu1_top_row) 160 VSUB.U8 Q10,Q7,Q6 @sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 163 VADD.I8 Q6,Q0,Q8 @edge_idx = vaddq_s8(const_2, sign_up) 166 VADD.I8 Q6,Q6,Q10 @edge_idx = vaddq_s8(edge_idx, sign_down) 239 VCGT.U8 Q6,Q5,Q9 @vcgtq_u8(pu1_cur_row, pu1_top_row) 241 VSUB.U8 Q10,Q7,Q6 @sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 295 VCGT.U8 Q6,Q5,Q14 @vcgtq_u8(pu1_cur_row, pu1_top_row [all...] |
ihevc_sao_edge_offset_class0.s | 153 VEXT.8 Q7,Q7,Q6,#15 @pu1_cur_row_tmp = vextq_u8(pu1_cur_row_tmp, pu1_cur_row, 15) 157 VCGT.U8 Q8,Q6,Q7 @vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 162 VCLT.U8 Q9,Q6,Q7 @vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 177 VEXT.8 Q7,Q6,Q7,#1 @pu1_cur_row_tmp = vextq_u8(pu1_cur_row, pu1_cur_row_tmp, 1) 185 VCGT.U8 Q8,Q6,Q7 @vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 188 VCLT.U8 Q9,Q6,Q7 @vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 295 VEXT.8 Q7,Q7,Q6,#15 @pu1_cur_row_tmp = vextq_u8(pu1_cur_row_tmp, pu1_cur_row, 15) 297 VCGT.U8 Q8,Q6,Q7 @vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 298 VCLT.U8 Q9,Q6,Q7 @vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 303 VEXT.8 Q7,Q6,Q7,#1 @pu1_cur_row_tmp = vextq_u8(pu1_cur_row, pu1_cur_row_tmp, 1 [all...] |
ihevc_sao_edge_offset_class2.s | 247 VCGT.U8 Q7,Q6,Q5 @vcgtq_u8(pu1_cur_row, pu1_top_row) 251 VCLT.U8 Q8,Q6,Q5 @vcltq_u8(pu1_cur_row, pu1_top_row) 292 VCGT.U8 Q5,Q6,Q9 @I vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 295 VCLT.U8 Q9,Q6,Q9 @I vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 316 VMOV Q6,Q8 @I pu1_cur_row = pu1_next_row 359 VCGT.U8 Q12,Q6,Q11 @II vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 363 VCLT.U8 Q11,Q6,Q11 @II vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 424 VMOV Q6,Q15 @III pu1_cur_row = pu1_next_row 460 VCGT.U8 Q5,Q6,Q9 @vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 464 VCLT.U8 Q9,Q6,Q9 @vcltq_u8(pu1_cur_row, pu1_next_row_tmp [all...] |
ihevc_sao_edge_offset_class2_chroma.s | 339 VCGT.U8 Q7,Q6,Q5 @vcgtq_u8(pu1_cur_row, pu1_top_row) 343 VCLT.U8 Q8,Q6,Q5 @vcltq_u8(pu1_cur_row, pu1_top_row) 401 VCGT.U8 Q10,Q6,Q9 @I vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 403 VCLT.U8 Q11,Q6,Q9 @I vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 425 VMOV Q6,Q8 @I pu1_cur_row = pu1_next_row 475 VCGT.U8 Q11,Q6,Q14 @II vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 483 VCLT.U8 Q12,Q6,Q14 @II vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 558 VMOV Q6,Q15 @III pu1_cur_row = pu1_next_row 620 VCGT.U8 Q11,Q6,Q9 @vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 624 VCLT.U8 Q12,Q6,Q9 @vcltq_u8(pu1_cur_row, pu1_next_row_tmp [all...] |
ihevc_sao_edge_offset_class3.s | 262 VCGT.U8 Q7,Q6,Q5 @vcgtq_u8(pu1_cur_row, pu1_top_row) 272 VCLT.U8 Q8,Q6,Q5 @vcltq_u8(pu1_cur_row, pu1_top_row) 307 VCGT.U8 Q5,Q6,Q9 @I vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 308 VCLT.U8 Q9,Q6,Q9 @I vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 331 VMOV Q6,Q8 384 VCGT.U8 Q12,Q6,Q9 @II vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 388 VCLT.U8 Q13,Q6,Q9 @II vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 457 VMOV Q6,Q15 @II pu1_cur_row = pu1_next_row 498 VCGT.U8 Q12,Q6,Q9 @vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 501 VCLT.U8 Q13,Q6,Q9 @vcltq_u8(pu1_cur_row, pu1_next_row_tmp [all...] |
ihevc_sao_edge_offset_class3_chroma.s | 329 VCGT.U8 Q7,Q6,Q5 @vcgtq_u8(pu1_cur_row, pu1_top_row) 333 VCLT.U8 Q8,Q6,Q5 @vcltq_u8(pu1_cur_row, pu1_top_row) 393 VCGT.U8 Q10,Q6,Q9 @I vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 395 VCLT.U8 Q11,Q6,Q9 @I vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 420 VMOV Q6,Q8 @I pu1_cur_row = pu1_next_row 476 VCGT.U8 Q11,Q6,Q14 @II vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 486 VCLT.U8 Q12,Q6,Q14 @II vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 563 VMOV Q6,Q15 @III pu1_cur_row = pu1_next_row 627 VCGT.U8 Q10,Q6,Q9 @vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) [all...] |
/external/chromium_org/third_party/openmax_dl/dl/sp/src/arm/neon/ |
armSP_FFT_CToC_SC16_Radix2_ls_unsafe_s.S | 91 #define qT1 Q6.S32
|
armSP_FFT_CToC_SC16_Radix2_ps_unsafe_s.S | 87 #define qT1 Q6.S32
|
armSP_FFT_CToC_SC16_Radix4_unsafe_s.S | 108 #define qT2 Q6.S32 119 #define qY0 Q6.S16
|
armSP_FFT_CToC_SC32_Radix4_unsafe_s.S | 110 #define qT2 Q6.S64 122 #define qY0 Q6.S32
|
omxSP_FFTInv_CCSToR_S16_Sfs_s.S | 104 #define qT0 Q6.S64 106 #define qT0s Q6.S16
|
armSP_FFTInv_CCSToR_S32_preTwiddleRadix2_unsafe_s.S | 112 #define qT0 Q6.S64
|
armSP_FFT_CToC_FC32_Radix4_fs_unsafe_s.S | 93 #define qY2 Q6.F32
|
armSP_FFT_CToC_SC16_Radix4_fs_unsafe_s.S | 104 #define qY2 Q6.S16
|
/bionic/libm/upstream-freebsd/lib/msun/bsdsrc/ |
b_tgamma.c | 108 #define Q6 -1.76012741431666995019222898833e-03 254 q = Q0 +z*(Q1+z*(Q2+z*(Q3+z*(Q4+z*(Q5+z*(Q6+z*(Q7+z*Q8)))))));
|
/frameworks/av/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p10/src/ |
omxVCM4P10_PredictIntraChroma_8x8_s.s | 136 qLeftDiff QN Q6.S16 156 qC4 QN Q6.S16 164 qSum4 QN Q6.S16
|