/external/libhevc/decoder/arm/ |
ihevcd_fmt_conv_420sp_to_rgba8888.s | 200 VMULL.S16 Q7,D5,D0[1] @//(U-128)*C2 FOR G 201 VMLAL.S16 Q7,D7,D0[2] @//Q7 = (U-128)*C2 + (V-128)*C3 215 VQSHRN.S32 D13,Q7,#13 @//D13 = [(U-128)*C2 + (V-128)*C3]>>13 4 16-BIT VALUES 218 VADDW.U8 Q7,Q4,D30 @//Q7 - HAS Y + B 226 VQMOVUN.S16 D14,Q7 233 VZIP.16 Q7,Q8 245 VZIP.32 Q7,Q10 259 VADDW.U8 Q7,Q4,D28 @//Q7 - HAS Y + [all...] |
/frameworks/av/media/libstagefright/codecs/amrwbenc/src/asm/ARMV7/ |
Filt_6k_7k_neon.s | 66 VLD1.S16 {Q6, Q7}, [r7]! @ signal[48] ~ signal[63] 76 VSHR.S16 Q11, Q7, #2 96 VLD1.S16 {Q6, Q7}, [r4]! @x[16] ~ X[31] 116 VEXT.8 Q6,Q6,Q7,#2 121 VEXT.8 Q7,Q7,Q8,#2 138 VEXT.8 Q6,Q6,Q7,#2 143 VEXT.8 Q7,Q7,Q8,#2 160 VEXT.8 Q6,Q6,Q7,# [all...] |
syn_filt_neon.s | 82 VDUP.S32 Q7, D10[0] 84 VSUB.S32 Q9, Q10, Q7
|
Dot_p_neon.s | 42 VLD1.S16 {Q6, Q7}, [r0]! @load 16 Word16 x[] 79 VLD1.S16 {Q6, Q7}, [r0]!
|
Norm_Corr_neon.s | 76 VLD1.S16 {Q6, Q7}, [r14]! 119 VLD1.S16 {Q6, Q7}, [r12]! @ load 16 x[] 140 VLD1.S16 {Q6, Q7}, [r12]! @ load 16 x[]
|
pred_lt4_1_neon.s | 57 VLD1.S16 {Q6, Q7}, [r4]! @load 16 x[]
|
/external/libhevc/common/arm/ |
ihevc_sao_edge_offset_class0.s | 153 VEXT.8 Q7,Q7,Q6,#15 @pu1_cur_row_tmp = vextq_u8(pu1_cur_row_tmp, pu1_cur_row, 15) 157 VCGT.U8 Q8,Q6,Q7 @vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 162 VCLT.U8 Q9,Q6,Q7 @vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 177 VEXT.8 Q7,Q6,Q7,#1 @pu1_cur_row_tmp = vextq_u8(pu1_cur_row, pu1_cur_row_tmp, 1) 185 VCGT.U8 Q8,Q6,Q7 @vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 188 VCLT.U8 Q9,Q6,Q7 @vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 195 VADD.I8 Q7,Q1,Q10 @edge_idx = vaddq_s8(const_2, sign_left) 198 VADD.I8 Q7,Q7,Q11 @edge_idx = vaddq_s8(edge_idx, sign_right [all...] |
ihevc_sao_edge_offset_class0_chroma.s | 159 VEXT.8 Q7,Q7,Q6,#14 @pu1_cur_row_tmp = vextq_u8(pu1_cur_row_tmp, pu1_cur_row, 14) 163 VCGT.U8 Q8,Q6,Q7 @vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 167 VCLT.U8 Q9,Q6,Q7 @vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 187 VEXT.8 Q7,Q6,Q7,#2 @pu1_cur_row_tmp = vextq_u8(pu1_cur_row, pu1_cur_row_tmp, 2) 191 VCGT.U8 Q8,Q6,Q7 @vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 194 VCLT.U8 Q9,Q6,Q7 @vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 200 VADD.U8 Q7,Q1,Q10 @edge_idx = vaddq_s8(const_2, sign_left) 202 VADD.U8 Q7,Q7,Q11 @edge_idx = vaddq_s8(edge_idx, sign_right [all...] |
ihevc_sao_edge_offset_class2.s | 247 VCGT.U8 Q7,Q6,Q5 @vcgtq_u8(pu1_cur_row, pu1_top_row) 255 VSUB.U8 Q7,Q8,Q7 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 293 VADD.I8 Q12,Q0,Q7 @I edge_idx = vaddq_s8(const_2, sign_up) 304 VNEG.S8 Q7,Q5 @I sign_up = vnegq_s8(sign_down) 306 VEXT.8 Q7,Q7,Q7,#15 @I sign_up = vextq_s8(sign_up, sign_up, 15) 377 VADD.I8 Q11,Q0,Q7 @II edge_idx = vaddq_s8(const_2, sign_up) 382 VNEG.S8 Q7,Q12 @II sign_up = vnegq_s8(sign_down [all...] |
ihevc_sao_edge_offset_class2_chroma.s | 339 VCGT.U8 Q7,Q6,Q5 @vcgtq_u8(pu1_cur_row, pu1_top_row) 355 VSUB.U8 Q7,Q8,Q7 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 406 VADD.I8 Q9,Q0,Q7 @I edge_idx = vaddq_s8(const_2, sign_up) 410 VNEG.S8 Q7,Q11 @I sign_up = vnegq_s8(sign_down) 413 VEXT.8 Q7,Q7,Q7,#14 @I sign_up = vextq_s8(sign_up, sign_up, 14) 499 VADD.I8 Q13,Q0,Q7 @II edge_idx = vaddq_s8(const_2, sign_up) 509 VNEG.S8 Q7,Q12 @II sign_up = vnegq_s8(sign_down [all...] |
ihevc_sao_edge_offset_class3.s | 262 VCGT.U8 Q7,Q6,Q5 @vcgtq_u8(pu1_cur_row, pu1_top_row) 275 VSUB.U8 Q7,Q8,Q7 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 311 VADD.I8 Q9,Q0,Q7 @I edge_idx = vaddq_s8(const_2, sign_up) 314 VNEG.S8 Q7,Q5 @I sign_up = vnegq_s8(sign_down) 316 VEXT.8 Q7,Q7,Q7,#1 @I sign_up = vextq_s8(sign_up, sign_up, 1) 400 VADD.I8 Q13,Q0,Q7 @II edge_idx = vaddq_s8(const_2, sign_up) 402 VNEG.S8 Q7,Q12 @II sign_up = vnegq_s8(sign_down [all...] |
ihevc_sao_edge_offset_class3_chroma.s | 329 VCGT.U8 Q7,Q6,Q5 @vcgtq_u8(pu1_cur_row, pu1_top_row) 337 VSUB.U8 Q7,Q8,Q7 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 398 VADD.I8 Q9,Q0,Q7 @I edge_idx = vaddq_s8(const_2, sign_up) 401 VNEG.S8 Q7,Q11 @I sign_up = vnegq_s8(sign_down) 404 VEXT.8 Q7,Q7,Q7,#2 @I sign_up = vextq_s8(sign_up, sign_up, 2) 502 VADD.I8 Q13,Q0,Q7 @II edge_idx = vaddq_s8(const_2, sign_up) 509 VNEG.S8 Q7,Q12 @II sign_up = vnegq_s8(sign_down [all...] |
ihevc_sao_edge_offset_class1.s | 135 VCLT.U8 Q7,Q5,Q4 @vcltq_u8(pu1_cur_row, pu1_top_row) 137 VSUB.U8 Q8,Q7,Q6 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 152 VCLT.U8 Q7,Q5,Q9 @vcltq_u8(pu1_cur_row, pu1_top_row) 155 VSUB.U8 Q10,Q7,Q6 @sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 227 VCLT.U8 Q7,Q5,Q9 @vcltq_u8(pu1_cur_row, pu1_top_row) 228 VSUB.U8 Q10,Q7,Q6 @sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 278 VCLT.U8 Q7,Q5,Q4 @vcltq_u8(pu1_cur_row, pu1_top_row) 279 VSUB.U8 Q8,Q7,Q6 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 294 VCLT.U8 Q7,Q5,Q9 @vcltq_u8(pu1_cur_row, pu1_next_row) 297 VSUB.U8 Q10,Q7,Q6 @sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt) [all...] |
ihevc_sao_edge_offset_class1_chroma.s | 140 VCLT.U8 Q7,Q5,Q14 @vcltq_u8(pu1_cur_row, pu1_top_row) 142 VSUB.U8 Q8,Q7,Q6 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 157 VCLT.U8 Q7,Q5,Q9 @vcltq_u8(pu1_cur_row, pu1_top_row) 160 VSUB.U8 Q10,Q7,Q6 @sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 240 VCLT.U8 Q7,Q5,Q9 @vcltq_u8(pu1_cur_row, pu1_top_row) 241 VSUB.U8 Q10,Q7,Q6 @sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 296 VCLT.U8 Q7,Q5,Q14 @vcltq_u8(pu1_cur_row, pu1_top_row) 297 VSUB.U8 Q8,Q7,Q6 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 312 VCLT.U8 Q7,Q5,Q9 @vcltq_u8(pu1_cur_row, pu1_next_row) 315 VSUB.U8 Q10,Q7,Q6 @sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt) [all...] |
/frameworks/av/media/libstagefright/codecs/aacenc/src/asm/ARMV7/ |
Radix4FFT_v7.s | 97 VQDMULH.S32 Q11, Q7, Q15 @ MULHIGH(sinx, t1) 98 VQDMULH.S32 Q12, Q7, Q14 @ MULHIGH(cosx, t1) 102 VSUB.S32 Q7, Q12, Q13 @ MULHIGH(cosx, t1) - MULHIGH(sinx, t0) 105 VSUB.S32 Q5, Q7, Q9 @ r5 = r7 - t1@ 107 VADD.S32 Q7, Q7, Q9 @ r7 = r7 + t1@ 115 VSUB.S32 Q11, Q3, Q7 @ xptr[1] = r3 - r7@ 125 VADD.S32 Q11, Q3, Q7 @ xptr[1] = r3 + r7@
|
PrePostMDCT_v7.s | 47 VREV64.32 Q8, Q7 110 VREV64.32 Q9, Q7
|
/bionic/libm/upstream-freebsd/lib/msun/bsdsrc/ |
b_tgamma.c | 109 #define Q7 9.35021023573788935372153030556e-05 254 q = Q0 +z*(Q1+z*(Q2+z*(Q3+z*(Q4+z*(Q5+z*(Q6+z*(Q7+z*Q8)))))));
|
/external/chromium_org/third_party/openmax_dl/dl/sp/src/arm/neon/ |
armSP_FFT_CToC_SC16_Radix4_ls_unsafe_s.S | 134 #define qY3 Q7.S16 143 #define qT2 Q7.S32
|
armSP_FFT_CToC_SC16_Radix4_unsafe_s.S | 109 #define qT3 Q7.S32 120 #define qY1 Q7.S16
|
armSP_FFT_CToC_SC32_Radix4_ls_unsafe_s.S | 111 #define qT0 Q7.S64 132 #define qZ0 Q7.S32
|
armSP_FFT_CToC_SC32_Radix4_unsafe_s.S | 111 #define qT3 Q7.S64 123 #define qY1 Q7.S32
|
omxSP_FFTInv_CCSToR_S16_Sfs_s.S | 105 #define qT1 Q7.S64 107 #define qT1s Q7.S16
|
armSP_FFTInv_CCSToR_S32_preTwiddleRadix2_unsafe_s.S | 113 #define qT1 Q7.S64
|
armSP_FFT_CToC_FC32_Radix4_fs_unsafe_s.S | 94 #define qY3 Q7.F32
|
/frameworks/av/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p10/src/ |
omxVCM4P10_PredictIntraChroma_8x8_s.s | 135 qLeft7minus0 QN Q7.S16 157 qC5 QN Q7.S16 165 qSum5 QN Q7.S16
|