/frameworks/av/media/libstagefright/codecs/amrwbenc/src/asm/ARMV7/ |
convolve_neon.s | 46 VMOV.S32 Q10, #0 63 VMOV.S32 r5, D20[0] 84 VMOV.S32 Q10, #0 99 VMOV.S32 r5, D20[0] 122 VMOV.S32 Q10, #0 138 VMOV.S32 r5, D20[0] 148 VMOV.S32 Q10, #0 163 VMOV.S32 r5, D20[0]
|
Filt_6k_7k_neon.s | 93 VMOV.S16 D7[3], r5 @set fir_6k_7K = 0 98 VMOV.S16 Q15, #0 189 VMOV.S16 D8,D9 191 VMOV.S16 D9,D10 193 VMOV.S16 D10,D11 194 VMOV.S16 D11,D12 196 VMOV.S16 D12,D13 198 VMOV.S16 D13,D14 199 VMOV.S16 D14,D15 201 VMOV.S16 D15,D1 [all...] |
syn_filt_neon.s | 51 VMOV.S16 D8[0], r5 86 VMOV.S16 r9, D20[0]
|
/external/libhevc/decoder/arm/ |
ihevcd_fmt_conv_420sp_to_rgba8888.s | 117 VMOV.16 D0[0],R10 @//C1 120 VMOV.16 D0[1],R10 @//C2 123 VMOV.16 D0[2],R10 @//C3 126 VMOV.16 D0[3],R10 @//C4 175 @VMOV.I8 Q1,#128 229 VMOV.I8 D17,#0 239 VMOV.I8 D23,#0 280 VMOV.I8 D17,#0 290 VMOV.I8 D23,#0 312 @VMOV.I8 Q1,#12 [all...] |
/external/libhevc/common/arm/ |
ihevc_sao_edge_offset_class0_chroma.s | 82 VMOV.I8 Q1,#2 @const_2 = vdupq_n_s8(2) 86 VMOV.I16 Q2,#0 @const_min_clip = vdupq_n_s16(0) 90 VMOV.I16 Q3,#255 @const_max_clip = vdupq_n_u16((1 << bit_depth) - 1) 96 VMOV.S8 Q4,#0xFF @au1_mask = vdupq_n_s8(-1) 126 VMOV.8 D8[0],r12 @vsetq_lane_s8(pu1_avail[0], au1_mask, 0) 127 VMOV.8 D8[1],r12 @vsetq_lane_s8(pu1_avail[0], au1_mask, 1) 132 VMOV.16 D8[0],r12 @au1_mask = vsetq_lane_s8(-1, au1_mask, 0) 138 VMOV.8 D9[6],r12 @au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 14) 139 VMOV.8 D9[7],r12 @au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15) 153 VMOV.16 D15[3],r11 @vsetq_lane_u16(pu1_src_left[ht - row], pu1_cur_row_tmp, 14,15 [all...] |
ihevc_sao_edge_offset_class3_chroma.s | 275 VMOV.I8 Q0,#2 @const_2 = vdupq_n_s8(2) 276 VMOV.I16 Q1,#0 @const_min_clip = vdupq_n_s16(0) 277 VMOV.I16 Q2,#255 @const_max_clip = vdupq_n_u16((1 << bit_depth) - 1) 285 VMOV.S8 Q4,#0xFF @au1_mask = vdupq_n_s8(-1) 301 VMOV.8 D8[0],r8 @au1_mask = vsetq_lane_s8(-1, au1_mask, 0) 305 VMOV.8 D8[1],r8 @au1_mask = vsetq_lane_s8(-1, au1_mask, 0) 309 VMOV.8 D9[6],r8 @au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15) 310 VMOV.8 D9[7],r8 @au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15) 320 VMOV.I8 Q9,#0 351 VMOV.I8 Q9,#0 @ [all...] |
ihevc_sao_edge_offset_class2_chroma.s | 263 VMOV.I8 Q0,#2 @const_2 = vdupq_n_s8(2) 267 VMOV.I16 Q1,#0 @const_min_clip = vdupq_n_s16(0) 271 VMOV.I16 Q2,#255 @const_max_clip = vdupq_n_u16((1 << bit_depth) - 1) 293 VMOV.S8 Q4,#0xFF @au1_mask = vdupq_n_s8(-1) 307 VMOV.8 D8[0],r8 @au1_mask = vsetq_lane_s8(-1, au1_mask, 0) 310 VMOV.8 D8[1],r8 @au1_mask = vsetq_lane_s8(-1, au1_mask, 0) 314 VMOV.8 D9[6],r8 @au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15) 315 VMOV.8 D9[7],r8 @au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15) 364 VMOV.I8 Q9,#0 368 VMOV.16 D18[0],r5 @I pu1_next_row_tmp = vsetq_lane_u8(pu1_src_cpy[src_strd + 16], (…) [all...] |
ihevc_sao_edge_offset_class0.s | 78 VMOV.I8 Q1,#2 @const_2 = vdupq_n_s8(2) 82 VMOV.I16 Q2,#0 @const_min_clip = vdupq_n_s16(0) 86 VMOV.I16 Q3,#255 @const_max_clip = vdupq_n_u16((1 << bit_depth) - 1) 92 VMOV.S8 Q4,#0xFF @au1_mask = vdupq_n_s8(-1) 122 VMOV.8 D8[0],r12 @vsetq_lane_s8(pu1_avail[0], au1_mask, 0) 127 VMOV.8 D8[0],r12 @au1_mask = vsetq_lane_s8(-1, au1_mask, 0) 133 VMOV.8 D9[7],r12 @au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15) 147 VMOV.8 D15[7],r11 @vsetq_lane_u8(pu1_src_left[ht - row], pu1_cur_row_tmp, 15) 161 VMOV.8 D29[7],r11 @II Iteration vsetq_lane_u8(pu1_src_left[ht - row], pu1_cur_row_tmp, 15) 173 VMOV.8 D14[0],r11 @pu1_cur_row_tmp = vsetq_lane_u8(pu1_src_cpy[16], pu1_cur_row_t (…) [all...] |
ihevc_sao_edge_offset_class2.s | 181 VMOV.I8 Q0,#2 @const_2 = vdupq_n_s8(2) 185 VMOV.I16 Q1,#0 @const_min_clip = vdupq_n_s16(0) 189 VMOV.I16 Q2,#255 @const_max_clip = vdupq_n_u16((1 << bit_depth) - 1) 203 VMOV.S8 Q4,#0xFF @au1_mask = vdupq_n_s8(-1) 221 VMOV.8 d8[0],r8 @au1_mask = vsetq_lane_s8((-1||pu1_avail[0]), au1_mask, 0) 225 VMOV.8 d9[7],r8 @au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15) 265 VMOV.I8 Q9,#0 275 VMOV.8 D18[0],r5 @I pu1_next_row_tmp = vsetq_lane_u8(pu1_src_cpy[src_strd + 16], pu1_next_row_tmp, 0) 290 VMOV.8 D14[0],r4 @I sign_up = sign_up = vsetq_lane_s8(SIGN(pu1_src_cpy[0] - pu1_src_left_cpy[ht_tmp - 1 - row]), sign_up, 0) 317 VMOV Q6,Q8 @I pu1_cur_row = pu1_next_ro [all...] |
/frameworks/av/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p10/src_gcc/ |
omxVCM4P10_FilterDeblockingChroma_HorEdge_I_s.S | 40 VMOV.I8 d14,#0 41 VMOV.I8 d15,#0x1 42 VMOV.I16 d1,#0x4 59 VMOV.32 d26[0],r6
|
omxVCM4P10_TransformDequantLumaDCFromPair_s.S | 60 VMOV.I32 q3,#0x2 61 VMOV.I32 q4,#0x2 62 VMOV.I32 q5,#0x2 63 VMOV.I32 q6,#0x2
|
armVCM4P10_InterpolateLuma_HalfDiagVerHor4x4_unsafe_s.S | 33 VMOV.I8 d30,#0x5 34 VMOV.I8 d31,#0x14 82 VMOV.I16 d31,#0x14 90 VMOV.I16 d30,#0x5
|
omxVCM4P10_FilterDeblockingChroma_VerEdge_I_s.S | 39 VMOV.I8 d14,#0 40 VMOV.I8 d15,#0x1 41 VMOV.I16 d1,#0x4 73 VMOV.32 d26[0],r6
|
omxVCM4P10_FilterDeblockingLuma_HorEdge_I_s.S | 40 VMOV.I8 d14,#0 41 VMOV.I8 d15,#0x1
|
armVCM4P10_DeblockingChroma_unsafe_s.S | 32 VMOV d28,d18
|
armVCM4P10_TransformResidual4x4_s.S | 32 VMOV.I16 d4,#0
|
omxVCM4P10_DequantTransformResidualFromPairAndAdd_s.S | 75 VMOV.I16 d4,#0 117 VMOV d4,r3,r5 120 VMOV d5,r3,r5
|
/frameworks/av/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p10/src/ |
omxVCM4P10_TransformDequantChromaDCFromPair_s.s | 89 VMOV dZero, #0 138 VMOV dInvTrCoeff, c0w0, c1w0
|
omxVCM4P10_FilterDeblockingChroma_HorEdge_I_s.s | 124 VMOV dMask_0, #0 125 VMOV dMask_1, #1 126 VMOV dMask_4, #4 152 VMOV.U32 dBS3210[0], bS3210
|
omxVCM4P10_TransformDequantLumaDCFromPair_s.s | 202 VMOV qRound0,#2 ;// Set the Round Value 203 VMOV qRound1,#2 204 VMOV qRound2,#2 205 VMOV qRound3,#2
|
armVCM4P10_InterpolateLuma_HalfDiagVerHor4x4_unsafe_s.s | 135 VMOV dTCoeff5, #5 136 VMOV dTCoeff20, #20 209 VMOV dCoeff20, #20 219 VMOV dCoeff5, #5
|
omxVCM4P10_FilterDeblockingChroma_VerEdge_I_s.s | 143 VMOV dMask_0, #0 144 VMOV dMask_1, #1 145 VMOV dMask_4, #4 206 VMOV.U32 dBS3210[0], bS3210
|
/frameworks/av/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p2/src/ |
omxVCM4P2_QuantInvInter_I_s.s | 120 VMOV d2QP0,doubleQP 125 VMOV dQP10,QP ;// If QP is even then QP1=QP-1 else QP1=QP
|
/frameworks/av/media/libstagefright/codecs/on2/h264dec/source/arm_neon_asm/ |
h264bsdClearMbLayer.s | 50 VMOV qZero, #0
|
/frameworks/av/media/libstagefright/codecs/on2/h264dec/source/arm_neon_asm_gcc/ |
h264bsdClearMbLayer.S | 50 VMOV.I8 qZero, #0
|