HomeSort by relevance Sort by last modified time
    Searched refs:VMOV (Results 1 - 25 of 62) sorted by null

1 2 3

  /frameworks/av/media/libstagefright/codecs/amrwbenc/src/asm/ARMV7/
convolve_neon.s 46 VMOV.S32 Q10, #0
63 VMOV.S32 r5, D20[0]
84 VMOV.S32 Q10, #0
99 VMOV.S32 r5, D20[0]
122 VMOV.S32 Q10, #0
138 VMOV.S32 r5, D20[0]
148 VMOV.S32 Q10, #0
163 VMOV.S32 r5, D20[0]
Filt_6k_7k_neon.s 93 VMOV.S16 D7[3], r5 @set fir_6k_7K = 0
98 VMOV.S16 Q15, #0
189 VMOV.S16 D8,D9
191 VMOV.S16 D9,D10
193 VMOV.S16 D10,D11
194 VMOV.S16 D11,D12
196 VMOV.S16 D12,D13
198 VMOV.S16 D13,D14
199 VMOV.S16 D14,D15
201 VMOV.S16 D15,D1
    [all...]
syn_filt_neon.s 51 VMOV.S16 D8[0], r5
86 VMOV.S16 r9, D20[0]
  /external/libhevc/decoder/arm/
ihevcd_fmt_conv_420sp_to_rgba8888.s 117 VMOV.16 D0[0],R10 @//C1
120 VMOV.16 D0[1],R10 @//C2
123 VMOV.16 D0[2],R10 @//C3
126 VMOV.16 D0[3],R10 @//C4
175 @VMOV.I8 Q1,#128
229 VMOV.I8 D17,#0
239 VMOV.I8 D23,#0
280 VMOV.I8 D17,#0
290 VMOV.I8 D23,#0
312 @VMOV.I8 Q1,#12
    [all...]
  /external/libhevc/common/arm/
ihevc_sao_edge_offset_class0_chroma.s 82 VMOV.I8 Q1,#2 @const_2 = vdupq_n_s8(2)
86 VMOV.I16 Q2,#0 @const_min_clip = vdupq_n_s16(0)
90 VMOV.I16 Q3,#255 @const_max_clip = vdupq_n_u16((1 << bit_depth) - 1)
96 VMOV.S8 Q4,#0xFF @au1_mask = vdupq_n_s8(-1)
126 VMOV.8 D8[0],r12 @vsetq_lane_s8(pu1_avail[0], au1_mask, 0)
127 VMOV.8 D8[1],r12 @vsetq_lane_s8(pu1_avail[0], au1_mask, 1)
132 VMOV.16 D8[0],r12 @au1_mask = vsetq_lane_s8(-1, au1_mask, 0)
138 VMOV.8 D9[6],r12 @au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 14)
139 VMOV.8 D9[7],r12 @au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15)
153 VMOV.16 D15[3],r11 @vsetq_lane_u16(pu1_src_left[ht - row], pu1_cur_row_tmp, 14,15
    [all...]
ihevc_sao_edge_offset_class3_chroma.s 275 VMOV.I8 Q0,#2 @const_2 = vdupq_n_s8(2)
276 VMOV.I16 Q1,#0 @const_min_clip = vdupq_n_s16(0)
277 VMOV.I16 Q2,#255 @const_max_clip = vdupq_n_u16((1 << bit_depth) - 1)
285 VMOV.S8 Q4,#0xFF @au1_mask = vdupq_n_s8(-1)
301 VMOV.8 D8[0],r8 @au1_mask = vsetq_lane_s8(-1, au1_mask, 0)
305 VMOV.8 D8[1],r8 @au1_mask = vsetq_lane_s8(-1, au1_mask, 0)
309 VMOV.8 D9[6],r8 @au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15)
310 VMOV.8 D9[7],r8 @au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15)
320 VMOV.I8 Q9,#0
351 VMOV.I8 Q9,#0 @
    [all...]
ihevc_sao_edge_offset_class2_chroma.s 263 VMOV.I8 Q0,#2 @const_2 = vdupq_n_s8(2)
267 VMOV.I16 Q1,#0 @const_min_clip = vdupq_n_s16(0)
271 VMOV.I16 Q2,#255 @const_max_clip = vdupq_n_u16((1 << bit_depth) - 1)
293 VMOV.S8 Q4,#0xFF @au1_mask = vdupq_n_s8(-1)
307 VMOV.8 D8[0],r8 @au1_mask = vsetq_lane_s8(-1, au1_mask, 0)
310 VMOV.8 D8[1],r8 @au1_mask = vsetq_lane_s8(-1, au1_mask, 0)
314 VMOV.8 D9[6],r8 @au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15)
315 VMOV.8 D9[7],r8 @au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15)
364 VMOV.I8 Q9,#0
368 VMOV.16 D18[0],r5 @I pu1_next_row_tmp = vsetq_lane_u8(pu1_src_cpy[src_strd + 16], (…)
    [all...]
ihevc_sao_edge_offset_class0.s 78 VMOV.I8 Q1,#2 @const_2 = vdupq_n_s8(2)
82 VMOV.I16 Q2,#0 @const_min_clip = vdupq_n_s16(0)
86 VMOV.I16 Q3,#255 @const_max_clip = vdupq_n_u16((1 << bit_depth) - 1)
92 VMOV.S8 Q4,#0xFF @au1_mask = vdupq_n_s8(-1)
122 VMOV.8 D8[0],r12 @vsetq_lane_s8(pu1_avail[0], au1_mask, 0)
127 VMOV.8 D8[0],r12 @au1_mask = vsetq_lane_s8(-1, au1_mask, 0)
133 VMOV.8 D9[7],r12 @au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15)
147 VMOV.8 D15[7],r11 @vsetq_lane_u8(pu1_src_left[ht - row], pu1_cur_row_tmp, 15)
161 VMOV.8 D29[7],r11 @II Iteration vsetq_lane_u8(pu1_src_left[ht - row], pu1_cur_row_tmp, 15)
173 VMOV.8 D14[0],r11 @pu1_cur_row_tmp = vsetq_lane_u8(pu1_src_cpy[16], pu1_cur_row_t (…)
    [all...]
ihevc_sao_edge_offset_class2.s 181 VMOV.I8 Q0,#2 @const_2 = vdupq_n_s8(2)
185 VMOV.I16 Q1,#0 @const_min_clip = vdupq_n_s16(0)
189 VMOV.I16 Q2,#255 @const_max_clip = vdupq_n_u16((1 << bit_depth) - 1)
203 VMOV.S8 Q4,#0xFF @au1_mask = vdupq_n_s8(-1)
221 VMOV.8 d8[0],r8 @au1_mask = vsetq_lane_s8((-1||pu1_avail[0]), au1_mask, 0)
225 VMOV.8 d9[7],r8 @au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15)
265 VMOV.I8 Q9,#0
275 VMOV.8 D18[0],r5 @I pu1_next_row_tmp = vsetq_lane_u8(pu1_src_cpy[src_strd + 16], pu1_next_row_tmp, 0)
290 VMOV.8 D14[0],r4 @I sign_up = sign_up = vsetq_lane_s8(SIGN(pu1_src_cpy[0] - pu1_src_left_cpy[ht_tmp - 1 - row]), sign_up, 0)
317 VMOV Q6,Q8 @I pu1_cur_row = pu1_next_ro
    [all...]
  /frameworks/av/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p10/src_gcc/
omxVCM4P10_FilterDeblockingChroma_HorEdge_I_s.S 40 VMOV.I8 d14,#0
41 VMOV.I8 d15,#0x1
42 VMOV.I16 d1,#0x4
59 VMOV.32 d26[0],r6
omxVCM4P10_TransformDequantLumaDCFromPair_s.S 60 VMOV.I32 q3,#0x2
61 VMOV.I32 q4,#0x2
62 VMOV.I32 q5,#0x2
63 VMOV.I32 q6,#0x2
armVCM4P10_InterpolateLuma_HalfDiagVerHor4x4_unsafe_s.S 33 VMOV.I8 d30,#0x5
34 VMOV.I8 d31,#0x14
82 VMOV.I16 d31,#0x14
90 VMOV.I16 d30,#0x5
omxVCM4P10_FilterDeblockingChroma_VerEdge_I_s.S 39 VMOV.I8 d14,#0
40 VMOV.I8 d15,#0x1
41 VMOV.I16 d1,#0x4
73 VMOV.32 d26[0],r6
omxVCM4P10_FilterDeblockingLuma_HorEdge_I_s.S 40 VMOV.I8 d14,#0
41 VMOV.I8 d15,#0x1
armVCM4P10_DeblockingChroma_unsafe_s.S 32 VMOV d28,d18
armVCM4P10_TransformResidual4x4_s.S 32 VMOV.I16 d4,#0
omxVCM4P10_DequantTransformResidualFromPairAndAdd_s.S 75 VMOV.I16 d4,#0
117 VMOV d4,r3,r5
120 VMOV d5,r3,r5
  /frameworks/av/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p10/src/
omxVCM4P10_TransformDequantChromaDCFromPair_s.s 89 VMOV dZero, #0
138 VMOV dInvTrCoeff, c0w0, c1w0
omxVCM4P10_FilterDeblockingChroma_HorEdge_I_s.s 124 VMOV dMask_0, #0
125 VMOV dMask_1, #1
126 VMOV dMask_4, #4
152 VMOV.U32 dBS3210[0], bS3210
omxVCM4P10_TransformDequantLumaDCFromPair_s.s 202 VMOV qRound0,#2 ;// Set the Round Value
203 VMOV qRound1,#2
204 VMOV qRound2,#2
205 VMOV qRound3,#2
armVCM4P10_InterpolateLuma_HalfDiagVerHor4x4_unsafe_s.s 135 VMOV dTCoeff5, #5
136 VMOV dTCoeff20, #20
209 VMOV dCoeff20, #20
219 VMOV dCoeff5, #5
omxVCM4P10_FilterDeblockingChroma_VerEdge_I_s.s 143 VMOV dMask_0, #0
144 VMOV dMask_1, #1
145 VMOV dMask_4, #4
206 VMOV.U32 dBS3210[0], bS3210
  /frameworks/av/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p2/src/
omxVCM4P2_QuantInvInter_I_s.s 120 VMOV d2QP0,doubleQP
125 VMOV dQP10,QP ;// If QP is even then QP1=QP-1 else QP1=QP
  /frameworks/av/media/libstagefright/codecs/on2/h264dec/source/arm_neon_asm/
h264bsdClearMbLayer.s 50 VMOV qZero, #0
  /frameworks/av/media/libstagefright/codecs/on2/h264dec/source/arm_neon_asm_gcc/
h264bsdClearMbLayer.S 50 VMOV.I8 qZero, #0

Completed in 550 milliseconds

1 2 3