HomeSort by relevance Sort by last modified time
    Searched refs:d24 (Results 51 - 75 of 168) sorted by null

1 23 4 5 6 7

  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/arm/neon/
sixtappredict8x8_neon.asm 65 vdup.8 d0, d24[0] ;first_pass filter (d0-d5)
66 vdup.8 d1, d24[4]
149 vqrshrun.s16 d24, q9, #7
156 vst1.u8 {d24}, [lr]!
284 vmlsl.u8 q5, d24, d4
293 vmlal.u8 q4, d24, d5
300 vmull.u8 q10, d24, d3
339 vdup.8 d0, d24[0] ;first_pass filter (d0-d5)
340 vdup.8 d1, d24[4]
421 vqrshrun.s16 d24, q9, #
    [all...]
sixtappredict4x4_neon.asm 61 vdup.8 d0, d24[0] ;first_pass filter (d0-d5)
63 vdup.8 d1, d24[4]
220 vext.8 d24, d28, d29, #4
241 vmlsl.u8 q6, d24, d1
246 vmlal.u8 q5, d24, d3 ;(src_ptr[1] * vp8_filter[3])
276 vdup.8 d0, d24[0] ;first_pass filter (d0-d5)
278 vdup.8 d1, d24[4]
379 vext.8 d24, d28, d29, #4
393 vmlsl.u8 q6, d24, d1
398 vmlal.u8 q5, d24, d3 ;(src_ptr[1] * vp8_filter[3]
    [all...]
vp8_subpixelvariance16x16_neon.asm 195 vld1.u8 {d24, d25}, [lr]!
200 vmull.u8 q3, d24, d0
209 vmlal.u8 q1, d24, d1 ;(src_ptr[pixel_step] * Filter[1])
322 vld1.u8 {d24, d25}, [r0], r1
327 vmull.u8 q3, d24, d0
336 vmlal.u8 q1, d24, d1 ;(src_ptr[pixel_step] * Filter[1])
393 vmlal.s16 q9, d24, d24
  /external/libhevc/common/arm/
ihevc_intra_pred_filters_luma_mode_11_to_17.s 327 vrshrn.i16 d24, q12, #5 @round shft (row 0)
337 vst1.8 d24, [r2], r3 @st (row 0)
371 vrshrn.i16 d24, q12, #5 @round shft (row 4)
381 vst1.8 d24, [r2], r3 @st (row 4)
449 vst1.8 d24, [r5], r3 @st (row 4)
450 vrshrn.i16 d24, q11, #5 @round shft (row 5)
470 vst1.8 d24, [r5], r3 @(from previous loop)st (row 5)
496 vrshrn.i16 d24, q12, #5 @round shft (row 0)
512 vst1.8 d24, [r2], r3 @st (row 0)
555 vrshrn.i16 d24, q12, #5 @round shft (row 4
    [all...]
ihevc_deblk_luma_horz.s 112 vdup.32 d24,r10 @ -2 value
207 vaddl.u8 q5,d24,d28
267 vaddl.u8 q3,d25,d24
279 vqadd.u8 d16,d24,d1
283 vqsub.u8 d17,d24,d1
413 @ vdup.32 d24,r11 @ -2 value
435 vsubl.u8 q6,d27,d24
478 vmov d18,d24
491 vsubl.u8 q7,d14,d24
498 vmovl.u8 q8,d24
    [all...]
ihevc_intra_pred_chroma_mode_3_to_9.s 214 vrshrn.i16 d24, q12, #5 @round shft (row 0)
224 vst1.8 d24, [r2], r3 @st (row 0)
260 vrshrn.i16 d24, q12, #5 @round shft (row 4)
270 vst1.8 d24, [r2], r3 @st (row 4)
347 vst1.8 d24, [r5], r3 @st (row 4)
394 vrshrn.i16 d24, q12, #5 @round shft (row 0)
409 vst1.8 d24, [r2], r3 @st (row 0)
453 vrshrn.i16 d24, q12, #5 @round shft (row 4)
474 vst1.8 d24, [r5], r3 @st (row 4)
475 vrshrn.i16 d24, q11, #5 @round shft (row 5
    [all...]
ihevc_intra_pred_filters_chroma_mode_11_to_17.s 328 vrshrn.i16 d24, q12, #5 @round shft (row 0)
338 vst1.8 d24, [r2], r3 @st (row 0)
374 vrshrn.i16 d24, q12, #5 @round shft (row 4)
384 vst1.8 d24, [r2], r3 @st (row 4)
458 vst1.8 d24, [r5], r3 @st (row 4)
459 vrshrn.i16 d24, q11, #5 @round shft (row 5)
487 vst1.8 d24, [r5], r3 @(from previous loop)st (row 5)
514 vrshrn.i16 d24, q12, #5 @round shft (row 0)
530 vst1.8 d24, [r2], r3 @st (row 0)
575 vrshrn.i16 d24, q12, #5 @round shft (row 4
    [all...]
ihevc_inter_pred_chroma_vert.s 287 vqrshrun.s16 d24,q12,#6
294 vst1.8 {d24},[r7],r3 @stores the loaded value
348 vqrshrun.s16 d24,q12,#6
355 vst1.8 {d24},[r7],r3 @stores the loaded value
377 vqrshrun.s16 d24,q12,#6
378 vst1.8 {d24},[r7],r3 @stores the loaded value
ihevc_inter_pred_chroma_vert_w16inp_w16out.s 237 vqshrn.s32 d24,q12,#6 @right shift
254 vst1.32 {d24},[r9] @stores the loaded value
287 vqshrn.s32 d24,q12,#6 @right shift
306 vst1.32 {d24},[r9] @stores the loaded value
319 vqshrn.s32 d24,q12,#6 @right shift
322 vst1.32 {d24},[r9] @stores the loaded value
ihevc_inter_pred_filters_luma_horz.s 138 vdup.8 d24,d2[0] @coeffabs_0 = vdup_lane_u8(coeffabs, 0)
220 vmlsl.u8 q4,d0,d24 @mul_res = vmlsl_u8(src[0_0], coeffabs_0)@
244 vmlsl.u8 q5,d12,d24 @mul_res = vmlsl_u8(src[0_0], coeffabs_0)@
300 vmlsl.u8 q4,d0,d24 @mul_res = vmlsl_u8(src[0_0], coeffabs_0)@
316 vmlsl.u8 q10,d1,d24
349 vmlsl.u8 q5,d0,d24 @mul_res = vmlsl_u8(src[0_0], coeffabs_0)@
368 vmlsl.u8 q11,d1,d24
400 vmlsl.u8 q4,d0,d24 @mul_res = vmlsl_u8(src[0_0], coeffabs_0)@
508 vmlsl.u8 q4,d0,d24
ihevc_deblk_luma_vert.s 113 vld1.8 {d24},[r7],r1
122 vtrn.8 d24,d1
129 vtrn.16 d24,d2
141 vdup.32 d4,d24[1]
396 vdup.32 d7,d24[0]
432 vmov.i8 d24,#0x1
490 vaddw.u8 q8,q8,d24
549 vaddw.u8 q1,q1,d24
ihevc_intra_pred_chroma_mode_27_to_33.s 215 vsub.u8 d24,d1,d25 @(iv)32-fract(dup_const_32_fract)
218 vmull.u8 q11,d20,d24 @(iv)vmull_u8(ref_main_idx, dup_const_32_fract)
292 vsub.u8 d24,d1,d25 @(viii)32-fract(dup_const_32_fract)
297 vmull.u8 q11,d20,d24 @(viii)vmull_u8(ref_main_idx, dup_const_32_fract)
356 vsub.u8 d24,d1,d25 @(iv)32-fract(dup_const_32_fract)
360 vmull.u8 q11,d20,d24 @(iv)vmull_u8(ref_main_idx, dup_const_32_fract)
435 vsub.u8 d24,d1,d25 @(viii)32-fract(dup_const_32_fract)
436 vmull.u8 q11,d20,d24 @(viii)vmull_u8(ref_main_idx, dup_const_32_fract)
ihevc_intra_pred_filters_chroma_mode_19_to_25.s 323 vsub.u8 d24,d1,d25 @(iv)32-fract(dup_const_32_fract)
325 vmull.u8 q11,d20,d24 @(iv)vmull_u8(ref_main_idx, dup_const_32_fract)
401 vsub.u8 d24,d1,d25 @(viii)32-fract(dup_const_32_fract)
407 vmull.u8 q11,d20,d24 @(viii)vmull_u8(ref_main_idx, dup_const_32_fract)
470 vsub.u8 d24,d1,d25 @(iv)32-fract(dup_const_32_fract)
473 vmull.u8 q11,d20,d24 @(iv)vmull_u8(ref_main_idx, dup_const_32_fract)
551 vsub.u8 d24,d1,d25 @(viii)32-fract(dup_const_32_fract)
552 vmull.u8 q11,d20,d24 @(viii)vmull_u8(ref_main_idx, dup_const_32_fract)
ihevc_intra_pred_filters_luma_mode_19_to_25.s 325 vsub.u8 d24,d1,d25 @(iv)32-fract(dup_const_32_fract)
327 vmull.u8 q11,d20,d24 @(iv)vmull_u8(ref_main_idx, dup_const_32_fract)
398 vsub.u8 d24,d1,d25 @(viii)32-fract(dup_const_32_fract)
404 vmull.u8 q11,d20,d24 @(viii)vmull_u8(ref_main_idx, dup_const_32_fract)
466 vsub.u8 d24,d1,d25 @(iv)32-fract(dup_const_32_fract)
469 vmull.u8 q11,d20,d24 @(iv)vmull_u8(ref_main_idx, dup_const_32_fract)
544 vsub.u8 d24,d1,d25 @(viii)32-fract(dup_const_32_fract)
545 vmull.u8 q11,d20,d24 @(viii)vmull_u8(ref_main_idx, dup_const_32_fract)
  /frameworks/av/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p10/src_gcc/
armVCM4P10_InterpolateLuma_HalfDiagHorVer4x4_unsafe_s.S 127 VMLA.I16 d24,d2,d31
137 VSUB.I16 d24,d24,d8
154 VADDL.S16 q2,d14,d24
162 VADDL.S16 q0,d18,d24
  /external/libavc/common/arm/
ih264_resi_trans_quant_a9.s 130 vld1.u8 d24, [r0], r3 @load first 8 pix src row 4
136 vsubl.u8 q3, d24, d25 @find residue row 4
174 vadd.s16 d24, d18, d19 @x5 = x0 + x1;
180 vst1.s16 d24[0], [r10] @Store the dc value to alternate dc sddress
230 vpadd.u8 d24, d22, d23 @I Pair add nnz4
234 vsub.u8 d26, d25, d24 @I invert current nnz
327 vld2.u8 {d23, d24}, [r1], r4 @load first 8 pix pred row 4
369 vadd.s16 d24, d18, d19 @x5 = x0 + x1;
375 vst1.s16 d24[0], [r10] @Store Unquantized dc value to dc alte address
424 vpadd.u8 d24, d22, d23 @I Pair add nnz
    [all...]
ih264_inter_pred_luma_horz_qpel_a9q.s 197 vext.8 d24, d5, d6, #3 @//extract a[3] (column1,row1)
203 vmlal.u8 q7, d24, d1 @// a0 + a5 + 20a2 + 20a3 (column1,row1)
233 vext.8 d24, d5, d6, #3 @//extract a[3] (column1,row1)
238 vmlal.u8 q7, d24, d1 @// a0 + a5 + 20a2 + 20a3 (column1,row1)
  /external/llvm/test/MC/ARM/
neon-shiftaccum-encoding.s 74 vrsra.s32 d7, d24, #32
92 vrsra.s32 d24, #32
109 @ CHECK: vrsra.s32 d7, d24, #32 @ encoding: [0x38,0x73,0xa0,0xf2]
126 @ CHECK: vrsra.s32 d24, d24, #32 @ encoding: [0x38,0x83,0xe0,0xf2]
152 vsri.32 d24, d13, #32
187 @ CHECK: vsri.32 d24, d13, #32 @ encoding: [0x1d,0x84,0xe0,0xf3]
neont2-shiftaccum-encoding.s 77 vrsra.s32 d7, d24, #32
95 vrsra.s32 d24, #32
112 @ CHECK: vrsra.s32 d7, d24, #32 @ encoding: [0xa0,0xef,0x38,0x73]
129 @ CHECK: vrsra.s32 d24, d24, #32 @ encoding: [0xe0,0xef,0x38,0x83]
155 vsri.32 d24, d13, #32
190 @ CHECK: vsri.32 d24, d13, #32 @ encoding: [0xe0,0xff,0x1d,0x84]
  /toolchain/binutils/binutils-2.25/gas/testsuite/gas/mn10300/
am33-2.c 40 #define d24(shift) signed_constant(24, shift, 1) macro
214 lparen, d24(0), comma, amreg (28), rparen, tick_random);
287 lparen, d24 (0), comma, amreg (28), rparen, comma, freg (24, 32));
289 lparen, amreg (28), plus, comma, d24 (0), rparen, comma,
295 freg (28, 33), comma, lparen, d24 (0), comma, amreg (24), rparen);
298 lparen, amreg (24), plus, comma, d24 (0), rparen);
392 lparen, d24 (0), comma, amreg (28), rparen, comma, dreg (24, 32));
394 lparen, amreg (28), plus, comma, d24 (0), rparen, comma,
401 lparen, d24 (0), comma, amreg (24), rparen);
404 lparen, amreg (24), plus, comma, d24 (0), rparen)
    [all...]
  /external/boringssl/src/crypto/curve25519/asm/
x25519-asm-arm.S 95 vld1.8 {d24},[r2]
251 vld1.8 {d24-d25},[r5,: 128]!
504 vld1.8 {d24},[r2,: 64]
520 vtrn.32 d24,d25
542 vmlal.s32 q8,d24,d6
552 vmlal.s32 q2,d24,d0
584 vmlal.s32 q7,d24,d21
590 vmlal.s32 q5,d24,d20
609 vmlal.s32 q4,d24,d29
611 vmlal.s32 q8,d24,d2
    [all...]
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/common/arm/neon/
vp9_loopfilter_neon.asm 458 vabd.u8 d24, d18, d17 ; m6 = abs(q3 - q2)
466 vmax.u8 d23, d23, d24 ; m3 = max(m5, m6)
472 vabd.u8 d24, d6, d7 ; m9 = abs(p0 - q0)
479 vqadd.u8 d24, d24, d24 ; b = abs(p0 - q0) * 2
492 vqadd.u8 d24, d24, d23 ; a = b + a
497 vcge.u8 d24, d0, d24 ; a > blimi
    [all...]
  /frameworks/rs/cpu_ref/
rsCpuIntrinsics_neon_3DLUT.S 64 vshll.u16 q8, d24, #8
66 vmlsl.u16 q8, d24, \zr0
175 vshll.u16 q6, d24, #2
  /external/libmpeg2/common/arm/
icv_variance_a9.s 96 vaddl.u16 q12, d24, d25
  /system/core/debuggerd/arm64/
crashglue.S 59 fmov d24, 24.0

Completed in 3337 milliseconds

1 23 4 5 6 7