/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/arm/neon/ |
sixtappredict8x8_neon.asm | 65 vdup.8 d0, d24[0] ;first_pass filter (d0-d5) 66 vdup.8 d1, d24[4] 149 vqrshrun.s16 d24, q9, #7 156 vst1.u8 {d24}, [lr]! 284 vmlsl.u8 q5, d24, d4 293 vmlal.u8 q4, d24, d5 300 vmull.u8 q10, d24, d3 339 vdup.8 d0, d24[0] ;first_pass filter (d0-d5) 340 vdup.8 d1, d24[4] 421 vqrshrun.s16 d24, q9, # [all...] |
sixtappredict4x4_neon.asm | 61 vdup.8 d0, d24[0] ;first_pass filter (d0-d5) 63 vdup.8 d1, d24[4] 220 vext.8 d24, d28, d29, #4 241 vmlsl.u8 q6, d24, d1 246 vmlal.u8 q5, d24, d3 ;(src_ptr[1] * vp8_filter[3]) 276 vdup.8 d0, d24[0] ;first_pass filter (d0-d5) 278 vdup.8 d1, d24[4] 379 vext.8 d24, d28, d29, #4 393 vmlsl.u8 q6, d24, d1 398 vmlal.u8 q5, d24, d3 ;(src_ptr[1] * vp8_filter[3] [all...] |
vp8_subpixelvariance16x16_neon.asm | 195 vld1.u8 {d24, d25}, [lr]! 200 vmull.u8 q3, d24, d0 209 vmlal.u8 q1, d24, d1 ;(src_ptr[pixel_step] * Filter[1]) 322 vld1.u8 {d24, d25}, [r0], r1 327 vmull.u8 q3, d24, d0 336 vmlal.u8 q1, d24, d1 ;(src_ptr[pixel_step] * Filter[1]) 393 vmlal.s16 q9, d24, d24
|
/external/libhevc/common/arm/ |
ihevc_intra_pred_filters_luma_mode_11_to_17.s | 327 vrshrn.i16 d24, q12, #5 @round shft (row 0) 337 vst1.8 d24, [r2], r3 @st (row 0) 371 vrshrn.i16 d24, q12, #5 @round shft (row 4) 381 vst1.8 d24, [r2], r3 @st (row 4) 449 vst1.8 d24, [r5], r3 @st (row 4) 450 vrshrn.i16 d24, q11, #5 @round shft (row 5) 470 vst1.8 d24, [r5], r3 @(from previous loop)st (row 5) 496 vrshrn.i16 d24, q12, #5 @round shft (row 0) 512 vst1.8 d24, [r2], r3 @st (row 0) 555 vrshrn.i16 d24, q12, #5 @round shft (row 4 [all...] |
ihevc_deblk_luma_horz.s | 112 vdup.32 d24,r10 @ -2 value 207 vaddl.u8 q5,d24,d28 267 vaddl.u8 q3,d25,d24 279 vqadd.u8 d16,d24,d1 283 vqsub.u8 d17,d24,d1 413 @ vdup.32 d24,r11 @ -2 value 435 vsubl.u8 q6,d27,d24 478 vmov d18,d24 491 vsubl.u8 q7,d14,d24 498 vmovl.u8 q8,d24 [all...] |
ihevc_intra_pred_chroma_mode_3_to_9.s | 214 vrshrn.i16 d24, q12, #5 @round shft (row 0) 224 vst1.8 d24, [r2], r3 @st (row 0) 260 vrshrn.i16 d24, q12, #5 @round shft (row 4) 270 vst1.8 d24, [r2], r3 @st (row 4) 347 vst1.8 d24, [r5], r3 @st (row 4) 394 vrshrn.i16 d24, q12, #5 @round shft (row 0) 409 vst1.8 d24, [r2], r3 @st (row 0) 453 vrshrn.i16 d24, q12, #5 @round shft (row 4) 474 vst1.8 d24, [r5], r3 @st (row 4) 475 vrshrn.i16 d24, q11, #5 @round shft (row 5 [all...] |
ihevc_intra_pred_filters_chroma_mode_11_to_17.s | 328 vrshrn.i16 d24, q12, #5 @round shft (row 0) 338 vst1.8 d24, [r2], r3 @st (row 0) 374 vrshrn.i16 d24, q12, #5 @round shft (row 4) 384 vst1.8 d24, [r2], r3 @st (row 4) 458 vst1.8 d24, [r5], r3 @st (row 4) 459 vrshrn.i16 d24, q11, #5 @round shft (row 5) 487 vst1.8 d24, [r5], r3 @(from previous loop)st (row 5) 514 vrshrn.i16 d24, q12, #5 @round shft (row 0) 530 vst1.8 d24, [r2], r3 @st (row 0) 575 vrshrn.i16 d24, q12, #5 @round shft (row 4 [all...] |
ihevc_inter_pred_chroma_vert.s | 287 vqrshrun.s16 d24,q12,#6 294 vst1.8 {d24},[r7],r3 @stores the loaded value 348 vqrshrun.s16 d24,q12,#6 355 vst1.8 {d24},[r7],r3 @stores the loaded value 377 vqrshrun.s16 d24,q12,#6 378 vst1.8 {d24},[r7],r3 @stores the loaded value
|
ihevc_inter_pred_chroma_vert_w16inp_w16out.s | 237 vqshrn.s32 d24,q12,#6 @right shift 254 vst1.32 {d24},[r9] @stores the loaded value 287 vqshrn.s32 d24,q12,#6 @right shift 306 vst1.32 {d24},[r9] @stores the loaded value 319 vqshrn.s32 d24,q12,#6 @right shift 322 vst1.32 {d24},[r9] @stores the loaded value
|
ihevc_inter_pred_filters_luma_horz.s | 138 vdup.8 d24,d2[0] @coeffabs_0 = vdup_lane_u8(coeffabs, 0) 220 vmlsl.u8 q4,d0,d24 @mul_res = vmlsl_u8(src[0_0], coeffabs_0)@ 244 vmlsl.u8 q5,d12,d24 @mul_res = vmlsl_u8(src[0_0], coeffabs_0)@ 300 vmlsl.u8 q4,d0,d24 @mul_res = vmlsl_u8(src[0_0], coeffabs_0)@ 316 vmlsl.u8 q10,d1,d24 349 vmlsl.u8 q5,d0,d24 @mul_res = vmlsl_u8(src[0_0], coeffabs_0)@ 368 vmlsl.u8 q11,d1,d24 400 vmlsl.u8 q4,d0,d24 @mul_res = vmlsl_u8(src[0_0], coeffabs_0)@ 508 vmlsl.u8 q4,d0,d24
|
ihevc_deblk_luma_vert.s | 113 vld1.8 {d24},[r7],r1 122 vtrn.8 d24,d1 129 vtrn.16 d24,d2 141 vdup.32 d4,d24[1] 396 vdup.32 d7,d24[0] 432 vmov.i8 d24,#0x1 490 vaddw.u8 q8,q8,d24 549 vaddw.u8 q1,q1,d24
|
ihevc_intra_pred_chroma_mode_27_to_33.s | 215 vsub.u8 d24,d1,d25 @(iv)32-fract(dup_const_32_fract) 218 vmull.u8 q11,d20,d24 @(iv)vmull_u8(ref_main_idx, dup_const_32_fract) 292 vsub.u8 d24,d1,d25 @(viii)32-fract(dup_const_32_fract) 297 vmull.u8 q11,d20,d24 @(viii)vmull_u8(ref_main_idx, dup_const_32_fract) 356 vsub.u8 d24,d1,d25 @(iv)32-fract(dup_const_32_fract) 360 vmull.u8 q11,d20,d24 @(iv)vmull_u8(ref_main_idx, dup_const_32_fract) 435 vsub.u8 d24,d1,d25 @(viii)32-fract(dup_const_32_fract) 436 vmull.u8 q11,d20,d24 @(viii)vmull_u8(ref_main_idx, dup_const_32_fract)
|
ihevc_intra_pred_filters_chroma_mode_19_to_25.s | 323 vsub.u8 d24,d1,d25 @(iv)32-fract(dup_const_32_fract) 325 vmull.u8 q11,d20,d24 @(iv)vmull_u8(ref_main_idx, dup_const_32_fract) 401 vsub.u8 d24,d1,d25 @(viii)32-fract(dup_const_32_fract) 407 vmull.u8 q11,d20,d24 @(viii)vmull_u8(ref_main_idx, dup_const_32_fract) 470 vsub.u8 d24,d1,d25 @(iv)32-fract(dup_const_32_fract) 473 vmull.u8 q11,d20,d24 @(iv)vmull_u8(ref_main_idx, dup_const_32_fract) 551 vsub.u8 d24,d1,d25 @(viii)32-fract(dup_const_32_fract) 552 vmull.u8 q11,d20,d24 @(viii)vmull_u8(ref_main_idx, dup_const_32_fract)
|
ihevc_intra_pred_filters_luma_mode_19_to_25.s | 325 vsub.u8 d24,d1,d25 @(iv)32-fract(dup_const_32_fract) 327 vmull.u8 q11,d20,d24 @(iv)vmull_u8(ref_main_idx, dup_const_32_fract) 398 vsub.u8 d24,d1,d25 @(viii)32-fract(dup_const_32_fract) 404 vmull.u8 q11,d20,d24 @(viii)vmull_u8(ref_main_idx, dup_const_32_fract) 466 vsub.u8 d24,d1,d25 @(iv)32-fract(dup_const_32_fract) 469 vmull.u8 q11,d20,d24 @(iv)vmull_u8(ref_main_idx, dup_const_32_fract) 544 vsub.u8 d24,d1,d25 @(viii)32-fract(dup_const_32_fract) 545 vmull.u8 q11,d20,d24 @(viii)vmull_u8(ref_main_idx, dup_const_32_fract)
|
/frameworks/av/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p10/src_gcc/ |
armVCM4P10_InterpolateLuma_HalfDiagHorVer4x4_unsafe_s.S | 127 VMLA.I16 d24,d2,d31 137 VSUB.I16 d24,d24,d8 154 VADDL.S16 q2,d14,d24 162 VADDL.S16 q0,d18,d24
|
/external/libavc/common/arm/ |
ih264_resi_trans_quant_a9.s | 130 vld1.u8 d24, [r0], r3 @load first 8 pix src row 4 136 vsubl.u8 q3, d24, d25 @find residue row 4 174 vadd.s16 d24, d18, d19 @x5 = x0 + x1; 180 vst1.s16 d24[0], [r10] @Store the dc value to alternate dc sddress 230 vpadd.u8 d24, d22, d23 @I Pair add nnz4 234 vsub.u8 d26, d25, d24 @I invert current nnz 327 vld2.u8 {d23, d24}, [r1], r4 @load first 8 pix pred row 4 369 vadd.s16 d24, d18, d19 @x5 = x0 + x1; 375 vst1.s16 d24[0], [r10] @Store Unquantized dc value to dc alte address 424 vpadd.u8 d24, d22, d23 @I Pair add nnz [all...] |
ih264_inter_pred_luma_horz_qpel_a9q.s | 197 vext.8 d24, d5, d6, #3 @//extract a[3] (column1,row1) 203 vmlal.u8 q7, d24, d1 @// a0 + a5 + 20a2 + 20a3 (column1,row1) 233 vext.8 d24, d5, d6, #3 @//extract a[3] (column1,row1) 238 vmlal.u8 q7, d24, d1 @// a0 + a5 + 20a2 + 20a3 (column1,row1)
|
/external/llvm/test/MC/ARM/ |
neon-shiftaccum-encoding.s | 74 vrsra.s32 d7, d24, #32 92 vrsra.s32 d24, #32 109 @ CHECK: vrsra.s32 d7, d24, #32 @ encoding: [0x38,0x73,0xa0,0xf2] 126 @ CHECK: vrsra.s32 d24, d24, #32 @ encoding: [0x38,0x83,0xe0,0xf2] 152 vsri.32 d24, d13, #32 187 @ CHECK: vsri.32 d24, d13, #32 @ encoding: [0x1d,0x84,0xe0,0xf3]
|
neont2-shiftaccum-encoding.s | 77 vrsra.s32 d7, d24, #32 95 vrsra.s32 d24, #32 112 @ CHECK: vrsra.s32 d7, d24, #32 @ encoding: [0xa0,0xef,0x38,0x73] 129 @ CHECK: vrsra.s32 d24, d24, #32 @ encoding: [0xe0,0xef,0x38,0x83] 155 vsri.32 d24, d13, #32 190 @ CHECK: vsri.32 d24, d13, #32 @ encoding: [0xe0,0xff,0x1d,0x84]
|
/toolchain/binutils/binutils-2.25/gas/testsuite/gas/mn10300/ |
am33-2.c | 40 #define d24(shift) signed_constant(24, shift, 1) macro 214 lparen, d24(0), comma, amreg (28), rparen, tick_random); 287 lparen, d24 (0), comma, amreg (28), rparen, comma, freg (24, 32)); 289 lparen, amreg (28), plus, comma, d24 (0), rparen, comma, 295 freg (28, 33), comma, lparen, d24 (0), comma, amreg (24), rparen); 298 lparen, amreg (24), plus, comma, d24 (0), rparen); 392 lparen, d24 (0), comma, amreg (28), rparen, comma, dreg (24, 32)); 394 lparen, amreg (28), plus, comma, d24 (0), rparen, comma, 401 lparen, d24 (0), comma, amreg (24), rparen); 404 lparen, amreg (24), plus, comma, d24 (0), rparen) [all...] |
/external/boringssl/src/crypto/curve25519/asm/ |
x25519-asm-arm.S | 95 vld1.8 {d24},[r2] 251 vld1.8 {d24-d25},[r5,: 128]! 504 vld1.8 {d24},[r2,: 64] 520 vtrn.32 d24,d25 542 vmlal.s32 q8,d24,d6 552 vmlal.s32 q2,d24,d0 584 vmlal.s32 q7,d24,d21 590 vmlal.s32 q5,d24,d20 609 vmlal.s32 q4,d24,d29 611 vmlal.s32 q8,d24,d2 [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/common/arm/neon/ |
vp9_loopfilter_neon.asm | 458 vabd.u8 d24, d18, d17 ; m6 = abs(q3 - q2) 466 vmax.u8 d23, d23, d24 ; m3 = max(m5, m6) 472 vabd.u8 d24, d6, d7 ; m9 = abs(p0 - q0) 479 vqadd.u8 d24, d24, d24 ; b = abs(p0 - q0) * 2 492 vqadd.u8 d24, d24, d23 ; a = b + a 497 vcge.u8 d24, d0, d24 ; a > blimi [all...] |
/frameworks/rs/cpu_ref/ |
rsCpuIntrinsics_neon_3DLUT.S | 64 vshll.u16 q8, d24, #8 66 vmlsl.u16 q8, d24, \zr0 175 vshll.u16 q6, d24, #2
|
/external/libmpeg2/common/arm/ |
icv_variance_a9.s | 96 vaddl.u16 q12, d24, d25
|
/system/core/debuggerd/arm64/ |
crashglue.S | 59 fmov d24, 24.0
|