/external/libvpx/libvpx/vp8/common/arm/neon/ |
idct_dequant_full_2x_neon.c | 21 int16x8_t q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; local 88 q8 = vqdmulhq_n_s16(q4, cospi8sqrt2minus1); 94 q8 = vshrq_n_s16(q8, 1); 97 q4 = vqaddq_s16(q4, q8); 116 q8 = vqdmulhq_n_s16(q2tmp2.val[1], sinpi8sqrt2); 130 q8 = vqsubq_s16(q8, q11); 134 q5 = vqaddq_s16(q3, q8); 135 q6 = vqsubq_s16(q3, q8); [all...] |
mbloopfilter_neon.c | 22 uint8x16_t q8, // q1 41 q14u8 = vabdq_u8(q8, q7); 42 q1u8 = vabdq_u8(q9, q8); 59 q1u8 = vabdq_u8(q5, q8); 66 q8 = veorq_u8(q8, q0u8); 83 q1s8 = vqsubq_s8(vreinterpretq_s8_u8(q5), vreinterpretq_s8_u8(q8)); 138 q13s8 = vqsubq_s8(vreinterpretq_s8_u8(q8), q12s8); 158 uint8x16_t q5, q6, q7, q8, q9, q10; local 176 q8 = vld1q_u8(src) 206 uint8x16_t q5, q6, q7, q8, q9, q10; local 293 uint8x16_t q5, q6, q7, q8, q9, q10; local 457 uint8x16_t q5, q6, q7, q8, q9, q10; local [all...] |
vp8_loopfilter_neon.c | 23 uint8x16_t q8, // q1 39 q14u8 = vabdq_u8(q8, q7); 40 q3 = vabdq_u8(q9, q8); 55 q2u8 = vabdq_u8(q5, q8); 63 q8 = veorq_u8(q8, q10); 80 q1s8 = vqsubq_s8(vreinterpretq_s8_u8(q5), vreinterpretq_s8_u8(q8)); 115 q12s8 = vqsubq_s8(vreinterpretq_s8_u8(q8), q1s8); 130 uint8x16_t q5, q6, q7, q8, q9, q10; local 147 q8 = vld1q_u8(src) 173 uint8x16_t q5, q6, q7, q8, q9, q10; local 321 uint8x16_t q5, q6, q7, q8, q9, q10; local 433 uint8x16_t q5, q6, q7, q8, q9, q10; local [all...] |
/toolchain/binutils/binutils-2.27/gas/testsuite/gas/arm/ |
armv8-2-fp16-simd-thumb.d | 18 18: ff30 0dec vabd.f16 q0, q8, q14 19 1c: ef10 0fec vmax.f16 q0, q8, q14 20 20: ef30 0fec vmin.f16 q0, q8, q14 24 30: ffb5 0760 vabs.f16 q0, q8 26 38: ffb5 07e0 vneg.f16 q0, q8 49 94: ff10 0efc vacge.f16 q0, q8, q14 50 98: ff30 0efc vacgt.f16 q0, q8, q14 51 9c: ff3c 0ef0 vacgt.f16 q0, q14, q8 52 a0: ff1c 0ef0 vacge.f16 q0, q14, q8 53 a4: ef10 0eec vceq.f16 q0, q8, q1 [all...] |
neon-omit.d | 14 0[0-9a-f]+ <[^>]+> f20ee170 vand q7, q7, q8 15 0[0-9a-f]+ <[^>]+> f30ee170 veor q7, q7, q8 38 0[0-9a-f]+ <[^>]+> f320eede vacgt\.f32 q7, q8, q7 39 0[0-9a-f]+ <[^>]+> f32ee370 vcge\.u32 q7, q7, q8 40 0[0-9a-f]+ <[^>]+> f32ee360 vcgt\.u32 q7, q7, q8 41 0[0-9a-f]+ <[^>]+> f320e3de vcge\.u32 q7, q8, q7 42 0[0-9a-f]+ <[^>]+> f320e3ce vcgt\.u32 q7, q8, q7 57 0[0-9a-f]+ <[^>]+> f220a2ce vhsub\.s32 q5, q8, q7 60 0[0-9a-f]+ <[^>]+> f200e1dc vand q7, q8, q6 61 0[0-9a-f]+ <[^>]+> f300e1dc veor q7, q8, q [all...] |
/bionic/libc/arch-arm/kryo/bionic/ |
memcpy.S | 61 vld1.32 {q8, q9}, [r1]! 66 vst1.32 {q8, q9}, [r0]! 73 vld1.32 {q8, q9}, [r1]! 78 vst1.32 {q8, q9}, [r0]! 99 vld1.32 {q8}, [r1]! 100 vst1.32 {q8}, [r0]!
|
/external/boringssl/ios-arm/crypto/fipsmodule/ |
ghashv8-armx32.S | 20 vext.8 q8,q10,q11,#8 @ t0=0xc2....01 23 vand q10,q10,q8 26 vand q8,q8,q9 28 veor q12,q3,q8 @ twisted H 32 vext.8 q8,q12,q12,#8 @ Karatsuba pre-processing 34 veor q8,q8,q12 36 .byte 0xa0,0x2e,0xa0,0xf2 @ pmull q1,q8,q8 [all...] |
aesv8-armx32.S | 111 vtbl.8 d20,{q8},d4 112 vtbl.8 d21,{q8},d5 125 veor q9,q9,q8 127 vext.8 q8,q0,q8,#12 129 veor q8,q8,q9 131 veor q8,q8,q10 141 vld1.8 {q8},[r0 [all...] |
/external/boringssl/linux-arm/crypto/fipsmodule/ |
ghashv8-armx32.S | 19 vext.8 q8,q10,q11,#8 @ t0=0xc2....01 22 vand q10,q10,q8 25 vand q8,q8,q9 27 veor q12,q3,q8 @ twisted H 31 vext.8 q8,q12,q12,#8 @ Karatsuba pre-processing 33 veor q8,q8,q12 35 .byte 0xa0,0x2e,0xa0,0xf2 @ pmull q1,q8,q8 [all...] |
/external/capstone/suite/MC/ARM/ |
neont2-dup-encoding.s.cs | 6 0xa0,0xee,0xb0,0x5b = vdup.16 q8, r5 13 0xf4,0xff,0x48,0x0c = vdup.32 q8, d8[0] 19 0xfc,0xff,0x48,0x0c = vdup.32 q8, d8[1]
|
neon-minmax-encoding.s.cs | 18 0xe2,0xe6,0x20,0xf2 = vmax.s32 q7, q8, q9 package 21 0x60,0xc6,0x2e,0xf3 = vmax.u32 q6, q7, q8 25 0xe2,0x06,0x60,0xf2 = vmax.s32 q8, q8, q9 package 28 0x60,0xe6,0x2e,0xf3 = vmax.u32 q7, q7, q8 46 0xf2,0xe6,0x20,0xf2 = vmin.s32 q7, q8, q9 package 49 0x70,0xc6,0x2e,0xf3 = vmin.u32 q6, q7, q8 53 0xf2,0x06,0x60,0xf2 = vmin.s32 q8, q8, q9 package 56 0x70,0xe6,0x2e,0xf3 = vmin.u32 q7, q7, q8 [all...] |
neont2-minmax-encoding.s.cs | 18 0x20,0xef,0xe2,0xe6 = vmax.s32 q7, q8, q9 package 21 0x2e,0xff,0x60,0xc6 = vmax.u32 q6, q7, q8 25 0x60,0xef,0xe2,0x06 = vmax.s32 q8, q8, q9 package 28 0x2e,0xff,0x60,0xe6 = vmax.u32 q7, q7, q8 46 0x20,0xef,0xf2,0xe6 = vmin.s32 q7, q8, q9 package 49 0x2e,0xff,0x70,0xc6 = vmin.u32 q6, q7, q8 53 0x60,0xef,0xf2,0x06 = vmin.s32 q8, q8, q9 package 56 0x2e,0xff,0x70,0xe6 = vmin.u32 q7, q7, q8 [all...] |
/external/libhevc/common/arm/ |
ihevc_deblk_chroma_horz.s | 103 vmovl.u8 q8,d16 114 vsub.i16 q3,q2,q8 128 vdup.16 q8,r2 134 vzip.16 q2,q8 139 vmin.s16 q8,q3,q2 140 vmax.s16 q2,q9,q8
|
ihevc_deblk_luma_vert.s | 109 vmov.i16 q8,#0x2 231 vmla.i16 q10,q0,q8 282 vmla.i16 q13,q0,q8 404 vaddw.u8 q8,q0,d6 406 vrshrn.i16 d2,q8,#2 437 vmov.i16 q8,#0x3 451 vmul.i16 q8,q0,q8 452 vsub.i16 q8,q9,q8 [all...] |
/external/boringssl/src/crypto/curve25519/asm/ |
x25519-asm-arm.S | 102 vshr.u64 q8,q8,#6 109 vand q8,q8,q2 150 vadd.i64 q8,q8,q12 152 vadd.i64 q15,q8,q1 167 vsub.i64 q7,q8,q12 172 vadd.i64 q8,q6,q1 177 vshr.s64 q8,q8,#2 [all...] |
/external/llvm/test/MC/ARM/ |
neon-minmax-encoding.s | 21 vmax.s32 q7, q8, q9 24 vmax.u32 q6, q7, q8 29 vmax.s32 q8, q9 32 vmax.u32 q7, q8 51 @ CHECK: vmax.s32 q7, q8, q9 @ encoding: [0xe2,0xe6,0x20,0xf2] 54 @ CHECK: vmax.u32 q6, q7, q8 @ encoding: [0x60,0xc6,0x2e,0xf3] 58 @ CHECK: vmax.s32 q8, q8, q9 @ encoding: [0xe2,0x06,0x60,0xf2] 61 @ CHECK: vmax.u32 q7, q7, q8 @ encoding: [0x60,0xe6,0x2e,0xf3] 83 vmin.s32 q7, q8, q [all...] |
neont2-minmax-encoding.s | 23 vmax.s32 q7, q8, q9 26 vmax.u32 q6, q7, q8 31 vmax.s32 q8, q9 34 vmax.u32 q7, q8 53 @ CHECK: vmax.s32 q7, q8, q9 @ encoding: [0x20,0xef,0xe2,0xe6] 56 @ CHECK: vmax.u32 q6, q7, q8 @ encoding: [0x2e,0xff,0x60,0xc6] 60 @ CHECK: vmax.s32 q8, q8, q9 @ encoding: [0x60,0xef,0xe2,0x06] 63 @ CHECK: vmax.u32 q7, q7, q8 @ encoding: [0x2e,0xff,0x60,0xe6] 85 vmin.s32 q7, q8, q [all...] |
neon-shiftaccum-encoding.s | 10 vsra.s64 q8, q4, #64 45 @ CHECK: vsra.s64 q8, q4, #64 @ encoding: [0xd8,0x01,0xc0,0xf2] 86 vrsra.u32 q7, q8, #32 87 vrsra.u64 q8, q9, #64 104 vrsra.u32 q8, #32 121 @ CHECK: vrsra.u32 q7, q8, #32 @ encoding: [0x70,0xe3,0xa0,0xf3] 122 @ CHECK: vrsra.u64 q8, q9, #64 @ encoding: [0xf2,0x03,0xc0,0xf3] 138 @ CHECK: vrsra.u32 q8, q8, #32 @ encoding: [0x70,0x03,0xe0,0xf3] 146 vsli.8 q1, q8, # [all...] |
neont2-shiftaccum-encoding.s | 12 vsra.s64 q8, q4, #64 47 @ CHECK: vsra.s64 q8, q4, #64 @ encoding: [0xc0,0xef,0xd8,0x01] 89 vrsra.u32 q7, q8, #32 90 vrsra.u64 q8, q9, #64 107 vrsra.u32 q8, #32 124 @ CHECK: vrsra.u32 q7, q8, #32 @ encoding: [0xa0,0xff,0x70,0xe3] 125 @ CHECK: vrsra.u64 q8, q9, #64 @ encoding: [0xc0,0xff,0xf2,0x03] 141 @ CHECK: vrsra.u32 q8, q8, #32 @ encoding: [0xe0,0xff,0x70,0x03] 149 vsli.8 q1, q8, # [all...] |
/frameworks/rs/cpu_ref/ |
rsCpuIntrinsics_neon_Convolve.S | 69 vmull.s16 q8, d4, d0[0] 70 vmlal.s16 q8, d5, d0[1] 71 vmlal.s16 q8, d6, d0[2] 72 vmlal.s16 q8, d8, d0[3] 73 vmlal.s16 q8, d9, d1[0] 74 vmlal.s16 q8, d10, d1[1] 75 vmlal.s16 q8, d12, d1[2] 76 vmlal.s16 q8, d13, d1[3] 77 vmlal.s16 q8, d14, d2[0] 89 vshrn.i32 d16, q8, # [all...] |
rsCpuIntrinsics_neon_Resize.S | 59 vmovl.u8 q8, d16 87 vmovl.u8 q8, d16 447 vst1.u8 {q8}, [r0]! 449 vst1.u8 {q8,q9}, [r0]! 455 3: vshr.u16 q8, q6, #1 457 vqrdmulh.s16 q9, q8, q8 459 vqrdmulh.s16 q10, q9, q8 468 vhadd.s16 q0, q10, q8 553 vmull.s16 q8, d24, d [all...] |
/external/libavc/common/arm/ |
ih264_inter_pred_luma_horz_qpel_vert_hpel_a9q.s | 331 vaddl.u8 q8, d2, d8 337 vmls.u16 q6, q8, q12 339 vaddl.u8 q8, d6, d8 345 vmla.u16 q14, q8, q13 349 vext.16 q8, q6, q7, #2 354 vadd.s16 q8, q8, q9 365 vaddl.u8 q8, d5, d11 367 vmls.u16 q7, q8, q12 370 vaddl.u8 q8, d8, d1 [all...] |
ih264_inter_pred_filters_luma_vert_a9q.s | 130 vaddl.u8 q8, d2, d8 @ temp2 = src[1_0] + src[4_0] 138 vmls.u16 q7, q8, q12 @ temp -= temp2 * 5 139 vaddl.u8 q8, d2, d0 141 vmla.u16 q8, q6, q11 149 vmls.u16 q8, q9, q12 158 vqrshrun.s16 d30, q8, #5 160 vaddl.u8 q8, d5, d3 162 vmla.u16 q8, q6, q11 171 vmls.u16 q8, q13, q12 180 vqrshrun.s16 d31, q8, # [all...] |
/external/valgrind/none/tests/arm/ |
v8crypto_a.stdout.exp | [all...] |
v8crypto_t.stdout.exp | [all...] |