/external/llvm/test/MC/ARM/ |
neon-dup-encoding.s | 11 vdup.8 q8, r0 12 vdup.16 q8, r0 13 vdup.32 q8, r0 15 @ CHECK: vdup.8 q8, r0 @ encoding: [0x90,0x0b,0xe0,0xee] 16 @ CHECK: vdup.16 q8, r0 @ encoding: [0xb0,0x0b,0xa0,0xee] 17 @ CHECK: vdup.32 q8, r0 @ encoding: [0x90,0x0b,0xa0,0xee] 27 vdup.8 q8, d16[1] 28 vdup.16 q8, d16[1] 29 vdup.32 q8, d16[1] 31 @ CHECK: vdup.8 q8, d16[1] @ encoding: [0x60,0x0c,0xf3,0xf3 [all...] |
neon-add-encoding.s | 14 @ CHECK: vadd.f32 q8, q8, q9 @ encoding: [0xe2,0x0d,0x40,0xf2] 15 vadd.f32 q8, q8, q9 17 @ CHECK: vaddl.s8 q8, d17, d16 @ encoding: [0xa0,0x00,0xc1,0xf2] 18 vaddl.s8 q8, d17, d16 19 @ CHECK: vaddl.s16 q8, d17, d16 @ encoding: [0xa0,0x00,0xd1,0xf2] 20 vaddl.s16 q8, d17, d16 21 @ CHECK: vaddl.s32 q8, d17, d16 @ encoding: [0xa0,0x00,0xe1,0xf2] 22 vaddl.s32 q8, d17, d1 [all...] |
neon-cmp-encoding.s | 7 vceq.i8 q8, q8, q9 8 vceq.i16 q8, q8, q9 9 vceq.i32 q8, q8, q9 10 vceq.f32 q8, q8, q9 16 @ CHECK: vceq.i8 q8, q8, q9 @ encoding: [0xf2,0x08,0x40,0xf3 [all...] |
neon-shift-encoding.s | 12 vshl.u8 q8, q9, q8 13 vshl.u16 q8, q9, q8 14 vshl.u32 q8, q9, q8 15 vshl.u64 q8, q9, q8 16 vshl.i8 q8, q8, # [all...] |
neon-mul-encoding.s | 7 vmul.i8 q8, q8, q9 8 vmul.i16 q8, q8, q9 9 vmul.i32 q8, q8, q9 10 vmul.f32 q8, q8, q9 12 vmul.p8 q8, q8, q [all...] |
neont2-mov-encoding.s | 28 vmov.i8 q8, #0x8 29 vmov.i16 q8, #0x10 30 vmov.i16 q8, #0x1000 31 vmov.i32 q8, #0x20 32 vmov.i32 q8, #0x2000 33 vmov.i32 q8, #0x200000 34 vmov.i32 q8, #0x20000000 35 vmov.i32 q8, #0x20FF 36 vmov.i32 q8, #0x20FFFF 37 vmov.i64 q8, #0xFF0000FF0000FFF [all...] |
neont2-mul-accum-encoding.s | 9 vmla.i8 q9, q8, q10 10 vmla.i16 q9, q8, q10 11 vmla.i32 q9, q8, q10 12 vmla.f32 q9, q8, q10 13 vmla.i32 q12, q8, d3[0] 19 @ CHECK: vmla.i8 q9, q8, q10 @ encoding: [0x40,0xef,0xe4,0x29] 20 @ CHECK: vmla.i16 q9, q8, q10 @ encoding: [0x50,0xef,0xe4,0x29] 21 @ CHECK: vmla.i32 q9, q8, q10 @ encoding: [0x60,0xef,0xe4,0x29] 22 @ CHECK: vmla.f32 q9, q8, q10 @ encoding: [0x40,0xef,0xf4,0x2d] 23 @ CHECK: vmla.i32 q12, q8, d3[0] @ encoding: [0xe0,0xff,0xc3,0x80 [all...] |
neon-mov-encoding.s | 27 vmov.i8 q8, #0x8 28 vmov.i16 q8, #0x10 29 vmov.i16 q8, #0x1000 30 vmov.i32 q8, #0x20 31 vmov.i32 q8, #0x2000 32 vmov.i32 q8, #0x200000 33 vmov.i32 q8, #0x20000000 34 vmov.i32 q8, #0x20FF 35 vmov.i32 q8, #0x20FFFF 36 vmov.i64 q8, #0xFF0000FF0000FFF [all...] |
neon-shuffle-encoding.s | 5 vext.8 q8, q9, q8, #3 6 vext.8 q8, q9, q8, #7 8 vext.32 q8, q9, q8, #3 9 vext.64 q8, q9, q8, #1 13 vext.8 q3, q8, #3 16 vext.32 q5, q8, # [all...] |
neon-pairwise-encoding.s | 33 @ CHECK: vpaddl.s8 q8, q8 @ encoding: [0x60,0x02,0xf0,0xf3] 34 vpaddl.s8 q8, q8 35 @ CHECK: vpaddl.s16 q8, q8 @ encoding: [0x60,0x02,0xf4,0xf3] 36 vpaddl.s16 q8, q8 37 @ CHECK: vpaddl.s32 q8, q8 @ encoding: [0x60,0x02,0xf8,0xf3 [all...] |
neont2-dup-encoding.s | 9 vdup.16 q8, r5 16 @ CHECK: vdup.16 q8, r5 @ encoding: [0xa0,0xee,0xb0,0x5b] 24 vdup.32 q8, d8[0] 30 vdup.32 q8, d8[1] 37 @ CHECK: vdup.32 q8, d8[0] @ encoding: [0xf4,0xff,0x48,0x0c] 43 @ CHECK: vdup.32 q8, d8[1] @ encoding: [0xfc,0xff,0x48,0x0c]
|
neon-bitwise-encoding.s | 5 vand q8, q8, q9 8 @ CHECK: vand q8, q8, q9 @ encoding: [0xf2,0x01,0x40,0xf2] 11 veor q8, q8, q9 14 @ CHECK: veor q8, q8, q9 @ encoding: [0xf2,0x01,0x40,0xf3] 17 vorr q8, q8, q [all...] |
neon-minmax-encoding.s | 21 vmax.s32 q7, q8, q9 24 vmax.u32 q6, q7, q8 29 vmax.s32 q8, q9 32 vmax.u32 q7, q8 51 @ CHECK: vmax.s32 q7, q8, q9 @ encoding: [0xe2,0xe6,0x20,0xf2] 54 @ CHECK: vmax.u32 q6, q7, q8 @ encoding: [0x60,0xc6,0x2e,0xf3] 58 @ CHECK: vmax.s32 q8, q8, q9 @ encoding: [0xe2,0x06,0x60,0xf2] 61 @ CHECK: vmax.u32 q7, q7, q8 @ encoding: [0x60,0xe6,0x2e,0xf3] 83 vmin.s32 q7, q8, q [all...] |
neont2-minmax-encoding.s | 23 vmax.s32 q7, q8, q9 26 vmax.u32 q6, q7, q8 31 vmax.s32 q8, q9 34 vmax.u32 q7, q8 53 @ CHECK: vmax.s32 q7, q8, q9 @ encoding: [0x20,0xef,0xe2,0xe6] 56 @ CHECK: vmax.u32 q6, q7, q8 @ encoding: [0x2e,0xff,0x60,0xc6] 60 @ CHECK: vmax.s32 q8, q8, q9 @ encoding: [0x60,0xef,0xe2,0x06] 63 @ CHECK: vmax.u32 q7, q7, q8 @ encoding: [0x2e,0xff,0x60,0xe6] 85 vmin.s32 q7, q8, q [all...] |
neon-shiftaccum-encoding.s | 10 vsra.s64 q8, q4, #64 45 @ CHECK: vsra.s64 q8, q4, #64 @ encoding: [0xd8,0x01,0xc0,0xf2] 86 vrsra.u32 q7, q8, #32 87 vrsra.u64 q8, q9, #64 104 vrsra.u32 q8, #32 121 @ CHECK: vrsra.u32 q7, q8, #32 @ encoding: [0x70,0xe3,0xa0,0xf3] 122 @ CHECK: vrsra.u64 q8, q9, #64 @ encoding: [0xf2,0x03,0xc0,0xf3] 138 @ CHECK: vrsra.u32 q8, q8, #32 @ encoding: [0x70,0x03,0xe0,0xf3] 146 vsli.8 q1, q8, # [all...] |
neont2-shiftaccum-encoding.s | 12 vsra.s64 q8, q4, #64 47 @ CHECK: vsra.s64 q8, q4, #64 @ encoding: [0xc0,0xef,0xd8,0x01] 89 vrsra.u32 q7, q8, #32 90 vrsra.u64 q8, q9, #64 107 vrsra.u32 q8, #32 124 @ CHECK: vrsra.u32 q7, q8, #32 @ encoding: [0xa0,0xff,0x70,0xe3] 125 @ CHECK: vrsra.u64 q8, q9, #64 @ encoding: [0xc0,0xff,0xf2,0x03] 141 @ CHECK: vrsra.u32 q8, q8, #32 @ encoding: [0xe0,0xff,0x70,0x03] 149 vsli.8 q1, q8, # [all...] |
/external/libvpx/libvpx/vp8/common/arm/neon/ |
variance_neon.asm | 29 vmov.i8 q8, #0 ;q8 - sum 49 vpadal.s16 q8, q11 ;calculate sum 55 vpadal.s16 q8, q12 58 vpadal.s16 q8, q13 61 vpadal.s16 q8, q14 68 vpaddl.s32 q0, q8 ;accumulate sum 102 vmov.i8 q8, #0 ;q8 - sum 119 vpadal.s16 q8, q11 ;calculate su [all...] |
vp8_subpixelvariance16x16s_neon.asm | 37 vmov.i8 q8, #0 ;q8 - sum 75 vpadal.s16 q8, q4 ;sum 81 vpadal.s16 q8, q5 84 vpadal.s16 q8, q6 87 vpadal.s16 q8, q7 91 vpadal.s16 q8, q0 ;sum 94 vpadal.s16 q8, q1 97 vpadal.s16 q8, q2 100 vpadal.s16 q8, q [all...] |
vp8_subpixelvariance8x8_neon.asm | 51 vmull.u8 q8, d6, d0 61 vmlal.u8 q8, d7, d1 69 vqrshrn.u16 d24, q8, #7 78 vmull.u8 q8, d6, d0 90 vmlal.u8 q8, d7, d1 96 vqrshrn.u16 d28, q8, #7 120 vmull.u8 q8, d29, d0 129 vmlal.u8 q8, d30, d1 138 vqrshrn.u16 d29, q8, #7 159 vmov.i8 q8, #0 ;q8 - su [all...] |
/system/core/libpixelflinger/ |
col32cb16blend_neon.S | 70 // q8 = dst red0 78 vshr.u16 q8, q10, #11 // shift dst red0 to low 5 bits 83 vmul.u16 q8, q8, q3 // multiply dst red0 by src alpha 90 vadd.u16 q8, q8, q0 // add src red to dst red0 96 vshr.u16 q8, q8, #8 // shift down red0 104 vsli.u16 q10, q8, #11 // shift & insert red0 into blue0
|
/frameworks/av/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p10/src_gcc/ |
armVCM4P10_InterpolateLuma_HalfVer4x4_unsafe_s.S | 27 VADDL.U8 q8,d8,d11 32 VADDL.U8 q8,d9,d12 37 VADDL.U8 q8,d10,d13 42 VADDL.U8 q8,d11,d14
|
/external/libvpx/libvpx/vp8/encoder/arm/neon/ |
fastquantizeb_neon.asm | 45 vld1.s16 {q8, q9}, [r5@128] ; load quant_ptr [0-15] 54 vqdmulh.s16 q4, q8 ; y = ((Round+abs(z)) * Quant) >> 16 71 vqdmulh.s16 q10, q8 ; y2 = ((Round+abs(z)) * Quant) >> 16 103 vceq.s16 q8, q8 ; set q8 to all 1 112 vtst.16 q14, q4, q8 ; now find eob 113 vtst.16 q15, q5, q8 ; non-zero element is set to all 1 124 vtst.16 q2, q10, q8 ; now find eob 125 vtst.16 q3, q11, q8 ; non-zero element is set to all [all...] |
vp8_mse16x16_neon.asm | 30 vmov.i8 q7, #0 ;q7, q8, q9, q10 - sse 31 vmov.i8 q8, #0 49 vmlal.s16 q8, d23, d23 56 vmlal.s16 q8, d27, d27 62 vadd.u32 q7, q7, q8 100 vmull.s16 q8, d24, d24 104 vadd.u32 q7, q7, q8
|
/external/pixman/pixman/ |
pixman-arm-neon-asm.S | 183 vshll.u8 q8, d19, #8 185 vsri.u16 q14, q8, #5 263 vshll.u8 q8, d19, #8 276 vsri.u16 q14, q8, #5 381 vshll.u8 q8, d19, #8 383 vsri.u16 q14, q8, #5 424 vshll.u8 q8, d1, #8 430 vsri.u16 q14, q8, #5 435 vsri.u16 q14, q8, #5 444 vshll.u8 q8, d1, # [all...] |
/external/libvpx/libvpx/vp9/common/arm/neon/ |
vp9_short_idct32x32_1_add_neon.asm | 21 vld1.8 {q8}, [$src], $stride 33 vqadd.u8 q8, q8, $diff 45 vqsub.u8 q8, q8, $diff 57 vst1.8 {q8}, [$dst], $stride
|