/external/llvm/test/MC/ARM/ |
neont2-reciprocal-encoding.s | 15 @ CHECK: vrecps.f32 q8, q8, q9 @ encoding: [0x40,0xef,0xf2,0x0f] 16 vrecps.f32 q8, q8, q9 27 @ CHECK: vrsqrts.f32 q8, q8, q9 @ encoding: [0x60,0xef,0xf2,0x0f] 28 vrsqrts.f32 q8, q8, q9
|
neon-minmax-encoding.s | 21 vmax.s32 q7, q8, q9 25 vmax.f32 q9, q5, q1 29 vmax.s32 q8, q9 51 @ CHECK: vmax.s32 q7, q8, q9 @ encoding: [0xe2,0xe6,0x20,0xf2] 55 @ CHECK: vmax.f32 q9, q5, q1 @ encoding: [0x42,0x2f,0x4a,0xf2] 58 @ CHECK: vmax.s32 q8, q8, q9 @ encoding: [0xe2,0x06,0x60,0xf2] 83 vmin.s32 q7, q8, q9 87 vmin.f32 q9, q5, q1 91 vmin.s32 q8, q9 113 @ CHECK: vmin.s32 q7, q8, q9 @ encoding: [0xf2,0xe6,0x20,0xf2 [all...] |
neon-pairwise-encoding.s | 57 @ CHECK: vpadal.s8 q9, q8 @ encoding: [0x60,0x26,0xf0,0xf3] 58 vpadal.s8 q9, q8 59 @ CHECK: vpadal.s16 q9, q8 @ encoding: [0x60,0x26,0xf4,0xf3] 60 vpadal.s16 q9, q8 61 @ CHECK: vpadal.s32 q9, q8 @ encoding: [0x60,0x26,0xf8,0xf3] 62 vpadal.s32 q9, q8 63 @ CHECK: vpadal.u8 q9, q8 @ encoding: [0xe0,0x26,0xf0,0xf3] 64 vpadal.u8 q9, q8 65 @ CHECK: vpadal.u16 q9, q8 @ encoding: [0xe0,0x26,0xf4,0xf3] 66 vpadal.u16 q9, q [all...] |
neont2-minmax-encoding.s | 23 vmax.s32 q7, q8, q9 27 vmax.f32 q9, q5, q1 31 vmax.s32 q8, q9 53 @ CHECK: vmax.s32 q7, q8, q9 @ encoding: [0x20,0xef,0xe2,0xe6] 57 @ CHECK: vmax.f32 q9, q5, q1 @ encoding: [0x4a,0xef,0x42,0x2f] 60 @ CHECK: vmax.s32 q8, q8, q9 @ encoding: [0x60,0xef,0xe2,0x06] 85 vmin.s32 q7, q8, q9 89 vmin.f32 q9, q5, q1 93 vmin.s32 q8, q9 115 @ CHECK: vmin.s32 q7, q8, q9 @ encoding: [0x20,0xef,0xf2,0xe6 [all...] |
neon-v8.s | 45 vcvtp.u32.f32 q9, q8 46 @ CHECK: vcvtp.u32.f32 q9, q8 @ encoding: [0xe0,0x22,0xfb,0xf3] 62 vrintz.f32 q9, q4 63 @ CHECK: vrintz.f32 q9, q4 @ encoding: [0xc8,0x25,0xfa,0xf3] 80 vrintz.f32.f32 q9, q4 81 @ CHECK: vrintz.f32 q9, q4 @ encoding: [0xc8,0x25,0xfa,0xf3]
|
thumb-neon-v8.s | 45 vcvtp.u32.f32 q9, q8 46 @ CHECK: vcvtp.u32.f32 q9, q8 @ encoding: [0xfb,0xff,0xe0,0x22] 62 vrintz.f32 q9, q4 63 @ CHECK: vrintz.f32 q9, q4 @ encoding: [0xfa,0xff,0xc8,0x25] 80 vrintz.f32.f32 q9, q4 81 @ CHECK: vrintz.f32 q9, q4 @ encoding: [0xfa,0xff,0xc8,0x25]
|
neon-shift-encoding.s | 12 vshl.u8 q8, q9, q8 13 vshl.u16 q8, q9, q8 14 vshl.u32 q8, q9, q8 15 vshl.u64 q8, q9, q8 29 @ CHECK: vshl.u8 q8, q9, q8 @ encoding: [0xe2,0x04,0x40,0xf3] 30 @ CHECK: vshl.u16 q8, q9, q8 @ encoding: [0xe2,0x04,0x50,0xf3] 31 @ CHECK: vshl.u32 q8, q9, q8 @ encoding: [0xe2,0x04,0x60,0xf3] 32 @ CHECK: vshl.u64 q8, q9, q8 @ encoding: [0xe2,0x04,0x70,0xf3] 289 vrshl.s8 q8, q9, q8 290 vrshl.s16 q8, q9, q [all...] |
neon-shiftaccum-encoding.s | 9 vsra.s32 q9, q5, #32 44 @ CHECK: vsra.s32 q9, q5, #32 @ encoding: [0x5a,0x21,0xe0,0xf2] 87 vrsra.u64 q8, q9, #64 105 vrsra.u64 q9, #64 122 @ CHECK: vrsra.u64 q8, q9, #64 @ encoding: [0xf2,0x03,0xc0,0xf3] 139 @ CHECK: vrsra.u64 q9, q9, #64 @ encoding: [0xf2,0x23,0xc0,0xf3] 157 vsri.64 q9, q6, #64 192 @ CHECK: vsri.64 q9, q6, #64 @ encoding: [0xdc,0x24,0xc0,0xf3]
|
/external/libvpx/libvpx/vp9/common/arm/neon/ |
vp9_short_idct32x32_1_add_neon.asm | 22 vld1.8 {q9}, [$src], $stride 34 vqadd.u8 q9, q9, $diff 46 vqsub.u8 q9, q9, $diff 58 vst1.8 {q9}, [$dst], $stride
|
vp9_short_iht8x8_add_neon.asm | 109 vtrn.32 q9, q11 112 vtrn.16 q8, q9 159 vmull.s16 q9, d26, d3 167 vmlal.s16 q9, d22, d2 178 vqrshrn.s32 d12, q9, #14 ; >> 14 232 vadd.s16 q0, q9, q15 ; output[0] = step[0] + step[3] 235 vsub.s16 q3, q9, q15 ; output[3] = step[0] - step[3] 247 vmull.s16 q9, d28, d16 255 vmlsl.s16 q9, d26, d16 263 vqrshrn.s32 d10, q9, #14 ; >> 1 [all...] |
vp9_short_idct16x16_add_neon.asm | 29 vtrn.32 q9, q11 32 vtrn.16 q8, q9 53 vld2.s16 {q8,q9}, [r0]! 54 vld2.s16 {q9,q10}, [r0]! 128 vmull.s16 q9, d26, d3 136 vmlal.s16 q9, d22, d2 144 vqrshrn.s32 d12, q9, #14 ; >> 14 217 vadd.s16 q1, q9, q10 ; step1[1] = step2[1] + step2[2]; 218 vsub.s16 q2, q9, q10 ; step1[2] = step2[1] - step2[2]; 228 vmull.s16 q9, d28, d1 [all...] |
vp9_short_iht4x4_add_neon.asm | 48 vsub.s16 q9, q13, q14 68 vmull.s16 q9, d5, d19 ; s6 = sinpi_4_9 * x3 74 vsub.s32 q11, q11, q9 138 vtrn.32 q8, q9 153 vld1.s16 {q8,q9}, [r0]! 212 vrshr.s16 q9, q9, #4 221 vaddw.u8 q9, q9, d27 225 vqmovun.s16 d27, q9 [all...] |
vp9_short_idct32x32_add_neon.asm | 103 ; q6-q9 contain the results (out[j * 32 + 0-31]) 114 vrshr.s16 q9, q9, #6 119 vaddw.u8 q9, q9, d11 124 vqmovun.s16 d11, q9 137 ; q6-q9 contain the results (out[j * 32 + 0-31]) 148 vrshr.s16 q9, q9, #6 153 vaddw.u8 q9, q9, d1 [all...] |
vp9_avg_neon.asm | 39 vld1.8 {q8-q9}, [r6@128]! 42 vrhadd.u8 q1, q1, q9 54 vld1.8 {q8-q9}, [r6@128], r3 59 vrhadd.u8 q1, q1, q9
|
/external/pixman/pixman/ |
pixman-arm-neon-asm.S | 180 vqadd.u8 q9, q0, q11 184 vshll.u8 q9, d18, #8 186 vsri.u16 q14, q9, #11 256 vqadd.u8 q9, q0, q11 278 vshll.u8 q9, d18, #8 284 vsri.u16 q14, q9, #11 378 vqadd.u8 q9, q0, q11 382 vshll.u8 q9, d18, #8 384 vsri.u16 q14, q9, #11 426 vshll.u8 q9, d0, # [all...] |
pixman-arm-neon-asm-bilinear.S | 488 bilinear_load_dst dst_fmt, op, 1, d18, d19, q9 502 mask_fmt, op, 1, d0, d1, q0, d18, d19, q9 507 op, 1, d0, d1, q0, d18, d19, q9, \ 517 bilinear_load_dst dst_fmt, op, 2, d18, d19, q9 531 mask_fmt, op, 2, d0, d1, q0, d18, d19, q9 536 op, 2, d0, d1, q0, d18, d19, q9, \ 545 q3, q9, d4, d5, d16, d17, d18, d19 578 q3, q8, q9, q10 581 q3, q8, q9, q10, d23 917 vmull.u8 q9, d22, d2 [all...] |
/external/libvpx/libvpx/vp8/common/arm/neon/ |
vp8_subpixelvariance8x8_neon.asm | 52 vmull.u8 q9, d8, d0 62 vmlal.u8 q9, d9, d1 71 vqrshrn.u16 d25, q9, #7 79 vmull.u8 q9, d8, d0 91 vmlal.u8 q9, d9, d1 97 vqrshrn.u16 d29, q9, #7 160 vmov.i8 q9, #0 ;q9, q10 - sse 177 vmlal.s16 q9, d8, d8 ;sse 183 vmlal.s16 q9, d10, d1 [all...] |
idct_dequant_full_2x_neon.asm | 60 ; q9: 12 * cospi 64 vqdmulh.s16 q9, q5, d0[0] 78 vshr.s16 q9, q9, #1 83 vqadd.s16 q5, q5, q9 111 ; q9: 3 * sinpi : d1/temp2 115 vqdmulh.s16 q9, q7, d0[2] 132 ; q9: d1 = temp1 + temp2 134 vqadd.s16 q9, q10, q9 [all...] |
sixtappredict8x8_neon.asm | 85 vmull.u8 q9, d10, d0 95 vmlsl.u8 q9, d30, d1 105 vmlsl.u8 q9, d30, d4 115 vmlal.u8 q9, d30, d2 125 vmlal.u8 q9, d30, d5 142 vqadd.s16 q9, q5 149 vqrshrun.s16 d24, q9, #7 170 vmull.u8 q9, d8, d0 182 vmlsl.u8 q9, d28, d1 194 vmlsl.u8 q9, d28, d [all...] |
sixtappredict8x4_neon.asm | 81 vmull.u8 q9, d10, d0 91 vmlsl.u8 q9, d30, d1 101 vmlsl.u8 q9, d30, d4 111 vmlal.u8 q9, d30, d2 121 vmlal.u8 q9, d30, d5 136 vqadd.s16 q9, q5 143 vqrshrun.s16 d24, q9, #7 157 vmull.u8 q9, d8, d0 169 vmlsl.u8 q9, d28, d1 181 vmlsl.u8 q9, d28, d [all...] |
vp8_subpixelvariance16x16_neon.asm | 72 vmull.u8 q9, d5, d0 85 vmlal.u8 q9, d5, d1 103 vqrshrn.u16 d16, q9, #7 123 vmull.u8 q9, d2, d0 ;(src_ptr[0] * Filter[0]) 134 vmlal.u8 q9, d2, d1 ;(src_ptr[0] * Filter[1]) 163 vqrshrn.u16 d10, q9, #7 ;shift/round/saturate to u8 260 vmull.u8 q9, d5, d0 273 vmlal.u8 q9, d5, d1 291 vqrshrn.u16 d16, q9, #7 369 vmov.i8 q9, #0 ;q9, q10 - ss [all...] |
sixtappredict16x16_neon.asm | 90 vmull.u8 q9, d7, d0 108 vmlsl.u8 q9, d28, d1 ;-(src_ptr[-1] * vp8_filter[1]) 124 vmlsl.u8 q9, d28, d4 ;-(src_ptr[2] * vp8_filter[4]) 140 vmlal.u8 q9, d28, d5 ;(src_ptr[3] * vp8_filter[5]) 156 vmlal.u8 q9, d28, d2 ;(src_ptr[0] * vp8_filter[2]) 182 vqadd.s16 q9, q6 187 vqrshrun.s16 d7, q9, #7 259 vmull.u8 q9, d23, d3 266 vqadd.s16 q9, q5 271 vqrshrun.s16 d8, q9, # [all...] |
/frameworks/av/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p10/src_gcc/ |
armVCM4P10_InterpolateLuma_HalfDiagVerHor4x4_unsafe_s.S | 24 VADDL.U8 q9,d0,d10 30 VMLAL.U8 q9,d4,d31 34 VMLAL.U8 q9,d6,d31 38 VMLSL.U8 q9,d2,d30 42 VMLSL.U8 q9,d8,d30
|
/hardware/samsung_slsi/exynos5/libswconverter/ |
csc_interleave_memcpy_neon.s | 88 vld1.8 {q9}, [r12]! 97 vst2.8 {q8, q9}, [r10]!
|
/external/libvpx/libvpx/vp8/encoder/arm/neon/ |
vp8_memcpy_neon.asm | 40 vld1.8 {q8, q9}, [r1]! 43 vst1.8 {q8, q9}, [r0]!
|