/external/libvpx/libvpx/vp8/common/arm/neon/ |
idct_dequant_full_2x_neon.asm | 28 vld1.16 {q4, q5}, [r0] ; r q 47 vmul.i16 q5, q5, q1 52 ; q4: l4r4 q5: l12r12 62 vqdmulh.s16 q7, q5, d0[2] 64 vqdmulh.s16 q9, q5, d0[0] 81 ; q5: 12 + 12 * cospi : c1/temp2 83 vqadd.s16 q5, q5, q9 87 vqsub.s16 q2, q6, q5 [all...] |
sixtappredict4x4_neon.asm | 64 vld1.u8 {q5}, [r0], r1 89 vmov q6, q5 101 vshr.u64 q5, q6, #32 115 vshr.u64 q5, q6, #24 130 vld1.u8 {q5}, [r0], r1 155 vmov q6, q5 169 vshr.u64 q5, q6, #32 189 vshr.u64 q5, q6, #24 210 vld1.s32 {q5, q6}, [r3] ;load second_pass filter 217 vabs.s32 q7, q5 [all...] |
vp8_subpixelvariance16x16s_neon.asm | 57 vext.8 q5, q4, q5, #1 63 vrhadd.u8 q2, q4, q5 67 vsubl.u8 q5, d1, d23 81 vpadal.s16 q8, q5 113 vmull.s32 q5, d0, d0 150 vld1.8 {q5}, [r2], r3 209 vmull.s32 q5, d0, d0 255 vext.8 q5, q4, q5, # [all...] |
sixtappredict8x8_neon.asm | 74 vld1.u8 {q5}, [r0], r1 135 vmull.u8 q5, d30, d3 142 vqadd.s16 q9, q5 155 vld1.u8 {q5}, [r0], r1 165 ;vld1.u8 {q5}, [r0], r1 231 vmull.u8 q5, d29, d3 237 vqadd.s16 q10, q5 250 vld1.s32 {q5, q6}, [r3] ;load second_pass filter 255 vabs.s32 q7, q5 274 vmull.u8 q5, d20, d [all...] |
sixtappredict8x4_neon.asm | 71 vld1.u8 {q5}, [r0], r1 131 vmull.u8 q5, d30, d3 136 vqadd.s16 q9, q5 148 vld1.u8 {q5}, [r0], r1 218 vmull.u8 q5, d29, d3 224 vqadd.s16 q10, q5 239 vld1.s32 {q5, q6}, [r3] ;load second_pass filter 242 vabs.s32 q7, q5 256 vmull.u8 q5, d24, d0 261 vmlsl.u8 q5, d25, d [all...] |
buildintrapredictorsmby_neon.asm | 245 vmull.u8 q5, d17, d0 248 vsub.s16 q5, q5, q7 265 vqadd.s16 q9, q0, q5 268 vqadd.s16 q11, q1, q5 271 vqadd.s16 q13, q2, q5 274 vqadd.s16 q15, q3, q5 528 vmull.u8 q5, d17, d0 531 vsub.s16 q5, q5, q [all...] |
/external/valgrind/main/none/tests/arm/ |
neon128.c | 358 TESTINSN_imm("vmov.i32 q5", q5, 0x700); 372 TESTINSN_imm("vmvn.i32 q5", q5, 0x700); 391 TESTINSN_imm("vbic.i32 q5", q5, 0x700); 439 TESTINSN_bin("vand q4, q6, q5", q4, q6, i8, 0xff, q5, i16, 0x57); 445 TESTINSN_bin("vbic q4, q6, q5", q4, q6, i8, 0xff, q5, i16, 0x57) [all...] |
neon128.stdout.exp | 8 vmov.i32 q5, #0x700 :: Qd 0x00000700 0x00000700 0x00000700 0x00000700 9 vmov.i32 q5, #0x700 :: Qd 0x00000700 0x00000700 0x00000700 0x00000700 33 vmvn.i32 q5, #0x700 :: Qd 0xfffff8ff 0xfffff8ff 0xfffff8ff 0xfffff8ff 34 vmvn.i32 q5, #0x700 :: Qd 0xfffff8ff 0xfffff8ff 0xfffff8ff 0xfffff8ff 65 vbic.i32 q5, #0x700 :: Qd 0x55555055 0x55555055 0x55555055 0x55555055 66 vbic.i32 q5, #0x700 :: Qd 0x151d181d 0x141c181c 0x131b181b 0x121f181f 120 vand q4, q6, q5 :: Qd 0x00570057 0x00570057 0x00570057 0x00570057 Qm (i8)0x000000ff Qn (i16)0x00000057 125 vbic q4, q6, q5 :: Qd 0xffa8ffa8 0xffa8ffa8 0xffa8ffa8 0xffa8ffa8 Qm (i8)0x000000ff Qn (i16)0x00000057 140 veor q4, q6, q5 :: Qd 0xffa8ffa8 0xffa8ffa8 0xffa8ffa8 0xffa8ffa8 Qm (i8)0x000000ff Qn (i16)0x00000057 149 vbsl q4, q6, q5 :: Qd 0x55575557 0x55575557 0x55575557 0x55575557 Qm (i8)0x000000ff Qn (i16)0x0000005 [all...] |
/external/llvm/test/MC/ARM/ |
neon-shiftaccum-encoding.s | 9 vsra.s32 q9, q5, #32 18 vsra.u64 q4, q5, #25 27 vsra.s32 q5, #32 36 vsra.u64 q5, #25 44 @ CHECK: vsra.s32 q9, q5, #32 @ encoding: [0x5a,0x21,0xe0,0xf2] 53 @ CHECK: vsra.u64 q4, q5, #25 @ encoding: [0xda,0x81,0xa7,0xf3] 61 @ CHECK: vsra.s32 q5, q5, #32 @ encoding: [0x5a,0xa1,0xa0,0xf2] 70 @ CHECK: vsra.u64 q5, q5, #25 @ encoding: [0xda,0xa1,0xa7,0xf3 [all...] |
neont2-shiftaccum-encoding.s | 11 vsra.s32 q9, q5, #32 20 vsra.u64 q4, q5, #25 29 vsra.s32 q5, #32 38 vsra.u64 q5, #25 46 @ CHECK: vsra.s32 q9, q5, #32 @ encoding: [0xe0,0xef,0x5a,0x21] 55 @ CHECK: vsra.u64 q4, q5, #25 @ encoding: [0xa7,0xff,0xda,0x81] 63 @ CHECK: vsra.s32 q5, q5, #32 @ encoding: [0xa0,0xef,0x5a,0xa1] 72 @ CHECK: vsra.u64 q5, q5, #25 @ encoding: [0xa7,0xff,0xda,0xa1 [all...] |
neon-minmax-encoding.s | 20 vmax.s16 q4, q5, q6 25 vmax.f32 q9, q5, q1 28 vmax.s16 q5, q6 31 vmax.u16 q4, q5 50 @ CHECK: vmax.s16 q4, q5, q6 @ encoding: [0x4c,0x86,0x1a,0xf2] 55 @ CHECK: vmax.f32 q9, q5, q1 @ encoding: [0x42,0x2f,0x4a,0xf2] 57 @ CHECK: vmax.s16 q5, q5, q6 @ encoding: [0x4c,0xa6,0x1a,0xf2] 60 @ CHECK: vmax.u16 q4, q4, q5 @ encoding: [0x4a,0x86,0x18,0xf3] 82 vmin.s16 q4, q5, q [all...] |
neont2-minmax-encoding.s | 22 vmax.s16 q4, q5, q6 27 vmax.f32 q9, q5, q1 30 vmax.s16 q5, q6 33 vmax.u16 q4, q5 52 @ CHECK: vmax.s16 q4, q5, q6 @ encoding: [0x1a,0xef,0x4c,0x86] 57 @ CHECK: vmax.f32 q9, q5, q1 @ encoding: [0x4a,0xef,0x42,0x2f] 59 @ CHECK: vmax.s16 q5, q5, q6 @ encoding: [0x1a,0xef,0x4c,0xa6] 62 @ CHECK: vmax.u16 q4, q4, q5 @ encoding: [0x18,0xff,0x4a,0x86] 84 vmin.s16 q4, q5, q [all...] |
neon-shift-encoding.s | 116 vsra.s64 q4, q5, #63 123 vsra.s16 q5, #15 134 @ CHECK: vsra.s64 q4, q5, #63 @ encoding: [0xda,0x81,0x81,0xf2] 140 @ CHECK: vsra.s16 q5, q5, #15 @ encoding: [0x5a,0xa1,0x91,0xf2] 152 vsra.u64 q4, q5, #63 159 vsra.u16 q5, #15 170 @ CHECK: vsra.u64 q4, q5, #63 @ encoding: [0xda,0x81,0x81,0xf3] 176 @ CHECK: vsra.u16 q5, q5, #15 @ encoding: [0x5a,0xa1,0x91,0xf3 [all...] |
neont2-pairwise-encoding.s | 22 vpaddl.s16 q5, q6 23 vpaddl.s32 q6, q5 35 @ CHECK: vpaddl.s16 q5, q6 @ encoding: [0xb4,0xff,0x4c,0xa2] 36 @ CHECK: vpaddl.s32 q6, q5 @ encoding: [0xb8,0xff,0x4a,0xc2] 49 vpadal.s16 q5, q11 62 @ CHECK: vpadal.s16 q5, q11 @ encoding: [0xb4,0xff,0x66,0xa6]
|
neon-bitwise-encoding.s | 241 vand q6, q5 242 vand.s8 q6, q5 247 veor q6, q5 248 veor.8 q6, q5 253 veor q6, q5 254 veor.i8 q6, q5 259 vclt.s16 q5, #0 262 vceq.s16 q5, q3 265 vcgt.s16 q5, q3 268 vcge.s16 q5, q [all...] |
/external/libvpx/libvpx/vp9/common/arm/neon/ |
vp9_short_idct32x32_add_neon.asm | 180 vrshr.s16 q5, q5, #6 185 vaddw.u8 q5, q5, d5 190 vqmovun.s16 d5, q5 214 vrshr.s16 q5, q5, #6 219 vaddw.u8 q5, q5, d5 224 vqmovun.s16 d5, q5 [all...] |
vp9_short_iht8x8_add_neon.asm | 135 vmull.s16 q5, d26, d2 143 vmlsl.s16 q5, d22, d3 151 vqrshrn.s32 d10, q5, #14 ; >> 14 241 vsub.s16 q13, q4, q5 ; step2[5] = step1[4] - step1[5] 242 vadd.s16 q4, q4, q5 ; step2[4] = step1[4] + step1[5] 273 vadd.s16 q10, q2, q5 ; output[2] = step1[2] + step1[5]; 276 vsub.s16 q13, q2, q5 ; output[5] = step1[2] - step1[5]; 310 vmull.s16 q5, d22, d30 318 vmlal.s16 q5, d24, d31 326 vadd.s32 q11, q1, q5 [all...] |
vp9_short_idct16x16_add_neon.asm | 92 vmull.s16 q5, d18, d1 100 vmlal.s16 q5, d30, d0 111 vqrshrn.s32 d14, q5, #14 ; >> 14 206 vsub.s16 q13, q4, q5 ; step2[5] = step1[4] - step1[5]; 207 vadd.s16 q4, q4, q5 ; step2[4] = step1[4] + step1[5]; 250 vadd.s16 q10, q2, q5 ; step2[2] = step1[2] + step1[5]; 253 vsub.s16 q13, q2, q5 ; step2[5] = step1[2] - step1[5]; 376 vmull.s16 q5, d25, d31 384 vmlal.s16 q5, d23, d30 395 vqrshrn.s32 d13, q5, #14 ; >> 1 [all...] |
/hardware/samsung_slsi/exynos5/libswconverter/ |
csc_ARGB8888_to_YUV420SP_NEON.s | 39 @q5: temp3, output 79 vand.u16 q5,#0x00FF @G 85 vmls.u16 q8,q5,q12 @-(74 * G[k]) 90 vmls.u16 q7,q5,q14 @q0:U -(94 * G[k]) @128<<6+ 32 + v>>2 107 vmla.u16 q7,q5,q15 @q0 += 129 *G[k] 121 vshr.u16 q5,q5,#8 @G 125 vmla.u16 q0,q5,q15 @q0 += 129 *G[k] 152 vand.u16 q5,#0x00FF @G 158 vmla.u16 q7,q5,q15 @q0 += 129 *G[k [all...] |
csc_interleave_memcpy_neon.s | 86 vld1.8 {q5}, [r12]! 95 vst2.8 {q4, q5}, [r10]!
|
/frameworks/av/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p10/src_gcc/ |
omxVCM4P10_TransformDequantLumaDCFromPair_s.S | 48 VMOV.I32 q5,#0x2 52 VMLAL.S16 q5,d2,d5 56 VSHRN.I32 d2,q5,#2
|
/external/libvpx/libvpx/vp9/common/mips/dspr2/ |
vp9_loopfilter_macros_dspr2.h | 370 "preceu.ph.qbl %[q5_l], %[q5] \n\t" \ 379 [q4] "r" (q4), [q5] "r" (q5), [q6] "r" (q6), [q7] "r" (q7) \ 410 "preceu.ph.qbr %[q5_r], %[q5] \n\t" \ 419 [q4] "r" (q4), [q5] "r" (q5), [q6] "r" (q6), [q7] "r" (q7) \ 451 "precr.qb.ph %[q5], %[q5_l], %[q5_r] \n\t" \ 457 [q5] "=&r" (q5),[q6] "=&r" (q6) \
|
/external/libvpx/libvpx/vp8/encoder/arm/neon/ |
fastquantizeb_neon.asm | 38 vabs.s16 q5, q1 50 vadd.s16 q5, q7 55 vqdmulh.s16 q5, q9 64 veor.s16 q5, q3 75 vshr.s16 q5, #1 80 vsub.s16 q5, q3 90 vst1.s16 {q4, q5}, [r7] ; store: qcoeff = x1 99 vmul.s16 q3, q7, q5 113 vtst.16 q15, q5, q8 ; non-zero element is set to all 1
|
vp8_memcpy_neon.asm | 36 vld1.8 {q4, q5}, [r1]! 39 vst1.8 {q4, q5}, [r0]!
|
/external/libvpx/libvpx/vp9/common/ |
vp9_loopfilter_filters.c | 223 q4 = *oq4, q5 = *oq5, q6 = *oq6, q7 = *oq7; local 237 q0 + q1 + q2 + q3 + q4 + q5, 4); 239 q0 + q1 + q2 + q3 + q4 + q5 + q6, 4); 241 q0 * 2 + q1 + q2 + q3 + q4 + q5 + q6 + q7, 4); 243 q0 + q1 * 2 + q2 + q3 + q4 + q5 + q6 + q7 * 2, 4); 245 q0 + q1 + q2 * 2 + q3 + q4 + q5 + q6 + q7 * 3, 4); 247 q0 + q1 + q2 + q3 * 2 + q4 + q5 + q6 + q7 * 4, 4); 249 q0 + q1 + q2 + q3 + q4 * 2 + q5 + q6 + q7 * 5, 4); 251 q0 + q1 + q2 + q3 + q4 + q5 * 2 + q6 + q7 * 6, 4); 253 q0 + q1 + q2 + q3 + q4 + q5 + q6 * 2 + q7 * 7, 4) [all...] |