/external/libvpx/libvpx/vp8/common/arm/neon/ |
sixtappredict4x4_neon.asm | 86 vmull.u8 q8, d20, d5 96 vmlal.u8 q8, d10, d0 103 vmlsl.u8 q8, d20, d1 110 vmlsl.u8 q8, d10, d4 117 vmlal.u8 q8, d20, d2 128 vqadd.s16 q8, q10 134 vqrshrun.s16 d28, q8, #7 151 vmull.u8 q8, d20, d5 163 vmlal.u8 q8, d10, d0 173 vmlsl.u8 q8, d20, d [all...] |
idct_dequant_full_2x_neon.asm | 59 ; q8: 4 * cospi 63 vqdmulh.s16 q8, q4, d0[0] ; cospi8sqrt2minus1 77 vshr.s16 q8, q8, #1 82 vqadd.s16 q4, q4, q8 110 ; q8: 1 * sinpi : c1/temp1 114 vqdmulh.s16 q8, q5, d0[2] ; sinpi8sqrt2 131 ; q8: c1 = temp1 - temp2 133 vqsub.s16 q8, q8, q1 [all...] |
sixtappredict8x4_neon.asm | 80 vmull.u8 q8, d8, d0 90 vmlsl.u8 q8, d29, d1 100 vmlsl.u8 q8, d29, d4 110 vmlal.u8 q8, d29, d2 120 vmlal.u8 q8, d29, d5 135 vqadd.s16 q8, q4 142 vqrshrun.s16 d23, q8, #7 156 vmull.u8 q8, d6, d0 ;(src_ptr[-2] * vp8_filter[0]) 168 vmlsl.u8 q8, d27, d1 ;-(src_ptr[-1] * vp8_filter[1]) 180 vmlsl.u8 q8, d27, d4 ;-(src_ptr[2] * vp8_filter[4] [all...] |
sixtappredict8x8_neon.asm | 84 vmull.u8 q8, d8, d0 94 vmlsl.u8 q8, d29, d1 104 vmlsl.u8 q8, d29, d4 114 vmlal.u8 q8, d29, d2 124 vmlal.u8 q8, d29, d5 141 vqadd.s16 q8, q4 148 vqrshrun.s16 d23, q8, #7 169 vmull.u8 q8, d6, d0 ;(src_ptr[-2] * vp8_filter[0]) 181 vmlsl.u8 q8, d27, d1 ;-(src_ptr[-1] * vp8_filter[1]) 193 vmlsl.u8 q8, d27, d4 ;-(src_ptr[2] * vp8_filter[4] [all...] |
vp8_subpixelvariance16x16_neon.asm | 71 vmull.u8 q8, d3, d0 94 vmlal.u8 q8, d3, d1 ;(src_ptr[0] * Filter[1]) 102 vqrshrn.u16 d15, q8, #7 207 vmull.u8 q8, d29, d0 216 vmlal.u8 q8, d31, d1 227 vqrshrn.u16 d9, q8, #7 259 vmull.u8 q8, d3, d0 282 vmlal.u8 q8, d3, d1 ;(src_ptr[0] * Filter[1]) 290 vqrshrn.u16 d15, q8, #7 334 vmull.u8 q8, d29, d [all...] |
bilinearpredict8x8_neon.asm | 48 vmull.u8 q8, d6, d0 58 vmlal.u8 q8, d7, d1 66 vqrshrn.u16 d24, q8, #7 75 vmull.u8 q8, d6, d0 87 vmlal.u8 q8, d7, d1 93 vqrshrn.u16 d28, q8, #7 118 vmull.u8 q8, d29, d0 127 vmlal.u8 q8, d30, d1 136 vqrshrn.u16 d9, q8, #7
|
sixtappredict16x16_neon.asm | 89 vmull.u8 q8, d6, d0 ;(src_ptr[-2] * vp8_filter[0]) 100 vmlsl.u8 q8, d28, d1 ;-(src_ptr[-1] * vp8_filter[1]) 116 vmlsl.u8 q8, d28, d4 ;-(src_ptr[2] * vp8_filter[4]) 132 vmlal.u8 q8, d28, d5 ;(src_ptr[3] * vp8_filter[5]) 148 vmlal.u8 q8, d28, d2 ;(src_ptr[0] * vp8_filter[2]) 172 vqadd.s16 q8, q4 ;sum of all (src_data*filter_parameters) 186 vqrshrun.s16 d6, q8, #7 ;shift/round/saturate to u8 207 vabs.s32 q8, q6 258 vmull.u8 q8, d22, d3 265 vqadd.s16 q8, q [all...] |
bilinearpredict16x16_neon.asm | 63 vmull.u8 q8, d3, d0 86 vmlal.u8 q8, d3, d1 ;(src_ptr[0] * vp8_filter[1]) 94 vqrshrn.u16 d15, q8, #7 196 vmull.u8 q8, d29, d0 205 vmlal.u8 q8, d31, d1 216 vqrshrn.u16 d9, q8, #7 248 vmull.u8 q8, d3, d0 271 vmlal.u8 q8, d3, d1 ;(src_ptr[0] * vp8_filter[1]) 279 vqrshrn.u16 d15, q8, #7 320 vmull.u8 q8, d29, d [all...] |
bilinearpredict4x4_neon.asm | 58 vmull.u8 q8, d4, d0 62 vmlal.u8 q8, d10, d1 66 vqrshrn.u16 d29, q8, #7
|
/external/libvpx/libvpx/vp9/common/arm/neon/ |
vp9_short_iht4x4_add_neon.asm | 47 vadd.s16 q8, q13, q14 66 vmull.s16 q8, d4, d19 ; s5 = sinpi_2_9 * x3 71 vadd.s32 q10, q10, q8 73 vdup.32 q8, r0 ; duplicate sinpi_3_9 75 vmul.s32 q15, q15, q8 ; x2 = sinpi_3_9 * s7 138 vtrn.32 q8, q9 153 vld1.s16 {q8,q9}, [r0]! 211 vrshr.s16 q8, q8, #4 220 vaddw.u8 q8, q8, d2 [all...] |
vp9_short_idct16x16_add_neon.asm | 21 ; Transpose a 8x8 16bit data matrix. Datas are loaded in q8-q15. 28 vtrn.32 q8, q10 32 vtrn.16 q8, q9 46 ; idct16 stage1 - stage6 on all the elements loaded in q8-q15. The output 47 ; will be stored back into q8-q15 registers. This function will touch q0-q7 52 ; load elements of 0, 2, 4, 6, 8, 10, 12, 14 into q8 - q15 53 vld2.s16 {q8,q9}, [r0]! 216 vadd.s16 q0, q8, q11 ; step1[0] = step2[0] + step2[3]; 219 vsub.s16 q3, q8, q11 ; step1[3] = step2[0] - step2[3]; 248 vadd.s16 q8, q0, q15 ; step2[0] = step1[0] + step1[7] [all...] |
vp9_short_idct32x32_add_neon.asm | 113 vrshr.s16 q8, q8, #6 118 vaddw.u8 q8, q8, d10 123 vqmovun.s16 d10, q8 147 vrshr.s16 q8, q8, #6 152 vaddw.u8 q8, q8, d10 157 vqmovun.s16 d10, q8 [all...] |
vp9_avg_neon.asm | 39 vld1.8 {q8-q9}, [r6@128]! 41 vrhadd.u8 q0, q0, q8 54 vld1.8 {q8-q9}, [r6@128], r3 57 vrhadd.u8 q0, q0, q8
|
vp9_short_iht8x8_add_neon.asm | 101 ; Transpose a 8x8 16bits data matrix. Datas are loaded in q8-q15. 108 vtrn.32 q8, q10 112 vtrn.16 q8, q9 119 ; loaded in q8-q15. The IDCT constants are loaded in r3 - r9. The output 120 ; will be stored back into q8-q15 registers. This macro will touch q0-q7 213 vmull.s16 q8, d20, d1 221 vmlal.s16 q8, d28, d0 229 vqrshrn.s32 d30, q8, #14 ; >> 14 271 vadd.s16 q8, q0, q7 ; output[0] = step1[0] + step1[7]; 282 ; loaded in q8-q15. IADST constants are loaded in r0 - r12 registers. Th [all...] |
/hardware/samsung_slsi/exynos5/libswconverter/ |
csc_ARGB8888_to_YUV420SP_NEON.s | 44 vmov.u16 q8, #25 82 vmov.u16 q8,q9 @ CalcU() 83 vmla.u16 q8,q6,q13 @112 * B[k] 84 vmls.u16 q8,q4,q11 @q0:U -(38 * R[k]) @128<<6+ 32 + u>>2 85 vmls.u16 q8,q5,q12 @-(74 * G[k]) 86 vshr.u16 q8,q8, #8 @(128<<8+ 128 + u)>>8 95 vtrn.8 q8,q7 96 vst1.8 {q8}, [r1]! @write UV component to yuv420_buffer+linear_ylanesiez 102 vmov.u16 q8, #25 @#2 [all...] |
csc_interleave_memcpy_neon.s | 80 vld1.8 {q8}, [r11]! 97 vst2.8 {q8, q9}, [r10]!
|
/external/llvm/test/MC/ARM/ |
neon-v8.s | 34 vcvtn.s32.f32 q3, q8 35 @ CHECK: vcvtn.s32.f32 q3, q8 @ encoding: [0x60,0x61,0xbb,0xf3] 45 vcvtp.u32.f32 q9, q8 46 @ CHECK: vcvtp.u32.f32 q9, q8 @ encoding: [0xe0,0x22,0xfb,0xf3] 58 vrinta.f32 q8, q2 59 @ CHECK: vrinta.f32 q8, q2 @ encoding: [0x44,0x05,0xfa,0xf3]
|
thumb-neon-v8.s | 34 vcvtn.s32.f32 q3, q8 35 @ CHECK: vcvtn.s32.f32 q3, q8 @ encoding: [0xbb,0xff,0x60,0x61] 45 vcvtp.u32.f32 q9, q8 46 @ CHECK: vcvtp.u32.f32 q9, q8 @ encoding: [0xfb,0xff,0xe0,0x22] 58 vrinta.f32 q8, q2 59 @ CHECK: vrinta.f32 q8, q2 @ encoding: [0xfa,0xff,0x44,0x05]
|
neont2-pairwise-encoding.s | 25 vpaddl.u16 q8, q3 38 @ CHECK: vpaddl.u16 q8, q3 @ encoding: [0xf4,0xff,0xc6,0x02] 52 vpadal.u16 q8, q14 65 @ CHECK: vpadal.u16 q8, q14 @ encoding: [0xf4,0xff,0xec,0x06]
|
/external/clang/test/SemaCXX/ |
constexpr-nqueens.cpp | 55 constexpr Board q8 = buildBoard(8); variable 65 static_assert(q8.check(
|
/external/libvpx/libvpx/vp8/encoder/arm/neon/ |
vp8_memcpy_neon.asm | 40 vld1.8 {q8, q9}, [r1]! 43 vst1.8 {q8, q9}, [r0]!
|
/external/valgrind/main/none/tests/arm/ |
neon128.c | 383 TESTINSN_imm("vorr.i32 q8", q8, 0x700); 392 TESTINSN_imm("vbic.i16 q8", q8, 0x700); [all...] |
/external/libvpx/libvpx/vpx_scale/arm/neon/ |
vp8_vpxyv12_copyframe_func_neon.asm | 61 vld1.8 {q8, q9}, [r10]! 72 vst1.8 {q8, q9}, [r11]! 121 vld1.8 {q8, q9}, [r10]! 128 vst1.8 {q8, q9}, [r11]!
|
/external/chromium_org/third_party/openssl/openssl/crypto/poly1305/ |
poly1305_arm_asm.S | 977 # asm 2: vshr.u64 >t2=q8,<r1=q3,#26 978 vshr.u64 q8,q3,#26 992 # asm 2: vadd.i64 >r2=q8,<r2=q13,<t2=q8 993 vadd.i64 q8,q13,q8 1022 # asm 2: vshr.u64 >t3=q13,<r2=q8,#26 1023 vshr.u64 q13,q8,#26 1037 # asm 2: vand >x23=q9,<r2=q8,<mask=q6 1038 vand q9,q8,q [all...] |
/external/chromium_org/third_party/openssl/openssl/crypto/ |
armv4cpuid.S | 104 .word 0xf34001f0 @ veor q8, q8, q8
|