HomeSort by relevance Sort by last modified time
    Searched refs:q4 (Results 1 - 25 of 81) sorted by null

1 2 3 4

  /external/llvm/test/MC/ARM/
neon-bitwise-encoding.s 64 veor q4, q7, q3
65 veor.8 q4, q7, q3
66 veor.16 q4, q7, q3
67 veor.32 q4, q7, q3
68 veor.64 q4, q7, q3
70 veor.i8 q4, q7, q3
71 veor.i16 q4, q7, q3
72 veor.i32 q4, q7, q3
73 veor.i64 q4, q7, q3
75 veor.s8 q4, q7, q
    [all...]
neon-shiftaccum-encoding.s 10 vsra.s64 q8, q4, #64
18 vsra.u64 q4, q5, #25
28 vsra.s64 q4, #64
45 @ CHECK: vsra.s64 q8, q4, #64 @ encoding: [0xd8,0x01,0xc0,0xf2]
53 @ CHECK: vsra.u64 q4, q5, #25 @ encoding: [0xda,0x81,0xa7,0xf3]
62 @ CHECK: vsra.s64 q4, q4, #64 @ encoding: [0xd8,0x81,0x80,0xf2]
82 vrsra.s32 q3, q4, #32
83 vrsra.s64 q4, q5, #64
100 vrsra.s32 q4, #3
    [all...]
neont2-shiftaccum-encoding.s 12 vsra.s64 q8, q4, #64
20 vsra.u64 q4, q5, #25
30 vsra.s64 q4, #64
47 @ CHECK: vsra.s64 q8, q4, #64 @ encoding: [0xc0,0xef,0xd8,0x01]
55 @ CHECK: vsra.u64 q4, q5, #25 @ encoding: [0xa7,0xff,0xda,0x81]
64 @ CHECK: vsra.s64 q4, q4, #64 @ encoding: [0x80,0xef,0xd8,0x81]
85 vrsra.s32 q3, q4, #32
86 vrsra.s64 q4, q5, #64
103 vrsra.s32 q4, #3
    [all...]
neon-shift-encoding.s 116 vsra.s64 q4, q5, #63
122 vsra.s8 q4, #7
134 @ CHECK: vsra.s64 q4, q5, #63 @ encoding: [0xda,0x81,0x81,0xf2]
139 @ CHECK: vsra.s8 q4, q4, #7 @ encoding: [0x58,0x81,0x89,0xf2]
152 vsra.u64 q4, q5, #63
158 vsra.u8 q4, #7
170 @ CHECK: vsra.u64 q4, q5, #63 @ encoding: [0xda,0x81,0x81,0xf3]
175 @ CHECK: vsra.u8 q4, q4, #7 @ encoding: [0x58,0x81,0x89,0xf3
    [all...]
vfp4.s 18 @ ARM: vfma.f32 q2, q4, q0 @ encoding: [0x50,0x4c,0x08,0xf2]
19 @ THUMB: vfma.f32 q2, q4, q0 @ encoding: [0x08,0xef,0x50,0x4c]
20 vfma.f32 q2, q4, q0
44 @ ARM: vfms.f32 q2, q4, q0 @ encoding: [0x50,0x4c,0x28,0xf2]
45 @ THUMB: vfms.f32 q2, q4, q0 @ encoding: [0x28,0xef,0x50,0x4c]
46 vfms.f32 q2, q4, q0
neon-minmax-encoding.s 20 vmax.s16 q4, q5, q6
31 vmax.u16 q4, q5
50 @ CHECK: vmax.s16 q4, q5, q6 @ encoding: [0x4c,0x86,0x1a,0xf2]
60 @ CHECK: vmax.u16 q4, q4, q5 @ encoding: [0x4a,0x86,0x18,0xf3]
82 vmin.s16 q4, q5, q6
93 vmin.u16 q4, q5
112 @ CHECK: vmin.s16 q4, q5, q6 @ encoding: [0x5c,0x86,0x1a,0xf2]
122 @ CHECK: vmin.u16 q4, q4, q5 @ encoding: [0x5a,0x86,0x18,0xf3
    [all...]
neont2-minmax-encoding.s 22 vmax.s16 q4, q5, q6
33 vmax.u16 q4, q5
52 @ CHECK: vmax.s16 q4, q5, q6 @ encoding: [0x1a,0xef,0x4c,0x86]
62 @ CHECK: vmax.u16 q4, q4, q5 @ encoding: [0x18,0xff,0x4a,0x86]
84 vmin.s16 q4, q5, q6
95 vmin.u16 q4, q5
114 @ CHECK: vmin.s16 q4, q5, q6 @ encoding: [0x1a,0xef,0x5c,0x86]
124 @ CHECK: vmin.u16 q4, q4, q5 @ encoding: [0x18,0xff,0x5a,0x86
    [all...]
neont2-pairwise-encoding.s 21 vpaddl.s8 q4, q7
24 vpaddl.u8 q7, q4
34 @ CHECK: vpaddl.s8 q4, q7 @ encoding: [0xb0,0xff,0x4e,0x82]
37 @ CHECK: vpaddl.u8 q7, q4 @ encoding: [0xb0,0xff,0xc8,0xe2]
48 vpadal.s8 q4, q10
61 @ CHECK: vpadal.s8 q4, q10 @ encoding: [0xb0,0xff,0x64,0x86]
  /external/libvpx/vp8/decoder/arm/neon/
dequant_idct_neon.asm 28 vld1.16 {q3, q4}, [r0]
41 vmul.i16 q2, q4, q6
48 vqdmulh.s16 q4, q2, d0[0]
54 vshr.s16 q4, q4, #1
57 vqadd.s16 q4, q4, q2
77 vqdmulh.s16 q4, q2, d0[0]
85 vshr.s16 q4, q4, #
    [all...]
idct_dequant_dc_full_2x_neon.asm 31 vld1.16 {q4, q5}, [r0] ; r q
52 vmul.i16 q4, q4, q0
62 ; q4: l4r4 q5: l12r12
71 vqdmulh.s16 q6, q4, d0[2] ; sinpi8sqrt2
73 vqdmulh.s16 q8, q4, d0[0] ; cospi8sqrt2minus1
90 ; q4: 4 + 4 * cospi : d1/temp1
92 vqadd.s16 q4, q4, q8
98 vqadd.s16 q3, q4, q
    [all...]
idct_dequant_full_2x_neon.asm 31 vld1.16 {q4, q5}, [r0] ; r q
48 vmul.i16 q4, q4, q0
54 ; q4: l4r4 q5: l12r12
63 vqdmulh.s16 q6, q4, d0[2] ; sinpi8sqrt2
65 vqdmulh.s16 q8, q4, d0[0] ; cospi8sqrt2minus1
82 ; q4: 4 + 4 * cospi : d1/temp1
84 vqadd.s16 q4, q4, q8
90 vqadd.s16 q3, q4, q
    [all...]
  /external/libvpx/vp8/encoder/arm/neon/
fastquantizeb_neon.asm 34 vabs.s16 q4, q0 ;calculate x = abs(z)
37 vcge.s16 q10, q4, q10 ;x>=zbin
58 vadd.s16 q4, q6 ;x + Round
63 vqdmulh.s16 q4, q8 ;y = ((Round + abs(z)) * Quant) >> 16
69 vshr.s16 q4, #1 ;right shift 1 after vqdmulh
73 veor.s16 q4, q2 ; y^sz
78 vsub.s16 q4, q2 ; x1 = (y^sz) - sz = (y^sz) - (-1) (two's complement)
81 vand.s16 q4, q10 ;mask off x1 elements
86 vtst.16 q14, q4, q8 ;now find eob
87 vtst.16 q15, q5, q8 ;non-zero element is set to all 1 in q4, q
    [all...]
fastfdct4x4_neon.asm 50 vshl.i16 q4, q4, #1 ; c1, d1
56 vqdmulh.s16 q8, q4, d0[0]
57 vqdmulh.s16 q7, q4, d0[2]
62 vadd.s16 q8, q4, q8 ;d16:temp2 = ((c1 * x_c1)>>16) + c1; d17:temp2 = ((d1 * x_c1)>>16) + d1
86 vqdmulh.s16 q8, q4, d0[0]
87 vqdmulh.s16 q7, q4, d0[2]
92 vadd.s16 q8, q4, q8 ;d16:temp2 = ((c1 * x_c1)>>16) + c1; d17:temp2 = ((d1 * x_c1)>>16) + d1
100 vclt.s16 q4, q2, #0
103 vsub.s16 q2, q2, q4
    [all...]
vp8_subpixelvariance16x16s_neon.asm 57 vext.8 q5, q4, q5, #1
63 vrhadd.u8 q2, q4, q5
66 vsubl.u8 q4, d0, d22 ;diff
75 vpadal.s16 q8, q4 ;sum
147 vld1.u8 {q4}, [r0], r1
155 vrhadd.u8 q2, q2, q4
156 vrhadd.u8 q4, q4, q6
255 vext.8 q5, q4, q5, #1
260 vrhadd.u8 q2, q4, q
    [all...]
  /external/valgrind/main/none/tests/arm/
neon128.c 439 TESTINSN_bin("vand q4, q6, q5", q4, q6, i8, 0xff, q5, i16, 0x57);
445 TESTINSN_bin("vbic q4, q6, q5", q4, q6, i8, 0xff, q5, i16, 0x57);
452 TESTINSN_bin("vorr q4, q4, q4", q4, q4, i16, 0xff, q4, i16, 0xff)
    [all...]
neon128.stdout.exp 120 vand q4, q6, q5 :: Qd 0x00570057 0x00570057 0x00570057 0x00570057 Qm (i8)0x000000ff Qn (i16)0x00000057
125 vbic q4, q6, q5 :: Qd 0xffa8ffa8 0xffa8ffa8 0xffa8ffa8 0xffa8ffa8 Qm (i8)0x000000ff Qn (i16)0x00000057
131 vorr q4, q4, q4 :: Qd 0x00ff00ff 0x00ff00ff 0x00ff00ff 0x00ff00ff Qm (i16)0x000000ff Qn (i16)0x000000ff
136 vorn q4, q4, q4 :: Qd 0xffffffff 0xffffffff 0xffffffff 0xffffffff Qm (i16)0x000000ff Qn (i16)0x000000ff
140 veor q4, q6, q5 :: Qd 0xffa8ffa8 0xffa8ffa8 0xffa8ffa8 0xffa8ffa8 Qm (i8)0x000000ff Qn (i16)0x00000057
145 veor q4, q4, q4 :: Qd 0x00000000 0x00000000 0x00000000 0x00000000 Qm (i16)0x000000ff Qn (i16)0x000000f
    [all...]
  /external/libvpx/vp8/common/arm/neon/
shortidct4x4llm_neon.asm 41 vqdmulh.s16 q4, q2, d0[0]
47 vshr.s16 q4, q4, #1
50 vqadd.s16 q4, q4, q2
73 vqdmulh.s16 q4, q2, d0[0]
79 vshr.s16 q4, q4, #1
82 vqadd.s16 q4, q4, q
    [all...]
loopfiltersimplehorizontaledge_neon.asm 63 vqsub.s8 q4, q5, q8 ; q4: vp8_filter = vp8_signed_char_clamp(ps1-qs1)
77 ;vqadd.s8 q4, q4, q2 ; vp8_filter = vp8_signed_char_clamp(vp8_filter + 3 * ( qs0 - ps0))
82 vand q4, q4, q15 ; vp8_filter &= mask
84 vqadd.s8 q2, q4, q10 ; Filter2 = vp8_signed_char_clamp(vp8_filter+3)
85 vqadd.s8 q4, q4, q9 ; Filter1 = vp8_signed_char_clamp(vp8_filter+4)
87 vshr.s8 q4, q4, #3 ; Filter1 >>=
    [all...]
sixtappredict4x4_neon.asm 63 vld1.u8 {q4}, [r0], r1
89 vmov q4, q3 ;keep original src data in q4 q6
94 vshr.u64 q9, q4, #8 ;construct src_ptr[-1]
101 vshr.u64 q3, q4, #32 ;construct src_ptr[2]
108 vshr.u64 q9, q4, #16 ;construct src_ptr[0]
115 vshr.u64 q3, q4, #24 ;construct src_ptr[1]
126 vld1.u8 {q4}, [r0], r1
155 vmov q4, q3 ;keep original src data in q4 q
    [all...]
recon16x16mb_neon.asm 39 vmovl.u8 q4, d28
54 vadd.s16 q4, q4, q12
63 vqmovun.s16 d4, q4
90 vmovl.u8 q4, d28
101 vadd.s16 q4, q4, q12
111 vqmovun.s16 d4, q4
  /frameworks/rs/driver/
rsdIntrinsics_Convolve.S 33 vpush {q4-q7}
59 vmovl.u8 q4, d28
102 vpop {q4-q7}
116 vpush {q4-q7}
170 vpop {q4-q7}
184 vpush {q4-q7}
227 vpop {q4-q7}
241 vpush {q4-q7}
272 vpop {q4-q7}
292 vpush {q4-q7
    [all...]
  /frameworks/support/renderscript/v8/rs_support/driver/
rsdIntrinsics_Convolve.S 33 vpush {q4-q7}
59 vmovl.u8 q4, d28
102 vpop {q4-q7}
116 vpush {q4-q7}
170 vpop {q4-q7}
184 vpush {q4-q7}
227 vpop {q4-q7}
241 vpush {q4-q7}
272 vpop {q4-q7}
292 vpush {q4-q7
    [all...]
  /frameworks/av/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p10/src_gcc/
armVCM4P10_InterpolateLuma_HalfHor4x4_unsafe_s.S 24 VADDL.U8 q4,d14,d15
36 VADDL.U8 q4,d16,d17
48 VADDL.U8 q4,d18,d19
60 VADDL.U8 q4,d20,d21
omxVCM4P10_TransformDequantLumaDCFromPair_s.S 47 VMOV.I32 q4,#0x2
51 VMLAL.S16 q4,d1,d5
55 VSHRN.I32 d1,q4,#2
  /hardware/samsung_slsi/exynos5/libswconverter/
csc_ARGB8888_to_YUV420SP_NEON.s 38 @q4: B
78 vand.u16 q4,#0x00FF @R
84 vmls.u16 q8,q4,q11 @q0:U -(38 * R[k]) @128<<6+ 32 + u>>2
89 vmla.u16 q7,q4,q13 @112 * R[k]
106 vmul.u16 q7,q4,q14 @q0 = 66 *R[k]
120 vshr.u16 q4,q4,#8 @R
124 vmul.u16 q0,q4,q14 @q0 = 66 *R[k]
151 vand.u16 q4,#0x00FF @R
157 vmul.u16 q7,q4,q14 @q0 = 66 *R[k
    [all...]

Completed in 962 milliseconds

1 2 3 4