HomeSort by relevance Sort by last modified time
    Searched defs:q4 (Results 1 - 25 of 144) sorted by null

1 2 3 4 5 6

  /prebuilts/go/darwin-x86/test/ken/
divmod.go 21 q4 = +1
43 if n2/d2 != q4 || n2%d2 != r4 {
66 if in2/id2 != q4 || in2%id2 != r4 {
89 if bn2/bd2 != q4 || bn2%bd2 != r4 {
112 if sn2/sd2 != q4 || sn2%sd2 != r4 {
135 if ln2/ld2 != q4 || ln2%ld2 != r4 {
158 if qn2/qd2 != q4 || qn2%qd2 != r4 {
175 if n2/qd2 != q4 || n2%qd2 != r4 {
192 if qn2/d2 != q4 || qn2%d2 != r4 {
20 q4 = +1 const
  /prebuilts/go/linux-x86/test/ken/
divmod.go 21 q4 = +1
43 if n2/d2 != q4 || n2%d2 != r4 {
66 if in2/id2 != q4 || in2%id2 != r4 {
89 if bn2/bd2 != q4 || bn2%bd2 != r4 {
112 if sn2/sd2 != q4 || sn2%sd2 != r4 {
135 if ln2/ld2 != q4 || ln2%ld2 != r4 {
158 if qn2/qd2 != q4 || qn2%qd2 != r4 {
175 if n2/qd2 != q4 || n2%qd2 != r4 {
192 if qn2/d2 != q4 || qn2%d2 != r4 {
20 q4 = +1 const
  /external/clang/test/SemaCXX/
builtin-assume-aligned.cpp 25 constexpr void *q4 = __builtin_assume_aligned(&n, 4, -4); variable
extern-c.cpp 243 ExternCStruct4 *q4 = p4; // expected-error {{ambiguous}} member in namespace:tag_hiding
  /external/capstone/suite/MC/ARM/
vfp4-thumb.s.cs 5 0x08,0xef,0x50,0x4c = vfma.f32 q2, q4, q0 package
11 0x28,0xef,0x50,0x4c = vfms.f32 q2, q4, q0 package
vfp4.s.cs 5 0x50,0x4c,0x08,0xf2 = vfma.f32 q2, q4, q0 package
11 0x50,0x4c,0x28,0xf2 = vfms.f32 q2, q4, q0 package
neon-v8.s.cs 3 0x5c,0x4f,0x08,0xf3 = vmaxnm.f32 q2, q4, q6 package
8 0x4c,0x80,0xbb,0xf3 = vcvta.s32.f32 q4, q6
9 0xe4,0x80,0xbb,0xf3 = vcvta.u32.f32 q4, q10
20 0x6e,0x82,0xbb,0xf3 = vcvtp.s32.f32 q4, q15
23 0x48,0x24,0xba,0xf3 = vrintn.f32 q1, q4
29 0xc8,0x25,0xfa,0xf3 = vrintz.f32 q9, q4
31 0xc8,0x26,0xba,0xf3 = vrintm.f32 q1, q4
33 0xc8,0x27,0xba,0xf3 = vrintp.f32 q1, q4
37 0xc8,0x25,0xfa,0xf3 = vrintz.f32 q9, q4
38 0xc8,0x27,0xba,0xf3 = vrintp.f32 q1, q4
    [all...]
thumb-neon-v8.s.cs 3 0x08,0xff,0x5c,0x4f = vmaxnm.f32 q2, q4, q6 package
8 0xbb,0xff,0x4c,0x80 = vcvta.s32.f32 q4, q6
9 0xbb,0xff,0xe4,0x80 = vcvta.u32.f32 q4, q10
20 0xbb,0xff,0x6e,0x82 = vcvtp.s32.f32 q4, q15
23 0xba,0xff,0x48,0x24 = vrintn.f32 q1, q4
29 0xfa,0xff,0xc8,0x25 = vrintz.f32 q9, q4
31 0xba,0xff,0xc8,0x26 = vrintm.f32 q1, q4
33 0xba,0xff,0xc8,0x27 = vrintp.f32 q1, q4
37 0xfa,0xff,0xc8,0x25 = vrintz.f32 q9, q4
38 0xba,0xff,0xc8,0x27 = vrintp.f32 q1, q4
    [all...]
neon-minmax-encoding.s.cs 17 0x4c,0x86,0x1a,0xf2 = vmax.s16 q4, q5, q6
27 0x4a,0x86,0x18,0xf3 = vmax.u16 q4, q4, q5 package
45 0x5c,0x86,0x1a,0xf2 = vmin.s16 q4, q5, q6
55 0x5a,0x86,0x18,0xf3 = vmin.u16 q4, q4, q5 package
neon-mul-encoding.s.cs 56 0x63,0x88,0xa8,0xf3 = vmul.i32 q4, q4, d3[1] package
70 0x63,0xa8,0xa8,0xf3 = vmul.i32 q5, q4, d3[1] package
71 0x44,0x88,0xaa,0xf3 = vmul.i32 q4, q5, d4[0]
neont2-minmax-encoding.s.cs 17 0x1a,0xef,0x4c,0x86 = vmax.s16 q4, q5, q6
27 0x18,0xff,0x4a,0x86 = vmax.u16 q4, q4, q5 package
45 0x1a,0xef,0x5c,0x86 = vmin.s16 q4, q5, q6
55 0x18,0xff,0x5a,0x86 = vmin.u16 q4, q4, q5 package
neon-add-encoding.s.cs 41 0x62,0x80,0x08,0xf3 = vhadd.u8 q4, q4, q9 package
neon-sub-encoding.s.cs 20 0x4e,0x88,0x38,0xf3 = vsub.i64 q4, q4, q7 package
74 0x62,0x82,0x08,0xf3 = vhsub.u8 q4, q4, q9 package
  /external/flac/libFLAC/
lpc_intrin_avx2.c 64 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; local
69 q4 = _mm256_set1_epi32(0xffff & qlp_coeff[4 ]);
87 mull = _mm256_madd_epi16(q4, _mm256_loadu_si256((const __m256i*)(data+i-5 ))); summ = _mm256_add_epi32(summ, mull);
97 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10; local
102 q4 = _mm256_set1_epi32(0xffff & qlp_coeff[4 ]);
118 mull = _mm256_madd_epi16(q4, _mm256_loadu_si256((const __m256i*)(data+i-5 ))); summ = _mm256_add_epi32(summ, mull);
130 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9; local
135 q4 = _mm256_set1_epi32(0xffff & qlp_coeff[4 ]);
149 mull = _mm256_madd_epi16(q4, _mm256_loadu_si256((const __m256i*)(data+i-5 ))); summ = _mm256_add_epi32(summ, mull);
159 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8 local
190 __m256i q0, q1, q2, q3, q4, q5, q6, q7; local
215 __m256i q0, q1, q2, q3, q4, q5, q6; local
240 __m256i q0, q1, q2, q3, q4, q5; local
261 __m256i q0, q1, q2, q3, q4; local
419 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; local
452 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10; local
485 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9; local
514 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8; local
545 __m256i q0, q1, q2, q3, q4, q5, q6, q7; local
570 __m256i q0, q1, q2, q3, q4, q5, q6; local
595 __m256i q0, q1, q2, q3, q4, q5; local
616 __m256i q0, q1, q2, q3, q4; local
778 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; local
811 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10; local
844 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9; local
873 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8; local
904 __m256i q0, q1, q2, q3, q4, q5, q6, q7; local
929 __m256i q0, q1, q2, q3, q4, q5, q6; local
954 __m256i q0, q1, q2, q3, q4, q5; local
975 __m256i q0, q1, q2, q3, q4; local
    [all...]
lpc_intrin_sse2.c 70 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; local
75 q4 = _mm_cvtsi32_si128(0xffff & qlp_coeff[4]); q4 = _mm_shuffle_epi32(q4, _MM_SHUFFLE(0,0,0,0));
93 mull = _mm_madd_epi16(q4, _mm_loadu_si128((const __m128i*)(data+i-5))); summ = _mm_add_epi32(summ, mull);
103 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10; local
108 q4 = _mm_cvtsi32_si128(0xffff & qlp_coeff[4]); q4 = _mm_shuffle_epi32(q4, _MM_SHUFFLE(0,0,0,0));
124 mull = _mm_madd_epi16(q4, _mm_loadu_si128((const __m128i*)(data+i-5))); summ = _mm_add_epi32(summ, mull)
136 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9; local
165 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8; local
196 __m128i q0, q1, q2, q3, q4, q5, q6, q7; local
221 __m128i q0, q1, q2, q3, q4, q5, q6; local
246 __m128i q0, q1, q2, q3, q4, q5; local
267 __m128i q0, q1, q2, q3, q4; local
    [all...]
lpc_intrin_sse41.c 971 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; local
1004 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10; local
1037 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9; local
1066 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8; local
1097 __m128i q0, q1, q2, q3, q4, q5, q6, q7; local
1122 __m128i q0, q1, q2, q3, q4, q5, q6; local
1147 __m128i q0, q1, q2, q3, q4, q5; local
1168 __m128i q0, q1, q2, q3, q4; local
    [all...]
  /external/libvpx/libvpx/vp8/common/arm/neon/
dequant_idct_neon.c 27 int16x8_t q1, q2, q3, q4, q5, q6; local
38 q4 = vld1q_s16(input);
59 vmulq_u16(vreinterpretq_u16_s16(q4), vreinterpretq_u16_s16(q6)));
67 q4 = vqdmulhq_n_s16(q2, cospi8sqrt2minus1);
69 q4 = vshrq_n_s16(q4, 1);
71 q4 = vqaddq_s16(q4, q2);
73 d10 = vqsub_s16(vget_low_s16(q3), vget_high_s16(q4));
74 d11 = vqadd_s16(vget_high_s16(q3), vget_low_s16(q4));
    [all...]
idct_dequant_full_2x_neon.c 21 int16x8_t q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; local
41 q4 = vld1q_s16(q);
68 q4 = vmulq_s16(q4, q0);
74 dLow1 = vget_low_s16(q4);
75 dHigh1 = vget_high_s16(q4);
77 q4 = vcombine_s16(dHigh0, dHigh1);
86 q6 = vqdmulhq_n_s16(q4, sinpi8sqrt2);
88 q8 = vqdmulhq_n_s16(q4, cospi8sqrt2minus1);
97 q4 = vqaddq_s16(q4, q8)
    [all...]
vp8_loopfilter_neon.c 19 uint8x16_t q4, // p2
36 q11u8 = vabdq_u8(q3, q4);
37 q12u8 = vabdq_u8(q4, q5);
41 q4 = vabdq_u8(q10, q9);
45 q3 = vmaxq_u8(q3, q4);
129 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local
139 q4 = vld1q_u8(src);
153 vp8_loop_filter_neon(qblimit, qlimit, qthresh, q3, q4, q5, q6, q7, q8, q9,
172 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local
216 q4 = vcombine_u8(d8, d9)
320 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local
432 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local
    [all...]
mbloopfilter_neon.c 18 uint8x16_t q4, // p2
38 q11u8 = vabdq_u8(q3, q4);
39 q12u8 = vabdq_u8(q4, q5);
70 q4 = veorq_u8(q4, q0u8);
137 q0s8 = vqaddq_s8(vreinterpretq_s8_u8(q4), q0s8);
157 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local
168 q4 = vld1q_u8(src);
182 vp8_mbloop_filter_neon(qblimit, qlimit, qthresh, q3, q4, q5, q6, q7, q8, q9,
183 q10, &q4, &q5, &q6, &q7, &q8, &q9)
205 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local
292 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local
456 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local
    [all...]
  /external/eigen/test/
geo_quaternion.cpp 208 Quaternionx q4 = MCQuaternionUA(array3unaligned); local
212 VERIFY_IS_APPROX(q4.coeffs(), q3.coeffs());
  /external/guava/guava/src/com/google/common/cache/
Striped64.java 100 volatile long q0, q1, q2, q3, q4, q5, q6; field in class:Striped64.Cell
  /external/libvpx/libvpx/vpx_dsp/
loopfilter.c 64 uint8_t q4) {
67 mask |= (abs(q4 - q0) > thresh) * -1;
243 const uint8_t q0 = *oq0, q1 = *oq1, q2 = *oq2, q3 = *oq3, q4 = *oq4, local
256 p7 * 3 + p6 + p5 + p4 + p3 + p2 * 2 + p1 + p0 + q0 + q1 + q2 + q3 + q4,
259 q0 + q1 + q2 + q3 + q4 + q5,
262 q1 + q2 + q3 + q4 + q5 + q6,
265 q2 + q3 + q4 + q5 + q6 + q7,
268 q3 + q4 + q5 + q6 + q7 * 2,
271 p4 + p3 + p2 + p1 + p0 + q0 + q1 + q2 * 2 + q3 + q4 + q5 + q6 + q7 * 3,
274 p3 + p2 + p1 + p0 + q0 + q1 + q2 + q3 * 2 + q4 + q5 + q6 + q7 * 4, 4)
596 const uint16_t q4 = *oq4; local
    [all...]
  /external/libvpx/libvpx/vpx_dsp/mips/
loopfilter_filters_dspr2.h 514 const uint32_t q4 = *oq4, q5 = *oq5, q6 = *oq6, q7 = *oq7; local
522 /* addition of p6,p5,p4,p3,p2,p1,p0,q0,q1,q2,q3,q4,q5,q6
534 "addu.ph %[add_p6toq6], %[add_p6toq6], %[q4] \n\t"
542 [q3] "r"(q3), [q4] "r"(q4), [q5] "r"(q5), [q6] "r"(q6),
555 "subu.ph %[res_op6], %[res_op6], %[q4] \n\t"
569 "subu.ph %[res_op5], %[res_op5], %[q4] \n\t"
581 "subu.ph %[res_op4], %[res_op4], %[q4] \n\t"
591 "subu.ph %[res_op3], %[res_op3], %[q4] \n\t"
597 p0 + q0 + q1 + q2 + q3 + q4, 4) *
    [all...]
  /external/libjpeg-turbo/simd/
jsimd_arm_neon.S 111 JLONG q1, q2, q3, q4, q5, q6, q7; \
125 q4 = row5 + row1; \ define
127 MULTIPLY(q4, FIX_1_175875602); \
129 MULTIPLY(q4, FIX_1_175875602_MINUS_0_390180644); \
132 q4 = q6; \ define
138 q4 += MULTIPLY(row7, FIX_0_298631336_MINUS_0_899976223) + \
165 tmp0 = q4; \
279 vmov q4, q6
285 vmlsl.s16 q4, ROW1L, XFIX_0_899976223
320 vmlal.s16 q4, ROW7L, XFIX_0_298631336_MINUS_0_89997622
    [all...]

Completed in 459 milliseconds

1 2 3 4 5 6