/external/flac/libFLAC/ |
lpc_intrin_avx2.c | 64 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; local 70 q5 = _mm256_set1_epi32(0xffff & qlp_coeff[5 ]); 86 mull = _mm256_madd_epi16(q5, _mm256_loadu_si256((const __m256i*)(data+i-6 ))); summ = _mm256_add_epi32(summ, mull); 97 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10; local 103 q5 = _mm256_set1_epi32(0xffff & qlp_coeff[5 ]); 117 mull = _mm256_madd_epi16(q5, _mm256_loadu_si256((const __m256i*)(data+i-6 ))); summ = _mm256_add_epi32(summ, mull); 130 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9; local 136 q5 = _mm256_set1_epi32(0xffff & qlp_coeff[5 ]); 148 mull = _mm256_madd_epi16(q5, _mm256_loadu_si256((const __m256i*)(data+i-6 ))); summ = _mm256_add_epi32(summ, mull); 159 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8 local 190 __m256i q0, q1, q2, q3, q4, q5, q6, q7; local 215 __m256i q0, q1, q2, q3, q4, q5, q6; local 240 __m256i q0, q1, q2, q3, q4, q5; local 419 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; local 452 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10; local 485 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9; local 514 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8; local 545 __m256i q0, q1, q2, q3, q4, q5, q6, q7; local 570 __m256i q0, q1, q2, q3, q4, q5, q6; local 595 __m256i q0, q1, q2, q3, q4, q5; local 778 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; local 811 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10; local 844 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9; local 873 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8; local 904 __m256i q0, q1, q2, q3, q4, q5, q6, q7; local 929 __m256i q0, q1, q2, q3, q4, q5, q6; local 954 __m256i q0, q1, q2, q3, q4, q5; local [all...] |
lpc_intrin_sse2.c | 70 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; local 76 q5 = _mm_cvtsi32_si128(0xffff & qlp_coeff[5]); q5 = _mm_shuffle_epi32(q5, _MM_SHUFFLE(0,0,0,0)); 92 mull = _mm_madd_epi16(q5, _mm_loadu_si128((const __m128i*)(data+i-6))); summ = _mm_add_epi32(summ, mull); 103 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10; local 109 q5 = _mm_cvtsi32_si128(0xffff & qlp_coeff[5]); q5 = _mm_shuffle_epi32(q5, _MM_SHUFFLE(0,0,0,0)); 123 mull = _mm_madd_epi16(q5, _mm_loadu_si128((const __m128i*)(data+i-6))); summ = _mm_add_epi32(summ, mull) 136 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9; local 165 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8; local 196 __m128i q0, q1, q2, q3, q4, q5, q6, q7; local 221 __m128i q0, q1, q2, q3, q4, q5, q6; local 246 __m128i q0, q1, q2, q3, q4, q5; local [all...] |
lpc_intrin_sse41.c | 971 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; local 1004 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10; local 1037 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9; local 1066 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8; local 1097 __m128i q0, q1, q2, q3, q4, q5, q6, q7; local 1122 __m128i q0, q1, q2, q3, q4, q5, q6; local 1147 __m128i q0, q1, q2, q3, q4, q5; local [all...] |
/external/libvpx/libvpx/vp8/common/arm/neon/ |
dequant_idct_neon.c | 27 int16x8_t q1, q2, q3, q4, q5, q6; local 42 q5 = vld1q_s16(dq); 57 vmulq_u16(vreinterpretq_u16_s16(q3), vreinterpretq_u16_s16(q5)));
|
idct_dequant_full_2x_neon.c | 21 int16x8_t q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; local 44 q5 = vld1q_s16(q); 69 q5 = vmulq_s16(q5, q1); 81 dLow1 = vget_low_s16(q5); 82 dHigh1 = vget_high_s16(q5); 84 q5 = vcombine_s16(dHigh0, dHigh1); 87 q7 = vqdmulhq_n_s16(q5, sinpi8sqrt2); 89 q9 = vqdmulhq_n_s16(q5, cospi8sqrt2minus1); 98 q5 = vqaddq_s16(q5, q9) [all...] |
vp8_loopfilter_neon.c | 20 uint8x16_t q5, // p1 37 q12u8 = vabdq_u8(q4, q5); 38 q13u8 = vabdq_u8(q5, q6); 55 q2u8 = vabdq_u8(q5, q8); 66 q5 = veorq_u8(q5, q10); 80 q1s8 = vqsubq_s8(vreinterpretq_s8_u8(q5), vreinterpretq_s8_u8(q8)); 114 q13s8 = vqaddq_s8(vreinterpretq_s8_u8(q5), q1s8); 130 uint8x16_t q5, q6, q7, q8, q9, q10; local 141 q5 = vld1q_u8(src) 173 uint8x16_t q5, q6, q7, q8, q9, q10; local 321 uint8x16_t q5, q6, q7, q8, q9, q10; local 433 uint8x16_t q5, q6, q7, q8, q9, q10; local [all...] |
mbloopfilter_neon.c | 19 uint8x16_t q5, // p1 39 q12u8 = vabdq_u8(q4, q5); 40 q13u8 = vabdq_u8(q5, q6); 59 q1u8 = vabdq_u8(q5, q8); 69 q5 = veorq_u8(q5, q0u8); 83 q1s8 = vqsubq_s8(vreinterpretq_s8_u8(q5), vreinterpretq_s8_u8(q8)); 139 q12s8 = vqaddq_s8(vreinterpretq_s8_u8(q5), q12s8); 158 uint8x16_t q5, q6, q7, q8, q9, q10; local 170 q5 = vld1q_u8(src) 206 uint8x16_t q5, q6, q7, q8, q9, q10; local 293 uint8x16_t q5, q6, q7, q8, q9, q10; local 457 uint8x16_t q5, q6, q7, q8, q9, q10; local [all...] |
/external/guava/guava/src/com/google/common/cache/ |
Striped64.java | 100 volatile long q0, q1, q2, q3, q4, q5, q6; field in class:Striped64.Cell
|
/external/capstone/suite/MC/ARM/ |
neon-bitwise-encoding.s.cs | 92 0x5a,0xc1,0x0c,0xf2 = vand q6, q6, q5 93 0x5a,0xc1,0x0c,0xf2 = vand q6, q6, q5 97 0x5a,0xc1,0x0c,0xf3 = veor q6, q6, q5 98 0x5a,0xc1,0x0c,0xf3 = veor q6, q6, q5 102 0x5a,0xc1,0x0c,0xf3 = veor q6, q6, q5 103 0x5a,0xc1,0x0c,0xf3 = veor q6, q6, q5 107 0x4a,0xa2,0xb5,0xf3 = vclt.s16 q5, q5, #0 package 109 0x56,0xa8,0x1a,0xf3 = vceq.i16 q5, q5, q package 111 0x46,0xa3,0x1a,0xf2 = vcgt.s16 q5, q5, q3 package 113 0x56,0xa3,0x1a,0xf2 = vcge.s16 q5, q5, q3 package 115 0x4a,0xa0,0xb5,0xf3 = vcgt.s16 q5, q5, #0 package 117 0xca,0xa0,0xb5,0xf3 = vcge.s16 q5, q5, #0 package 119 0x4a,0xa1,0xb5,0xf3 = vceq.i16 q5, q5, #0 package 121 0xca,0xa1,0xb5,0xf3 = vcle.s16 q5, q5, #0 package 124 0x56,0xae,0x0a,0xf3 = vacge.f32 q5, q5, q3 package 126 0x56,0xae,0x2a,0xf3 = vacgt.f32 q5, q5, q3 package [all...] |
neon-minmax-encoding.s.cs | 17 0x4c,0x86,0x1a,0xf2 = vmax.s16 q4, q5, q6 package 22 0x42,0x2f,0x4a,0xf2 = vmax.f32 q9, q5, q1 package 24 0x4c,0xa6,0x1a,0xf2 = vmax.s16 q5, q5, q6 package 27 0x4a,0x86,0x18,0xf3 = vmax.u16 q4, q4, q5 45 0x5c,0x86,0x1a,0xf2 = vmin.s16 q4, q5, q6 package 50 0x42,0x2f,0x6a,0xf2 = vmin.f32 q9, q5, q1 package 52 0x5c,0xa6,0x1a,0xf2 = vmin.s16 q5, q5, q6 package 55 0x5a,0x86,0x18,0xf3 = vmin.u16 q4, q4, q5 [all...] |
neon-mul-encoding.s.cs | 57 0x44,0xa8,0xaa,0xf3 = vmul.i32 q5, q5, d4[0] package 70 0x63,0xa8,0xa8,0xf3 = vmul.i32 q5, q4, d3[1] 71 0x44,0x88,0xaa,0xf3 = vmul.i32 q4, q5, d4[0] package
|
neont2-minmax-encoding.s.cs | 17 0x1a,0xef,0x4c,0x86 = vmax.s16 q4, q5, q6 package 22 0x4a,0xef,0x42,0x2f = vmax.f32 q9, q5, q1 package 24 0x1a,0xef,0x4c,0xa6 = vmax.s16 q5, q5, q6 package 27 0x18,0xff,0x4a,0x86 = vmax.u16 q4, q4, q5 45 0x1a,0xef,0x5c,0x86 = vmin.s16 q4, q5, q6 package 50 0x6a,0xef,0x42,0x2f = vmin.f32 q9, q5, q1 package 52 0x1a,0xef,0x5c,0xa6 = vmin.s16 q5, q5, q6 package 55 0x18,0xff,0x5a,0x86 = vmin.u16 q4, q4, q5 [all...] |
neon-add-encoding.s.cs | 42 0x60,0xa0,0x1a,0xf3 = vhadd.u16 q5, q5, q8 package 110 0x4a,0xc8,0x0c,0xf2 = vadd.i8 q6, q6, q5
|
neon-sub-encoding.s.cs | 21 0x4c,0xad,0x2a,0xf2 = vsub.f32 q5, q5, q6 package 75 0x60,0xa2,0x1a,0xf3 = vhsub.u16 q5, q5, q8 package
|
neon-shift-encoding.s.cs | 57 0xda,0x81,0x81,0xf2 = vsra.s64 q4, q5, #63 package 63 0x5a,0xa1,0x91,0xf2 = vsra.s16 q5, q5, #15 package 73 0xda,0x81,0x81,0xf3 = vsra.u64 q4, q5, #63 package 79 0x5a,0xa1,0x91,0xf3 = vsra.u16 q5, q5, #15 package 89 0xda,0x84,0x81,0xf3 = vsri.64 q4, q5, #63 package 95 0x5a,0xa4,0x91,0xf3 = vsri.16 q5, q5, #15 package 105 0xda,0x85,0xbf,0xf3 = vsli.64 q4, q5, #6 package 111 0x5a,0xa5,0x9f,0xf3 = vsli.16 q5, q5, #15 package 203 0xca,0xa5,0x08,0xf3 = vrshl.u8 q5, q5, q12 package 219 0x5a,0xa0,0x88,0xf3 = vshr.u8 q5, q5, #8 package 235 0x5a,0xa2,0x88,0xf3 = vrshr.u8 q5, q5, #8 package [all...] |
/external/libvpx/libvpx/vpx_dsp/ |
loopfilter.c | 244 q5 = *oq5, q6 = *oq6, q7 = *oq7; local 259 q0 + q1 + q2 + q3 + q4 + q5, 262 q1 + q2 + q3 + q4 + q5 + q6, 265 q2 + q3 + q4 + q5 + q6 + q7, 268 q3 + q4 + q5 + q6 + q7 * 2, 271 p4 + p3 + p2 + p1 + p0 + q0 + q1 + q2 * 2 + q3 + q4 + q5 + q6 + q7 * 3, 274 p3 + p2 + p1 + p0 + q0 + q1 + q2 + q3 * 2 + q4 + q5 + q6 + q7 * 4, 4); 276 p2 + p1 + p0 + q0 + q1 + q2 + q3 + q4 * 2 + q5 + q6 + q7 * 5, 4); 278 p1 + p0 + q0 + q1 + q2 + q3 + q4 + q5 * 2 + q6 + q7 * 6, 4); 280 p0 + q0 + q1 + q2 + q3 + q4 + q5 + q6 * 2 + q7 * 7, 4) 597 const uint16_t q5 = *oq5; local [all...] |
/external/libvpx/libvpx/vpx_dsp/mips/ |
loopfilter_filters_dspr2.h | 514 const uint32_t q4 = *oq4, q5 = *oq5, q6 = *oq6, q7 = *oq7; local 522 /* addition of p6,p5,p4,p3,p2,p1,p0,q0,q1,q2,q3,q4,q5,q6 535 "addu.ph %[add_p6toq6], %[add_p6toq6], %[q5] \n\t" 542 [q3] "r"(q3), [q4] "r"(q4), [q5] "r"(q5), [q6] "r"(q6), 556 "subu.ph %[res_op6], %[res_op6], %[q5] \n\t" 570 "subu.ph %[res_op5], %[res_op5], %[q5] \n\t" 582 "subu.ph %[res_op4], %[res_op4], %[q5] \n\t" 592 "subu.ph %[res_op3], %[res_op3], %[q5] \n\t" 602 "subu.ph %[res_op2], %[res_op2], %[q5] \n\t [all...] |
loopfilter_mb_horiz_dspr2.c | 32 uint32_t p7, p6, p5, p4, p3, p2, p1, p0, q0, q1, q2, q3, q4, q5, q6, q7; local 97 "lw %[q5], (%[sq5]) \n\t" 102 [q7] "=&r"(q7), [q6] "=&r"(q6), [q5] "=&r"(q5), [q4] "=&r"(q4) 109 flatmask5(p7, p6, p5, p4, p0, q0, q4, q5, q6, q7, &flat2); 161 "sw %[q5], (%[sq5]) \n\t" 169 : [q6] "r"(q6), [q5] "r"(q5), [q4] "r"(q4), [q3] "r"(q3),
|
loopfilter_mb_vert_dspr2.c | 30 uint32_t p7, p6, p5, p4, p3, p2, p1, p0, q0, q1, q2, q3, q4, q5, q6, q7; local 83 "lw %[q5], +4(%[s3]) \n\t" 87 [q7] "=&r"(q7), [q6] "=&r"(q6), [q5] "=&r"(q5), [q4] "=&r"(q4) 198 /* transpose q4, q5, q6, q7 203 q5 q5_0 q5_1 q5_2 q5_3 210 q5 q4_1 q5_1 q26_1 q7_1 216 "precrq.qb.ph %[prim3], %[q5], %[q4] \n\t" 217 "precr.qb.ph %[prim4], %[q5], %[q4] \n\t" 225 "precrq.ph.w %[q5], %[q4], %[sec4] \n\t [all...] |
/external/libjpeg-turbo/simd/ |
jsimd_arm_neon.S | 111 JLONG q1, q2, q3, q4, q5, q6, q7; \ 124 q5 = row7 + row3; \ define 126 q6 = MULTIPLY(q5, FIX_1_175875602_MINUS_1_961570560) + \ 128 q7 = MULTIPLY(q5, FIX_1_175875602) + \ 140 q5 = q7; \ define 150 q5 += MULTIPLY(row5, FIX_2_053119869_MINUS_2_562915447) + \ 166 tmp1 = q5; \ 289 vmov q5, q7 298 vmlal.s16 q5, ROW5L, XFIX_2_053119869_MINUS_2_562915447 300 vmlsl.s16 q5, ROW3L, XFIX_2_56291544 [all...] |
/external/libvpx/libvpx/vpx_dsp/arm/ |
highbd_loopfilter_neon.c | 196 const uint16x8_t q3, const uint16x8_t q4, const uint16x8_t q5, 218 *op1 = apply_15_tap_filter_kernel(flat2, p7, p2, p1, q5, *op1, &sum); 225 *oq5 = apply_15_tap_filter_kernel(flat2, p2, q4, q5, q7, q5, &sum); 226 *oq6 = apply_15_tap_filter_kernel(flat2, p1, q5, q6, q7, q6, &sum); 320 const uint16x8_t q4, const uint16x8_t q5, const uint16x8_t q6, 338 q4, q5, q6, q7, op6, op5, op4, op3, op2, op1, op0, 542 const uint16x8_t q4, const uint16x8_t q5, 553 vst1q_u16(s + 5 * p, q5); 670 q3, q4, q5, q6, q7, op6, op5, op4, op3, op2, op1, op0, oq0, oq1, oq2, oq3 local 693 q3, q4, q5, q6, q7, op6, op5, op4, op3, op2, op1, op0, oq0, oq1, oq2, oq3, local [all...] |
loopfilter_neon.c | 312 const uint8x8_t q3, const uint8x8_t q4, const uint8x8_t q5, 334 *op1 = apply_15_tap_filter_8_kernel(flat2, p7, p2, p1, q5, *op1, &sum); 341 *oq5 = apply_15_tap_filter_8_kernel(flat2, p2, q4, q5, q7, q5, &sum); 342 *oq6 = apply_15_tap_filter_8_kernel(flat2, p1, q5, q6, q7, q6, &sum); 350 const uint8x16_t q3, const uint8x16_t q4, const uint8x16_t q5, 387 apply_15_tap_filter_16_kernel(flat2, p7, p2, p1, q5, *op1, &sum0, &sum1); 398 *oq5 = apply_15_tap_filter_16_kernel(flat2, p2, q4, q5, q7, q5, &sum0, &sum1); 399 *oq6 = apply_15_tap_filter_16_kernel(flat2, p1, q5, q6, q7, q6, &sum0, &sum1) 980 uint8x8_t p7, p6, p5, p4, p3, p2, p1, p0, q0, q1, q2, q3, q4, q5, q6, q7, op6, local 997 uint8x16_t p7, p6, p5, p4, p3, p2, p1, p0, q0, q1, q2, q3, q4, q5, q6, q7, local 1020 uint8x8_t p7, p6, p5, p4, p3, p2, p1, p0, q0, q1, q2, q3, q4, q5, q6, q7, op6, local 1052 uint8x16_t p7, p6, p5, p4, p3, p2, p1, p0, q0, q1, q2, q3, q4, q5, q6, q7, local [all...] |
/external/tensorflow/tensorflow/core/grappler/costs/ |
graph_properties_test.cc | 275 auto q5 = ops::RandomShuffleQueue( local 283 ops::QueueEnqueue(root.WithOpName("Enqueue5"), q5, {rnd, rnd2, rnd3}); 285 root.WithOpName("Dequeue5"), q5, [all...] |
/external/clang/test/SemaCXX/ |
overload-call.cpp | 205 double* q5 = quals_rank1(ap); local
|
/external/libvpx/libvpx/vp9/encoder/x86/ |
vp9_dct_ssse3.c | 90 const __m128i q5 = _mm_sub_epi16(in2, in5); local 146 const __m128i d0 = _mm_sub_epi16(q6, q5); 147 const __m128i d1 = _mm_add_epi16(q6, q5);
|