/prebuilts/sdk/renderscript/include/ |
rs_quaternion.rsh | 70 * q0: First quaternion. 75 rsQuaternionDot(const rs_quaternion* q0, const rs_quaternion* q1) { 76 return q0->w*q1->w + q0->x*q1->x + q0->y*q1->y + q0->z*q1->z; 266 * q0: First input quaternion. 272 rsQuaternionSlerp(rs_quaternion* q, const rs_quaternion* q0, const rs_quaternion* q1, float t) { 274 rsQuaternionSet(q, q0); 283 rsQuaternionSet(&tempq0, q0); [all...] |
/external/libvpx/libvpx/vpx_dsp/x86/ |
loopfilter_sse2.c | 69 // mask |= (abs(p0 - q0) * 2 + abs(p1 - q1) / 2 > blimit) * -1; 72 // mask |= (abs(q1 - q0) > limit) * -1; 398 __m128i p4, p3, p2, p1, p0, q0, q1, q2, q3, q4; local 413 q0 = _mm_loadu_si128((__m128i *)(s - 0 * p)); 424 const __m128i abs_q1q0 = abs_diff(q1, q0); 427 __m128i abs_p0q0 = abs_diff(p0, q0); 436 // mask |= (abs(p0 - q0) * 2 + abs(p1 - q1) / 2 > blimit) * -1; 439 // mask |= (abs(q1 - q0) > limit) * -1; 450 work = _mm_max_epu8(abs_diff(p2, p0), abs_diff(q2, q0)); 452 work = _mm_max_epu8(abs_diff(p3, p0), abs_diff(q3, q0)); 745 __m128i p3, p2, p1, p0, q0, q1, q2, q3; local 971 __m128i p3, p2, p1, p0, q0, q1, q2, q3; local 1210 __m128i p3, p2, p1, p0, q0, q1, q2, q3; local [all...] |
/frameworks/rs/api/ |
rs_quaternion.spec | 56 arg: const rs_quaternion* q0, "First quaternion." 62 return q0->w*q1->w + q0->x*q1->x + q0->y*q1->y + q0->z*q1->z; 238 arg: const rs_quaternion* q0, "First input quaternion." 246 rsQuaternionSet(q, q0); 255 rsQuaternionSet(&tempq0, q0); 258 float angle = rsQuaternionDot(q0, q1); 305 arg: const rs_quaternion* q0 [all...] |
/external/boringssl/linux-arm/crypto/modes/ |
ghash-armv4.S | 375 vrev64.8 q0,q0 386 veor q3,q0 @ inp^=Xi 391 vmull.p8 q0, d26, d0 @ E = A*B1 397 veor q8, q8, q0 @ L = E + F 401 vmull.p8 q0, d26, d0 @ I = A*B3 408 veor q10, q10, q0 @ N = I + J 418 vmull.p8 q0, d26, d6 @ D = A*B 423 veor q0, q0, q [all...] |
/external/libvpx/libvpx/vpx_dsp/arm/ |
sad_neon.c | 87 uint8x16_t q0, q4; local 94 q0 = vld1q_u8(src_ptr); 98 q12 = vabdl_u8(vget_low_u8(q0), vget_low_u8(q4)); 99 q13 = vabdl_u8(vget_high_u8(q0), vget_high_u8(q4)); 102 q0 = vld1q_u8(src_ptr); 106 q12 = vabal_u8(q12, vget_low_u8(q0), vget_low_u8(q4)); 107 q13 = vabal_u8(q13, vget_high_u8(q0), vget_high_u8(q4));
|
/frameworks/rs/cpu_ref/ |
rsCpuIntrinsics_neon_3DLUT.S | 142 /* q0,q1,q2,q5 source data 147 vmovl.u8 q0, d0 150 vmul.u16 q0, q0, d6[0] 159 vsra.u16 q0, q0, #8 163 vshr.u16 q12, q0, #8 167 vbic.u16 q0, #0xff00 171 /* q0,d2,q2 fractional offset
|
rsCpuIntrinsics_neon_YuvToRGB.S | 45 vaddw.u8 q0, q1, d14 // r0 = y0 * 149 + (v >> 1) 57 vhadd.u16 q0, q11 // r0 = (r0 + r2) >> 1 64 vqsub.u16 q0, q13 // r0 = satu16(r0 - (16 * 149 + (128 >> 1) + 128 * 204) >> 1) 71 vqrshrn.u16 d0, q0, #6 78 vzip.u8 q0, q1 196 vzip.8 q0, q2 198 vzip.8 q0, q1
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/common/x86/ |
vp9_loopfilter_intrin_avx2.c | 66 // mask |= (abs(p0 - q0) * 2 + abs(p1 - q1) / 2 > blimit) * -1; 69 // mask |= (abs(q1 - q0) > limit) * -1; 402 __m128i p4, p3, p2, p1, p0, q0, q1, q2, q3, q4; local 417 q0 = _mm_loadu_si128((__m128i *) (s - 0 * p)); 426 const __m128i abs_q1q0 = _mm_or_si128(_mm_subs_epu8(q1, q0), 427 _mm_subs_epu8(q0, q1)); 430 __m128i abs_p0q0 = _mm_or_si128(_mm_subs_epu8(p0, q0), 431 _mm_subs_epu8(q0, p0)); 443 // mask |= (abs(p0 - q0) * 2 + abs(p1 - q1) / 2 > blimit) * -1; 446 // mask |= (abs(q1 - q0) > limit) * -1 [all...] |
vp9_loopfilter_intrin_sse2.c | 66 // mask |= (abs(p0 - q0) * 2 + abs(p1 - q1) / 2 > blimit) * -1; 69 // mask |= (abs(q1 - q0) > limit) * -1; 388 __m128i p4, p3, p2, p1, p0, q0, q1, q2, q3, q4; local 397 q0 = _mm_loadu_si128((__m128i *)(s - 0 * p)); 412 _mm_store_si128((__m128i *)&aq[0 * 16], q0); 418 const __m128i abs_q1q0 = _mm_or_si128(_mm_subs_epu8(q1, q0), 419 _mm_subs_epu8(q0, q1)); 422 __m128i abs_p0q0 = _mm_or_si128(_mm_subs_epu8(p0, q0), 423 _mm_subs_epu8(q0, p0)); 435 // mask |= (abs(p0 - q0) * 2 + abs(p1 - q1) / 2 > blimit) * -1 874 __m128i p3, p2, p1, p0, q0, q1, q2, q3; local 1107 __m128i p3, p2, p1, p0, q0, q1, q2, q3; local 1346 __m128i p3, p2, p1, p0, q0, q1, q2, q3; local [all...] |
/external/libavc/common/arm/ |
ih264_iquant_itrans_recon_a9.s | 147 vmull.s16 q0, d16, d20 @ Q0 = p[i] = (x[i] * trns_coeff[i]) where i = 0..3 152 vshl.s32 q0, q0, q15 @ Q0 = q[i] = (p[i] << (qP/6)) where i = 0..3 157 vqrshrn.s32 d0, q0, #0x4 @ D0 = c[i] = ((q[i] + 32) >> 4) where i = 0..3 168 vadd.s16 d4, d0, d2 @x0 = q0 + q1; 170 vsub.s16 d5, d0, d2 @x1 = q0 - q1; 172 vshr.s16 d8, d1, #1 @q0>>1 175 vsub.s16 d6, d8, d3 @x2 = (q0 >> 1) - q1 [all...] |
ih264_inter_pred_luma_horz_hpel_vert_qpel_a9q.s | 151 vld1.32 {q0}, [r0], r2 @ row -2 load for horizontal filter 162 vld1.32 {q0}, [r0], r2 @ row -1 load for horizontal filter 176 vld1.32 {q0}, [r0], r2 @ row 0 load for horizontal filter 190 vld1.32 {q0}, [r0], r2 @ row 1 load for horizontal filter 204 vld1.32 {q0}, [r0], r2 @ row 2 load for horizontal filter 222 vld1.32 {q0}, [r0], r2 @ row 3 load for horizontal filter 237 vld1.32 {q0}, [r0], r2 @ row 4 load for hoorizontal filter 268 vld1.32 {q0}, [r0], r2 @ row 5 load for horizontal filter 306 vld1.32 {q0}, [r0], r2 @ row 6 load for horizontal filter 380 vld1.32 {q0}, [r8], r [all...] |
ih264_weighted_pred_a9q.s | 128 vdup.16 q0, r9 @Q0 = -log_wd (16-bit) 148 vrshl.s16 q2, q2, q0 @rounds off the weighted samples from rows 1,2 149 vrshl.s16 q3, q3, q0 @rounds off the weighted samples from rows 3,4 182 vrshl.s16 q2, q2, q0 @rounds off the weighted samples from row 1 183 vrshl.s16 q3, q3, q0 @rounds off the weighted samples from row 2 184 vrshl.s16 q4, q4, q0 @rounds off the weighted samples from row 3 186 vrshl.s16 q5, q5, q0 @rounds off the weighted samples from row 4 229 vrshl.s16 q6, q6, q0 @rounds off the weighted samples from row 1L 232 vrshl.s16 q7, q7, q0 @rounds off the weighted samples from row 1 [all...] |
ih264_iquant_itrans_recon_dc_a9.s | 137 asr r6, r6, #4 @q0 = (pi2_src[0]*pu2_iscal_mat[0]*pu2_weigh_mat[0] + rnd_fact)<<(u4_qp_div_6-4) 143 add r6, r6, #32 @i_macro = q0 + 32 145 vdup.s16 q0, r6 @copy transform output to Q0 154 vaddw.u8 q10, q0, d30 156 vaddw.u8 q11, q0, d31 256 asr r6, r6, #6 @q0 = (pi2_src[0]*pu2_iscal_mat[0]*pu2_weigh_mat[0] + rnd_fact)<<(u4_qp_div_6-4) 257 add r6, r6, #32 @i_macro = q0 + 32 259 vdup.s16 q8, r6 @copy transform output to Q0 266 vaddw.u8 q0, q8, d2 [all...] |
/external/libhevc/common/arm/ |
ihevc_weighted_pred_bi_default.s | 134 vmov.i16 q0,#0x40 @tmp_lvl_shift = 1 << (shift - 1) 136 vadd.s16 q0,q0,q2 276 vqadd.s16 q12,q12,q0 @vaddq_s32(i4_tmp1_t1, tmp_lvl_shift_t) 281 vqadd.s16 q11,q11,q0 @vaddq_s32(i4_tmp2_t1, tmp_lvl_shift_t) 287 vqadd.s16 q15,q15,q0 @vaddq_s32(i4_tmp1_t1, tmp_lvl_shift_t) iii iteration 291 vqadd.s16 q4,q4,q0 333 vqadd.s16 q12,q12,q0 @vaddq_s32(i4_tmp1_t1, tmp_lvl_shift_t) 337 vqadd.s16 q11,q11,q0 @vaddq_s32(i4_tmp2_t1, tmp_lvl_shift_t) 409 vqadd.s16 q11,q11,q0 [all...] |
ihevc_intra_pred_luma_vert.s | 195 @vsubl.u8 q0, d26, d22 198 @vshr.s16 q0, q0, #1 202 @vaddl.s8 q0, d24, d26 203 vmovl.s8 q0, d26 205 vqadd.s16 q0, q0, q15 211 vqmovun.s16 d25, q0 213 @vmovn.u16 d25, q0 337 @vaddl.s8 q0, d24, d2 [all...] |
/external/opencv3/3rdparty/libwebp/enc/ |
filter.c | 54 const int p1 = p[-2*step], p0 = p[-step], q0 = p[0], q1 = p[step]; local 55 const int a = 3 * (q0 - p0) + sclip1[1020 + p1 - q1]; 59 p[ 0] = clip1[255 + q0 - a1]; 64 const int p1 = p[-2*step], p0 = p[-step], q0 = p[0], q1 = p[step]; local 65 const int a = 3 * (q0 - p0); 71 p[ 0] = clip1[255 + q0 - a1]; 77 const int p1 = p[-2*step], p0 = p[-step], q0 = p[0], q1 = p[step]; local 78 return (abs0[255 + p1 - p0] > thresh) || (abs0[255 + q1 - q0] > thresh); 82 const int p1 = p[-2*step], p0 = p[-step], q0 = p[0], q1 = p[step]; local 83 return (2 * abs0[255 + p0 - q0] + abs1[255 + p1 - q1]) <= thresh 89 const int q0 = p[0], q1 = p[step], q2 = p[2*step], q3 = p[3*step]; local [all...] |
/external/libavc/common/x86/ |
ih264_iquant_itrans_recon_dc_ssse3.c | 114 WORD32 q0 = pi2_src[0]; local 125 INV_QUANT(q0, pu2_iscal_mat[0], pu2_weigh_mat[0], u4_qp_div_6, rnd_fact, 4); 128 q0 = pi2_dc_ld_addr[0]; // Restoring dc value for intra case 130 i_macro = ((q0 + 32) >> 6); 239 WORD32 q0 = pi2_src[0]; local 252 INV_QUANT(q0, pu2_iscale_mat[0], pu2_weigh_mat[0], qp_div, rnd_fact, 6); 253 i_macro = ((q0 + 32) >> 6); 399 WORD16 q0 = pi2_dc_src[0]; // DC value won't be dequantized for chroma inverse transform local 400 WORD16 i_macro = ((q0 + 32) >> 6);
|
/frameworks/av/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p10/src_gcc/ |
armVCM4P10_InterpolateLuma_HalfDiagHorVer4x4_unsafe_s.S | 144 VADDL.S16 q0,d12,d18 147 VMUL.I32 q4,q0,q14 150 VADDL.S16 q0,d14,d20 153 VMUL.I32 q4,q0,q14 156 VADDL.S16 q0,d16,d22 159 VMUL.I32 q4,q0,q14 162 VADDL.S16 q0,d18,d24 165 VMLS.I32 q3,q0,q14 170 VQMOVN.U16 d0,q0
|
omxVCM4P10_PredictIntraChroma_8x8_s.S | 182 VADDL.U8 q0,d0,d1 183 VDUP.16 q0,d1[3] 184 VSHL.I16 q0,q0,#4 189 VADD.I16 q2,q2,q0 190 VDUP.16 q0,d6[0] 198 VADD.I16 q0,q2,q0 206 VQRSHRUN.S16 d0,q0,#5
|
/external/opencv3/modules/imgproc/src/ |
moments.cpp | 735 double q0 = t0 * t0, q1 = t1 * t1; local 743 hu[3] = q0 + q1; 744 hu[5] = d * (q0 - q1) + n4 * t0 * t1; 746 t0 *= q0 - 3 * q1; 747 t1 *= 3 * q0 - q1; 749 q0 = m.nu30 - 3 * m.nu12; 752 hu[2] = q0 * q0 + q1 * q1; 753 hu[4] = q0 * t0 + q1 * t1; 754 hu[6] = q1 * t0 - q0 * t1 836 double q0 = t0 * t0, q1 = t1 * t1; local [all...] |
/external/boringssl/linux-arm/crypto/sha/ |
sha1-armv4-large.S | 493 vld1.8 {q0,q1},[r1]! @ handles unaligned 497 vrev32.8 q0,q0 @ yes, even on 500 vadd.i32 q8,q0,q14 510 vext.8 q8,q0,q1,#8 521 veor q8,q8,q0 700 veor q0,q0,q8 703 veor q0,q0,q [all...] |
/external/llvm/test/MC/ARM/ |
neon-v8.s | 9 vminnm.f32 q0, q13, q2 10 @ CHECK: vminnm.f32 q0, q13, q2 @ encoding: [0xd4,0x0f,0x2a,0xf3] 54 vrintx.f32 q0, q3 55 @ CHECK: vrintx.f32 q0, q3 @ encoding: [0xc6,0x04,0xba,0xf3] 76 vrintx.f32.f32 q0, q3 77 @ CHECK: vrintx.f32 q0, q3 @ encoding: [0xc6,0x04,0xba,0xf3]
|
thumb-neon-v8.s | 9 vminnm.f32 q0, q13, q2 10 @ CHECK: vminnm.f32 q0, q13, q2 @ encoding: [0x2a,0xff,0xd4,0x0f] 54 vrintx.f32 q0, q3 55 @ CHECK: vrintx.f32 q0, q3 @ encoding: [0xba,0xff,0xc6,0x04] 76 vrintx.f32.f32 q0, q3 77 @ CHECK: vrintx.f32 q0, q3 @ encoding: [0xba,0xff,0xc6,0x04]
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vpx_scale/arm/neon/ |
vp8_vpxyv12_copyframe_func_neon.asm | 60 vld1.8 {q0, q1}, [r8]! 71 vst1.8 {q0, q1}, [r9]! 120 vld1.8 {q0, q1}, [r8]! 127 vst1.8 {q0, q1}, [r9]! 179 vld1.8 {q0}, [r8]! 184 vst1.8 {q0}, [r9]!
|
/system/core/libpixelflinger/ |
col32cb16blend_neon.S | 58 vshll.u8 q0, d0, #5 // shift up red and widen 90 vadd.u16 q8, q8, q0 // add src red to dst red0 97 vadd.u16 q13, q13, q0 // add src red to dst red1
|