/external/apache-commons-math/src/main/java/org/apache/commons/math/ode/nonstiff/ |
MidpointStepInterpolator.java | 86 final double coeff1 = oneMinusThetaH * theta; local 94 interpolatedState[i] = currentState[i] + coeff1 * yDot1 - coeff2 * yDot2;
|
ClassicalRungeKuttaStepInterpolator.java | 92 final double coeff1 = s * ((-fourTheta + 5) * theta - 1); local 103 currentState[i] + coeff1 * yDot1 + coeff23 * yDot23 + coeff4 * yDot4;
|
ThreeEighthesStepInterpolator.java | 93 final double coeff1 = s * (1 - 7 * theta + 2 * fourTheta2); local 108 currentState[i] - coeff1 * yDot1 - coeff2 * yDot2 - coeff3 * yDot3 - coeff4 * yDot4;
|
GillStepInterpolator.java | 102 final double coeff1 = soMt * (1 - fourTheta); local 118 currentState[i] - coeff1 * yDot1 - coeff2 * yDot2 - coeff3 * yDot3 - coeff4 * yDot4;
|
/external/libvpx/libvpx/vp8/encoder/mips/msa/ |
encodeopt_msa.c | 18 v8i16 coeff, dq_coeff, coeff0, coeff1; local 26 ILVRL_H2_SH(coeff, dq_coeff, coeff0, coeff1); 27 HSUB_UH2_SW(coeff0, coeff1, diff0, diff1); 47 v8i16 coeff, coeff0, coeff1, coeff2, coeff3, coeff4; local 79 ILVRL_H2_SH(coeff, dq_coeff, coeff0, coeff1); 80 HSUB_UH2_SW(coeff0, coeff1, diff0, diff1); 83 ILVRL_H2_SH(coeff2, dq_coeff2, coeff0, coeff1); 84 HSUB_UH2_SW(coeff0, coeff1, diff0, diff1); 91 ILVRL_H2_SH(coeff3, dq_coeff3, coeff0, coeff1); 92 HSUB_UH2_SW(coeff0, coeff1, diff0, diff1) 113 v8i16 coeff, coeff0, coeff1, coeff2, coeff3, coeff4; local [all...] |
quantize_msa.c | 24 v8i16 coeff0, coeff1, z0, z1; local 35 LD_SH2(coeff_ptr, 8, coeff0, coeff1); 36 VSHF_H2_SH(coeff0, coeff1, coeff0, coeff1, zigzag_mask0, zigzag_mask1, z0, 38 LD_SH2(round, 8, coeff0, coeff1); 39 VSHF_H2_SH(coeff0, coeff1, coeff0, coeff1, zigzag_mask0, zigzag_mask1, round0, 41 LD_SH2(quant, 8, coeff0, coeff1); 42 VSHF_H2_SH(coeff0, coeff1, coeff0, coeff1, zigzag_mask0, zigzag_mask1, quant0 93 v8i16 coeff0, coeff1, z0, z1; local [all...] |
denoising_msa.c | 33 v16u8 coeff0, coeff1; local 77 ILVRL_B2_UB(mc_running_avg_y0, sig0, coeff0, coeff1); 78 HSUB_UB2_SH(coeff0, coeff1, diff0, diff1); 125 ILVRL_B2_UB(mc_running_avg_y1, sig1, coeff0, coeff1); 126 HSUB_UB2_SH(coeff0, coeff1, diff0, diff1); 204 ILVRL_B2_UB(mc_running_avg_y0, sig0, coeff0, coeff1); 205 HSUB_UB2_SH(coeff0, coeff1, diff0, diff1); 240 ILVRL_B2_UB(mc_running_avg_y1, sig1, coeff0, coeff1); 241 HSUB_UB2_SH(coeff0, coeff1, diff0, diff1);
|
/external/libvpx/libvpx/vp9/encoder/x86/ |
vp9_quantize_sse2.c | 42 __m128i coeff0, coeff1; local 57 coeff1 = load_tran_low(coeff_ptr + n_coeffs + 8); 61 coeff1_sign = _mm_srai_epi16(coeff1, 15); 63 qcoeff1 = _mm_xor_si128(coeff1, coeff1_sign); 85 coeff1 = _mm_mullo_epi16(qcoeff1, dequant); 88 store_tran_low(coeff1, dqcoeff_ptr + n_coeffs + 8); 98 zero_coeff1 = _mm_cmpeq_epi16(coeff1, zero); 117 __m128i coeff0, coeff1; local 124 coeff1 = load_tran_low(coeff_ptr + n_coeffs + 8); 128 coeff1_sign = _mm_srai_epi16(coeff1, 15) [all...] |
vp9_dct_ssse3.c | 294 __m128i coeff0, coeff1; local 309 coeff1 = *in[1]; 313 coeff1_sign = _mm_srai_epi16(coeff1, 15); 315 qcoeff1 = _mm_xor_si128(coeff1, coeff1_sign); 337 coeff1 = _mm_mullo_epi16(qcoeff1, dequant); 340 store_tran_low(coeff1, dqcoeff_ptr + n_coeffs + 8); 350 zero_coeff1 = _mm_cmpeq_epi16(coeff1, zero); 369 __m128i coeff0, coeff1; local 377 coeff1 = *in[index + 1]; 381 coeff1_sign = _mm_srai_epi16(coeff1, 15) [all...] |
vp9_dct_intrin_sse2.c | 461 __m128i coeff0, coeff1; local 476 coeff1 = *in[1]; 480 coeff1_sign = _mm_srai_epi16(coeff1, 15); 482 qcoeff1 = _mm_xor_si128(coeff1, coeff1_sign); 504 coeff1 = _mm_mullo_epi16(qcoeff1, dequant); 507 _mm_store_si128((__m128i *)(dqcoeff_ptr + n_coeffs) + 1, coeff1); 517 zero_coeff1 = _mm_cmpeq_epi16(coeff1, zero); 535 __m128i coeff0, coeff1; local 543 coeff1 = *in[index + 1]; 547 coeff1_sign = _mm_srai_epi16(coeff1, 15) [all...] |
/external/libvpx/libvpx/vpx_dsp/x86/ |
quantize_sse2.c | 39 __m128i coeff0, coeff1; local 60 coeff1 = load_tran_low(coeff_ptr + n_coeffs + 8); 64 coeff1_sign = _mm_srai_epi16(coeff1, 15); 66 qcoeff1 = _mm_xor_si128(coeff1, coeff1_sign); 100 coeff1 = _mm_mullo_epi16(qcoeff1, dequant); 103 store_tran_low(coeff1, dqcoeff_ptr + n_coeffs + 8); 113 zero_coeff1 = _mm_cmpeq_epi16(coeff1, zero); 130 __m128i coeff0, coeff1; local 138 coeff1 = load_tran_low(coeff_ptr + n_coeffs + 8); 142 coeff1_sign = _mm_srai_epi16(coeff1, 15) [all...] |
avg_intrin_sse2.c | 260 __m128i coeff1 = load_tran_low(coeff + 64); local 264 __m128i b0 = _mm_add_epi16(coeff0, coeff1); 265 __m128i b1 = _mm_sub_epi16(coeff0, coeff1); 275 coeff1 = _mm_add_epi16(b1, b3); 277 store_tran_low(coeff1, coeff + 64);
|
/frameworks/base/libs/hwui/utils/ |
Blur.cpp | 77 float coeff1 = 1.0f / (sqrt(2.0f * pi) * sigma); local 83 weights[r + intRadius] = coeff1 * pow(e, floatR * floatR * coeff2);
|
/frameworks/rs/cpu_ref/ |
rsCpuIntrinsicBlur.cpp | 79 float coeff1 = 1.0f / (sqrtf(2.0f * pi) * sigma); local 88 mFp[r + mIradius] = coeff1 * powf(e, floatR * floatR * coeff2);
|
/external/libvpx/libvpx/vpx_dsp/mips/ |
fwd_txfm_msa.c | 23 v8i16 coeff1 = { cospi_2_64, cospi_30_64, cospi_14_64, cospi_18_64, local 73 SPLATI_H2_SH(coeff1, 0, 1, cnst0, cnst1); 85 SPLATI_H2_SH(coeff1, 2, 3, cnst0, cnst1); 109 SPLATI_H2_SH(coeff1, 4, 5, cnst0, cnst1); 121 SPLATI_H2_SH(coeff1, 6, 7, cnst0, cnst1);
|
/external/apache-commons-math/src/main/java/org/apache/commons/math/estimation/ |
LevenbergMarquardtEstimator.java | 394 double coeff1 = 0; local 396 coeff1 += work1[j] * work1[j]; 399 coeff1 = coeff1 / pc2; 401 double preRed = coeff1 + 2 * coeff2; 402 double dirDer = -(coeff1 + coeff2);
|
/external/apache-commons-math/src/main/java/org/apache/commons/math/optimization/general/ |
LevenbergMarquardtOptimizer.java | 395 double coeff1 = 0; local 397 coeff1 += work1[j] * work1[j]; 400 coeff1 = coeff1 / pc2; 402 double preRed = coeff1 + 2 * coeff2; 403 double dirDer = -(coeff1 + coeff2); [all...] |
/external/webp/src/dsp/ |
enc_msa.c | 234 v8i16 coeff0, coeff1; local 237 LD_SH2(&out[0], 8, coeff0, coeff1); 239 coeff1 = __msa_add_a_h(coeff1, zero); 240 SRAI_H2_SH(coeff0, coeff1, 3); 242 coeff1 = __msa_min_s_h(coeff1, max_coeff_thr); 243 ST_SH2(coeff0, coeff1, &out[0], 8);
|