/frameworks/av/media/libstagefright/codecs/amrwbenc/src/ |
autocorr.c | 42 Word32 L_sum, L_sum1, L_tmp, F_LEN; 64 L_tmp = vo_L_mult(y[i], y[i]); 65 L_tmp = (L_tmp >> 8); 66 L_sum += L_tmp;
|
cor_h_x.c | 42 Word32 L_tmp, y32[L_SUBFR], L_tot; 54 L_tmp = 1; /* 1 -> to avoid null dn[] */ 58 L_tmp = L_add(L_tmp, vo_L_mult(*p1++, *p2++)); 60 y32[i] = L_tmp; 61 L_tmp = (L_tmp > 0)? L_tmp: (L_tmp == INT_MIN ? INT_MAX : -L_tmp); [all...] |
gpclip.c | 95 Word32 L_tmp; 96 L_tmp = (29491 * mem[1])<<1; 97 L_tmp += (3277 * gain_pit)<<1; 99 gain = extract_h(L_tmp);
|
hp400.c | 65 Word32 L_tmp; 81 L_tmp = 8192L; /* rounding to maximise precision */ 82 L_tmp += y1_lo * a[1]; 83 L_tmp += y2_lo * a[2]; 84 L_tmp = L_tmp >> 14; 85 L_tmp += (y1_hi * a[1] + y2_hi * a[2] + (x0 + x2)* b[0] + x1 * b[1]) << 1; 86 L_tmp <<= 1; /* coeff Q12 --> Q13 */ 89 y1_hi = (Word16)(L_tmp>>16); 90 y1_lo = (Word16)((L_tmp & 0xffff)>>1) [all...] |
hp50.c | 66 Word32 L_tmp; 83 L_tmp = 8192 ; /* rounding to maximise precision */ 84 L_tmp += y1_lo * a[1]; 85 L_tmp += y2_lo * a[2]; 86 L_tmp = L_tmp >> 14; 87 L_tmp += (y1_hi * a[1] + y2_hi * a[2] + (x0 + x2) * b[0] + x1 * b[1]) << 1; 88 L_tmp <<= 2; /* coeff Q12 --> Q13 */ 91 y1_hi = (Word16)(L_tmp>>16); 92 y1_lo = (Word16)((L_tmp & 0xffff)>>1) [all...] |
syn_filt.c | 42 Word32 L_tmp; 56 L_tmp = vo_mult32(a0, x[i]); 57 L_tmp -= vo_mult32((*p1++), (*p2--)); 58 L_tmp -= vo_mult32((*p1++), (*p2--)); 59 L_tmp -= vo_mult32((*p1++), (*p2--)); 60 L_tmp -= vo_mult32((*p1++), (*p2--)); 61 L_tmp -= vo_mult32((*p1++), (*p2--)); 62 L_tmp -= vo_mult32((*p1++), (*p2--)); 63 L_tmp -= vo_mult32((*p1++), (*p2--)); 64 L_tmp -= vo_mult32((*p1++), (*p2--)) [all...] |
hp_wsp.c | 59 Word32 L_tmp; 63 L_tmp = ((mem[i] << 16) + (mem[i + 1]<<1)); 64 L_tmp = L_shl(L_tmp, exp); 65 mem[i] = L_tmp >> 16; 66 mem[i + 1] = (L_tmp & 0xffff)>>1; 71 L_tmp = L_deposit_h(mem[i]); /* x[i] */ 72 L_tmp = L_shl(L_tmp, exp); 73 mem[i] = vo_round(L_tmp); [all...] |
dtx.c | 387 Word32 L_tmp; 406 L_tmp = 0; 410 L_tmp = L_add(L_tmp, L_deposit_l(isf_old[i * M + j])); 412 isf_aver[j] = L_tmp; 436 Word32 L_tmp, summin, summax, summax2nd; 491 L_tmp = 0; 495 L_tmp = L_mac(L_tmp, tmp, tmp); 497 st->D[i - 1] = L_tmp; [all...] |
pitch_f4.c | 175 Word32 L_tmp, L_tmp1, L_tmp2; 188 L_tmp = 0; 191 L_tmp = L_add(L_tmp, (xn[i] * xn[i])); 192 L_tmp = L_add(L_tmp, (xn[i+1] * xn[i+1])); 193 L_tmp = L_add(L_tmp, (xn[i+2] * xn[i+2])); 194 L_tmp = L_add(L_tmp, (xn[i+3] * xn[i+3])) [all...] |
q_gain2.c | 83 Word32 i, j, L_tmp, dist_min; 173 * L_tmp = MEAN_ENER - 10log10(energy of code/ L_subfr) * 177 L_tmp = Dot_product12_asm(code, code, L_subfr, &exp_code); 179 L_tmp = Dot_product12(code, code, L_subfr, &exp_code); 181 /* exp_code: -18 (code in Q9), -6 (/L_subfr), -31 (L_tmp Q31->Q0) */ 184 Log2(L_tmp, &exp, &frac); 186 L_tmp = Mpy_32_16(exp, frac, -24660); /* x -3.0103(Q13) -> Q14 */ 188 L_tmp += (MEAN_ENER * 8192)<<1; /* + MEAN_ENER in Q14 */ 194 L_tmp = (L_tmp << 10); /* From Q14 to Q24 * [all...] |
qpisf_2s.c | 236 Word32 i, j, L_tmp; 285 L_tmp = mean_isf[i] << 14; 288 L_tmp += (isf_buf[j * M + i] << 14); 290 ref_isf[i] = vo_round(L_tmp); 329 Word32 i, j, L_tmp; 379 L_tmp = (mean_isf[i] << 14); 382 L_tmp += (isf_buf[j * M + i] << 14); 384 ref_isf[i] = vo_round(L_tmp);
|
c4t64fx.c | 140 Word32 s, cor, L_tmp, L_index; 340 L_tmp = 0; 345 L_tmp = L_add(L_tmp, (H[i] * H[i]) << 1); 348 val = extract_h(L_tmp); 628 L_tmp = 0L; 634 if (vecSq2 > 0 && L_tmp > INT_MAX - vecSq2) { 635 L_tmp = INT_MAX; 636 } else if (vecSq2 < 0 && L_tmp < INT_MIN - vecSq2) { 637 L_tmp = INT_MIN [all...] |
voAMRWBEnc.c | 208 Word32 L_tmp, L_gain_code, L_max, L_tmp1; 281 L_tmp = new_speech[0] << 15; 282 L_tmp -= (st->mem_preemph * mu)<<1; 283 L_max = L_abs(L_tmp); 287 L_tmp = new_speech[i] << 15; 288 L_tmp -= (new_speech[i - 1] * mu)<<1; 289 L_tmp = L_abs(L_tmp); 290 if(L_tmp > L_max) 292 L_max = L_tmp; [all...] |
/frameworks/av/media/libstagefright/codecs/amrnb/dec/src/ |
pstfilt.cpp | 330 Word32 L_tmp; 371 L_tmp = L_mult (h[0], h[0]); 374 L_tmp = L_mac (L_tmp, h[i], h[i]); 376 temp1 = extract_h (L_tmp); 378 L_tmp = L_mult (h[0], h[1]); 381 L_tmp = L_mac (L_tmp, h[i], h[i + 1]); 383 temp2 = extract_h (L_tmp); 457 Word32 L_tmp; [all...] |
/frameworks/av/media/libstagefright/codecs/amrnb/common/src/ |
gc_pred.cpp | 338 Word32 L_tmp; 352 L_tmp = Mpy_32_16(exp, frac, -24660); // Q0.Q15 * Q13 -> Q14 354 * L_tmp = means_ener - 10log10(ener_code/L_SUBFR) 379 L_tmp = L_mac(L_tmp, 16678, 64); // Q14 395 L_tmp = L_mac(L_tmp, 17062, 64); // Q14 400 L_tmp = L_mac(L_tmp, 32588, 32); // Q14 405 L_tmp = L_mac(L_tmp, 32268, 32); // Q1 [all...] |
vad1.cpp | [all...] |
/frameworks/av/media/libstagefright/codecs/amrnb/enc/src/ |
qgain475.cpp | 411 Word32 L_tmp; 427 L_tmp = L_mult(g_code, gcode0); 428 L_tmp = L_shr(L_tmp, sub(10, exp_gcode0)); 429 *gain_cod = extract_h(L_tmp); 445 L_tmp = Mpy_32_16(exp, frac, 24660); // 24660 Q12 ~= 6.0206 = 20*log10(2) 446 qua_ener = pv_round (L_shl (L_tmp, 13)); // Q12 * Q0 = Q13 -> Q10 488 Word32 L_tmp; 505 L_tmp = ((Word32) g_code * gcode0) << 1; 507 L_tmp = L_shr(L_tmp, tmp, pOverflow) [all...] |
qua_gain.cpp | 227 Word32 L_tmp; 315 L_tmp = L_deposit_h(frac_coeff[i]); 316 L_tmp = L_shr(L_tmp, j, pOverflow); 317 L_Extract(L_tmp, &coeff[i], &coeff_lo[i], pOverflow); 351 L_tmp = Mpy_32_16(coeff[0], coeff_lo[0], g2_pitch, pOverflow); 353 L_tmp = L_add(L_tmp, L_tmp2, pOverflow); 356 L_tmp = L_add(L_tmp, L_tmp2, pOverflow) [all...] |
qgain795.cpp | 242 Word32 L_tmp; 296 L_tmp = L_deposit_h(frac_coeff[i]); 297 L_tmp = L_shr(L_tmp, j, pOverflow); 298 L_Extract(L_tmp, &coeff[i], &coeff_lo[i], pOverflow); 337 L_tmp = L_mult(g_code, g_code, pOverflow); 338 L_Extract(L_tmp, &g2_code_h, &g2_code_l, pOverflow); 340 L_tmp = L_mult(g_code, g_pitch, pOverflow); 341 L_Extract(L_tmp, &g_pit_cod_h, &g_pit_cod_l, pOverflow); 343 L_tmp = Mac_32(L_tmp0, coeff[2], coeff_lo[2] [all...] |
/frameworks/av/media/libstagefright/codecs/amrwb/src/ |
synthesis_amr_wb.cpp | 153 int32 L_tmp; 262 L_tmp = L_deposit_h(div_16by16(tmp, ener)); /* result is normalized */ 264 one_ov_sqrt_norm(&L_tmp, &exp); 265 L_tmp = shl_int32(L_tmp, exp + 1); /* L_tmp x 2, L_tmp in Q31 */ 267 tmp = (int16)(L_tmp >> 16); /* tmp = 2 x sqrt(ener_exc/ener_hf) */ 288 L_tmp = 1L; 292 L_tmp = mac_16by16_to_int32(L_tmp, synth[0], synth[0]) [all...] |
dec_gain2_amr_wb.cpp | 190 int32 L_tmp; 203 * L_tmp = 1.0 / sqrt(energy of code/ L_subfr) 206 L_tmp = Dot_product12(code, code, L_subfr, &exp); 209 one_ov_sqrt_norm(&L_tmp, &exp); 211 gcode_inov = extract_h(shl_int32(L_tmp, exp - 3)); /* g_code_inov in Q12 */ 259 L_tmp = tmp; 260 L_tmp += tmp1; 264 L_tmp += tmp; 265 L_tmp += tmp1; 267 qua_ener = (int16)(L_tmp >> 3) [all...] |
dtx_decoder_amr_wb.cpp | 675 int32 L_tmp; 695 L_tmp = 0; 699 L_tmp = add_int32(L_tmp, (int32)(isf_old[i * M + j])); 701 isf_aver[j] = L_tmp; 731 int32 L_tmp, summin, summax, summax2nd; 787 L_tmp = 0; 791 L_tmp = mac_16by16_to_int32(L_tmp, tmp, tmp); 793 st->D[i - 1] = L_tmp; [all...] |
pvamrwbdecoder.cpp | 323 int32 L_tmp, L_gain_code; 449 L_tmp = mul_16by16_to_int32(isf_tmp[i], sub_int16(32767, interpol_frac[j])); 450 L_tmp = mac_16by16_to_int32(L_tmp, isf[i], interpol_frac[j]); 451 HfIsf[i] = amr_wb_round(L_tmp); 531 L_tmp = 0; 535 L_tmp = mac_16by16_to_int32(L_tmp, tmp, tmp); 537 tmp = extract_h(shl_int32(L_tmp, 8)); 538 tmp = mult_int16(tmp, 26214); /* tmp = L_tmp*0.8/256 * [all...] |
/external/aac/libAACdec/src/ |
usacdec_fac.cpp | 188 FIXP_DBL L_tmp; 191 L_tmp = (FIXP_DBL)0; 194 L_tmp -= fMultDiv2(a[j], x[i - (j + 1)]); 197 L_tmp = scaleValue(L_tmp, a_exp + 1); 199 x[i] = scaleValueSaturate((x[i] >> 1) + (L_tmp >> 1),
|
usacdec_acelp.cpp | 149 FIXP_DBL L_tmp; 153 L_tmp = FX_COD2FX_DBL(x[i]); 154 L_tmp -= fMultDiv2(x[i - 1], TILT_CODE2); 155 x[i] = FX_DBL2FX_COD(L_tmp); 169 FIXP_DBL L_tmp; 172 L_tmp = FX_COD2FX_DBL(x[i]); 173 L_tmp += fMult(x[i - pit_lag], PIT_SHARP); 174 x[i] = FX_DBL2FX_COD(L_tmp); 305 FIXP_DBL ener_exc, L_tmp; 327 /* L_tmp = ener_code * gain_code*gain_code; * [all...] |