/external/speex/libspeex/ |
lpc_bfin.h | 56 ac0 = ADD32(ac0,SHR32(MULT16_16(x[j],x[j]),8)); 115 d = ADD32(d,SHR32(MULT16_16(x[j],x[j]), shift)); 124 d = ADD32(d,SHR32(MULT16_16(x[j],x[j-i]), shift)); 128 ac[i] = SHR32(ac32[i], ac_shift);
|
fixed_generic.h | 47 #define SHR32(a,shift) ((a) >> (shift)) 50 #define PSHR32(a,shift) (SHR32((a)+((EXTEND32(1)<<((shift))>>1)),shift)) 51 #define VSHR32(a, shift) (((shift)>0) ? SHR32(a, shift) : SHL32(a, -(shift)))
|
lpc.c | 141 ac0 = ADD32(ac0,SHR32(MULT16_16(x[j],x[j]),8)); 162 d = ADD32(d,SHR32(MULT16_16(x[j],x[j-i]), shift)); 165 ac[i] = SHR32(d, ac_shift);
|
preprocess.c | 98 if (SHR32(a,7) >= b) 104 a = SHR32(a,8); 105 b = SHR32(b,8); 109 a = SHR32(a,4); 110 b = SHR32(b,4); 114 a = SHR32(a,4); 115 b = SHR32(b,4); 124 if (SHR32(a,15) >= b) 130 a = SHR32(a,8); 131 b = SHR32(b,8) [all...] |
stereo.c | 164 e_left += SHR32(MULT16_16(data[2*i],data[2*i]),8); 165 e_right += SHR32(MULT16_16(data[2*i+1],data[2*i+1]),8); 172 e_tot += SHR32(MULT16_16(data[i],data[i]),8); 235 e_left = SHR32(MULT16_16(spx_sqrt(balance), e_right), 8); 262 e_left = SHR32(MULT16_16(spx_sqrt(balance), e_right), 8);
|
math_approx.h | 126 res = MULT16_16(EXTRACT16(SHR32(*seed,16)),std); 127 return EXTRACT16(PSHR32(SUB32(res, SHR32(res, 3)),14)); 299 x = DIV32_16(SHL32(EXTEND32(32767),29-e), EXTRACT16(SHR32(x, e-14)));
|
fixed_debug.h | 132 static inline int SHR32(long long a, int shift) 137 fprintf (stderr, "SHR32: inputs are not int: %d %d\n", (int)a, shift); 142 fprintf (stderr, "SHR32: output is not int: %d\n", (int)res); 164 #define PSHR32(a,shift) (SHR32(ADD32((a),((EXTEND32(1)<<((shift))>>1))),shift)) 165 #define VSHR32(a, shift) (((shift)>0) ? SHR32(a, shift) : SHL32(a, -(shift))) 269 #define MAC16_16_Q11(c,a,b) (EXTRACT16(ADD16((c),EXTRACT16(SHR32(MULT16_16((a),(b)),11))))) 270 #define MAC16_16_Q13(c,a,b) (EXTRACT16(ADD16((c),EXTRACT16(SHR32(MULT16_16((a),(b)),13))))) 271 #define MAC16_16_P13(c,a,b) (EXTRACT16(ADD32((c),SHR32(ADD32(4096,MULT16_16((a),(b))),13))))
|
ltp.c | 71 sum = ADD32(sum,SHR32(part,6)); 129 sum1 = ADD32(sum1,SHR32(part1,6)); 130 sum2 = ADD32(sum2,SHR32(part2,6)); 131 sum3 = ADD32(sum3,SHR32(part3,6)); 132 sum4 = ADD32(sum4,SHR32(part4,6)); 235 energy[i-start+1] = SUB32(ADD32(energy[i-start],SHR32(MULT16_16(sw[-i-1],sw[-i-1]),6)), SHR32(MULT16_16(sw[-i+len-1],sw[-i+len-1]),6)); 298 g = DIV32(SHL32(EXTEND32(corr16[i-start]),cshift), 10+SHR32(MULT16_16(spx_sqrt(e0),spx_sqrt(SHL32(EXTEND32(ener16[i-start]),eshift))),6));
|
vq.c | 86 dist=SUB32(SHR32(E[i],1),dist); 128 dist = ADD32(dist,SHR32(E[i],1));
|
filters.c | 118 y[i] = SHL32(MULT16_32_Q14(EXTRACT16(SHR32(x[i],7)),scale),7); 134 } else if (scale > SHR32(EXTEND32(SIG_SCALING), 2)) { 206 tmp = EXTRACT16(SHR32(x[i],sig_shift)); 208 tmp = EXTRACT16(SHR32(x[i+1],sig_shift)); 210 tmp = EXTRACT16(SHR32(x[i+2],sig_shift)); 212 tmp = EXTRACT16(SHR32(x[i+3],sig_shift)); 214 sum = ADD32(sum,SHR32(sum2,6)); 243 sum = ADD32(sum,SHR32(sum2,6)); 262 sum = ADD32(sum,SHR32(sum2,6)); 292 y[i] = EXTRACT16(SHR32(x[i], sig_shift)) [all...] |
kiss_fftr.c | 228 f2k.i = SHR32(ADD32(EXTEND32(st->tmpbuf[k].i), EXTEND32(st->tmpbuf[ncfft-k].i)),1); 237 f2k.r = SHR32(SUB32(EXTEND32(st->tmpbuf[k].r), EXTEND32(st->tmpbuf[ncfft-k].r)),1); 243 twr = SHR32(SUB32(MULT16_16(f2k.r,st->super_twiddles[k].r),MULT16_16(f2k.i,st->super_twiddles[k].i)), 1); 244 twi = SHR32(ADD32(MULT16_16(f2k.i,st->super_twiddles[k].r),MULT16_16(f2k.r,st->super_twiddles[k].i)), 1);
|
/external/chromium_org/third_party/opus/src/celt/ |
celt_lpc.c | 61 rr += SHR32(ac[i + 1],3); 64 lpc[i] = SHR32(r,3); 77 if (error<SHR32(ac[0],10)) 261 if (n&1) ac0 += SHR32(MULT16_16(xptr[0],xptr[0]),9); 264 ac0 += SHR32(MULT16_16(xptr[i],xptr[i]),9); 265 ac0 += SHR32(MULT16_16(xptr[i+1],xptr[i+1]),9); 302 ac[i] = SHR32(ac[i], shift2);
|
pitch.c | 69 Syy = ADD32(Syy, SHR32(MULT16_16(y[j],y[j]), yshift)); 100 Syy += SHR32(MULT16_16(y[i+len],y[i+len]),yshift) - SHR32(MULT16_16(y[i],y[i]),yshift); 173 x_lp[i] = SHR32(HALF32(HALF32(x[0][(2*i-1)]+x[0][(2*i+1)])+x[0][2*i]), shift); 174 x_lp[0] = SHR32(HALF32(HALF32(x[0][1])+x[0][0]), shift); 178 x_lp[i] += SHR32(HALF32(HALF32(x[1][(2*i-1)]+x[1][(2*i+1)])+x[1][2*i]), shift); 179 x_lp[0] += SHR32(HALF32(HALF32(x[1][1])+x[1][0]), shift); 187 ac[0] += SHR32(ac[0],13); 369 sum += SHR32(MULT16_16(x_lp[j],y[i+j]), shift); 513 pg = SHR32(frac_div32(best_xy,best_yy+1),16) [all...] |
fixed_debug.h | 45 #define MULT32_32_Q31(a,b) ADD32(ADD32(SHL32(MULT16_16(SHR32((a),16),SHR((b),16)),1), SHR32(MULT16_16SU(SHR32((a),16),((b)&0x0000ffff)),15)), SHR32(MULT16_16SU(SHR32((b),16),((a)&0x0000ffff)),15)) 48 #define MULT16_32_Q16(a,b) ADD32(MULT16_16((a),SHR32((b),16)), SHR32(MULT16_16SU((a),((b)&0x0000ffff)),16)) 59 #define SHR(a,b) SHR32(a,b) 182 static OPUS_INLINE int SHR32(opus_int64 a, int shift) 187 fprintf (stderr, "SHR32: inputs are not int: %d %d\n", (int)a, shift) [all...] |
fixed_generic.h | 72 #define SHR32(a,shift) ((a) >> (shift)) 77 #define PSHR32(a,shift) (SHR32((a)+((EXTEND32(1)<<((shift))>>1)),shift)) 79 #define VSHR32(a, shift) (((shift)>0) ? SHR32(a, shift) : SHL32(a, -(shift))) 93 #define HALF32(x) (SHR32(x,1))
|
vq.c | 51 Xptr[stride] = EXTRACT16(SHR32(MULT16_16(c,x2) + MULT16_16(s,x1), 15)); 52 *Xptr++ = EXTRACT16(SHR32(MULT16_16(c,x1) - MULT16_16(s,x2), 15)); 60 Xptr[stride] = EXTRACT16(SHR32(MULT16_16(c,x2) + MULT16_16(s,x1), 15)); 61 *Xptr-- = EXTRACT16(SHR32(MULT16_16(c,x1) - MULT16_16(s,x2), 15)); 272 Rxy = EXTRACT16(SHR32(ADD32(xy, EXTEND32(X[j])),rshift));
|
/external/libopus/celt/ |
celt_lpc.c | 61 rr += SHR32(ac[i + 1],3); 64 lpc[i] = SHR32(r,3); 77 if (error<SHR32(ac[0],10)) 261 if (n&1) ac0 += SHR32(MULT16_16(xptr[0],xptr[0]),9); 264 ac0 += SHR32(MULT16_16(xptr[i],xptr[i]),9); 265 ac0 += SHR32(MULT16_16(xptr[i+1],xptr[i+1]),9); 302 ac[i] = SHR32(ac[i], shift2);
|
pitch.c | 69 Syy = ADD32(Syy, SHR32(MULT16_16(y[j],y[j]), yshift)); 100 Syy += SHR32(MULT16_16(y[i+len],y[i+len]),yshift) - SHR32(MULT16_16(y[i],y[i]),yshift); 173 x_lp[i] = SHR32(HALF32(HALF32(x[0][(2*i-1)]+x[0][(2*i+1)])+x[0][2*i]), shift); 174 x_lp[0] = SHR32(HALF32(HALF32(x[0][1])+x[0][0]), shift); 178 x_lp[i] += SHR32(HALF32(HALF32(x[1][(2*i-1)]+x[1][(2*i+1)])+x[1][2*i]), shift); 179 x_lp[0] += SHR32(HALF32(HALF32(x[1][1])+x[1][0]), shift); 187 ac[0] += SHR32(ac[0],13); 369 sum += SHR32(MULT16_16(x_lp[j],y[i+j]), shift); 513 pg = SHR32(frac_div32(best_xy,best_yy+1),16) [all...] |
fixed_debug.h | 45 #define MULT32_32_Q31(a,b) ADD32(ADD32(SHL32(MULT16_16(SHR32((a),16),SHR((b),16)),1), SHR32(MULT16_16SU(SHR32((a),16),((b)&0x0000ffff)),15)), SHR32(MULT16_16SU(SHR32((b),16),((a)&0x0000ffff)),15)) 48 #define MULT16_32_Q16(a,b) ADD32(MULT16_16((a),SHR32((b),16)), SHR32(MULT16_16SU((a),((b)&0x0000ffff)),16)) 59 #define SHR(a,b) SHR32(a,b) 182 static OPUS_INLINE int SHR32(opus_int64 a, int shift) 187 fprintf (stderr, "SHR32: inputs are not int: %d %d\n", (int)a, shift) [all...] |
fixed_generic.h | 72 #define SHR32(a,shift) ((a) >> (shift)) 77 #define PSHR32(a,shift) (SHR32((a)+((EXTEND32(1)<<((shift))>>1)),shift)) 79 #define VSHR32(a, shift) (((shift)>0) ? SHR32(a, shift) : SHL32(a, -(shift))) 93 #define HALF32(x) (SHR32(x,1))
|
vq.c | 51 Xptr[stride] = EXTRACT16(SHR32(MULT16_16(c,x2) + MULT16_16(s,x1), 15)); 52 *Xptr++ = EXTRACT16(SHR32(MULT16_16(c,x1) - MULT16_16(s,x2), 15)); 60 Xptr[stride] = EXTRACT16(SHR32(MULT16_16(c,x2) + MULT16_16(s,x1), 15)); 61 *Xptr-- = EXTRACT16(SHR32(MULT16_16(c,x1) - MULT16_16(s,x2), 15)); 272 Rxy = EXTRACT16(SHR32(ADD32(xy, EXTEND32(X[j])),rshift));
|
/external/chromium_org/third_party/opus/src/celt/arm/ |
kiss_fft_armv5e.h | 86 (m).r = SHR32(SUB32(mr1__, mr2__), 1); \ 87 (m).i = SHR32(mi__, 1); \
|
/external/libopus/celt/arm/ |
kiss_fft_armv5e.h | 86 (m).r = SHR32(SUB32(mr1__, mr2__), 1); \ 87 (m).i = SHR32(mi__, 1); \
|
/external/chromium_org/third_party/opus/src/src/ |
mlp.c | 53 xx = EXTRACT16(SHR32(_x, 8)); 55 i = SHR32(ADD32(1024,MULT16_16(25, xx)),11); 57 xx -= EXTRACT16(SHR32(MULT16_16(20972,i),8));
|
/external/libopus/src/ |
mlp.c | 53 xx = EXTRACT16(SHR32(_x, 8)); 55 i = SHR32(ADD32(1024,MULT16_16(25, xx)),11); 57 xx -= EXTRACT16(SHR32(MULT16_16(20972,i),8));
|