/external/chromium_org/third_party/opus/src/silk/ |
resampler_private_up2_HQ.c | 63 out32_1 = silk_ADD32( S[ 0 ], X ); 64 S[ 0 ] = silk_ADD32( in32, X ); 69 out32_2 = silk_ADD32( S[ 1 ], X ); 70 S[ 1 ] = silk_ADD32( out32_1, X ); 75 out32_1 = silk_ADD32( S[ 2 ], X ); 76 S[ 2 ] = silk_ADD32( out32_2, X ); 84 out32_1 = silk_ADD32( S[ 3 ], X ); 85 S[ 3 ] = silk_ADD32( in32, X ); 90 out32_2 = silk_ADD32( S[ 4 ], X ); 91 S[ 4 ] = silk_ADD32( out32_1, X ) [all...] |
resampler_down2.c | 57 out32 = silk_ADD32( S[ 0 ], X ); 58 S[ 0 ] = silk_ADD32( in32, X ); 66 out32 = silk_ADD32( out32, S[ 1 ] ); 67 out32 = silk_ADD32( out32, X ); 68 S[ 1 ] = silk_ADD32( in32, X );
|
resampler_private_down_FIR.c | 92 res_Q6 = silk_SMULWB( silk_ADD32( buf_ptr[ 0 ], buf_ptr[ 23 ] ), FIR_Coefs[ 0 ] ); 93 res_Q6 = silk_SMLAWB( res_Q6, silk_ADD32( buf_ptr[ 1 ], buf_ptr[ 22 ] ), FIR_Coefs[ 1 ] ); 94 res_Q6 = silk_SMLAWB( res_Q6, silk_ADD32( buf_ptr[ 2 ], buf_ptr[ 21 ] ), FIR_Coefs[ 2 ] ); 95 res_Q6 = silk_SMLAWB( res_Q6, silk_ADD32( buf_ptr[ 3 ], buf_ptr[ 20 ] ), FIR_Coefs[ 3 ] ); 96 res_Q6 = silk_SMLAWB( res_Q6, silk_ADD32( buf_ptr[ 4 ], buf_ptr[ 19 ] ), FIR_Coefs[ 4 ] ); 97 res_Q6 = silk_SMLAWB( res_Q6, silk_ADD32( buf_ptr[ 5 ], buf_ptr[ 18 ] ), FIR_Coefs[ 5 ] ); 98 res_Q6 = silk_SMLAWB( res_Q6, silk_ADD32( buf_ptr[ 6 ], buf_ptr[ 17 ] ), FIR_Coefs[ 6 ] ); 99 res_Q6 = silk_SMLAWB( res_Q6, silk_ADD32( buf_ptr[ 7 ], buf_ptr[ 16 ] ), FIR_Coefs[ 7 ] ); 100 res_Q6 = silk_SMLAWB( res_Q6, silk_ADD32( buf_ptr[ 8 ], buf_ptr[ 15 ] ), FIR_Coefs[ 8 ] ); 101 res_Q6 = silk_SMLAWB( res_Q6, silk_ADD32( buf_ptr[ 9 ], buf_ptr[ 14 ] ), FIR_Coefs[ 9 ] ) [all...] |
ana_filt_bank_1.c | 58 out_1 = silk_ADD32( S[ 0 ], X ); 59 S[ 0 ] = silk_ADD32( in32, X ); 67 out_2 = silk_ADD32( S[ 1 ], X ); 68 S[ 1 ] = silk_ADD32( in32, X ); 71 outL[ k ] = (opus_int16)silk_SAT16( silk_RSHIFT_ROUND( silk_ADD32( out_2, out_1 ), 11 ) );
|
NSQ_del_dec.c | 376 n_LTP_Q14 = silk_SMULWB( silk_ADD32( shp_lag_ptr[ 0 ], shp_lag_ptr[ -2 ] ), HarmShapeFIRPacked_Q14 ); 453 tmp1 = silk_ADD32( n_AR_Q14, n_LF_Q14 ); /* Q14 */ 454 tmp2 = silk_ADD32( n_LTP_Q14, LPC_pred_Q14 ); /* Q13 */ 471 q1_Q10 = silk_ADD32( q1_Q10, offset_Q10 ); 472 q2_Q10 = silk_ADD32( q1_Q10, 1024 ); 477 q2_Q10 = silk_ADD32( q1_Q10, 1024 - QUANT_LEVEL_ADJUST_Q10 ); 486 q1_Q10 = silk_ADD32( silk_LSHIFT( q1_Q0, 10 ), QUANT_LEVEL_ADJUST_Q10 ); 487 q1_Q10 = silk_ADD32( q1_Q10, offset_Q10 ); 488 q2_Q10 = silk_ADD32( q1_Q10, 1024 ); 498 psSS[ 0 ].RD_Q10 = silk_ADD32( psDD->RD_Q10, rd1_Q10 ) [all...] |
NLSF_decode.c | 95 NLSF_Q15_tmp = silk_ADD32( pNLSF_Q15[ i ], silk_DIV32_16( silk_LSHIFT( (opus_int32)res_Q10[ i ], 14 ), W_tmp_Q9 ) );
|
NSQ.c | 284 n_LTP_Q13 = silk_SMULWB( silk_ADD32( shp_lag_ptr[ 0 ], shp_lag_ptr[ -2 ] ), HarmShapeFIRPacked_Q14 ); 309 q1_Q10 = silk_ADD32( q1_Q10, offset_Q10 ); 310 q2_Q10 = silk_ADD32( q1_Q10, 1024 ); 315 q2_Q10 = silk_ADD32( q1_Q10, 1024 - QUANT_LEVEL_ADJUST_Q10 ); 324 q1_Q10 = silk_ADD32( silk_LSHIFT( q1_Q0, 10 ), QUANT_LEVEL_ADJUST_Q10 ); 325 q1_Q10 = silk_ADD32( q1_Q10, offset_Q10 ); 326 q2_Q10 = silk_ADD32( q1_Q10, 1024 );
|
SigProc_FIX.h | 397 #define silk_MLA(a32, b32, c32) silk_ADD32((a32),((b32) * (c32))) 406 #define silk_SMLATT(a32, b32, c32) silk_ADD32((a32),((b32) >> 16) * ((c32) >> 16)) 429 #define silk_ADD32(a, b) ((a) + (b)) 447 #define silk_ADD_SAT16(a, b) (opus_int16)silk_SAT16( silk_ADD32( (opus_int32)(a), (b) ) ) 487 #define silk_ADD_LSHIFT32(a, b, shift) silk_ADD32((a), silk_LSHIFT32((b), (shift))) /* shift >= 0 */ 490 #define silk_ADD_RSHIFT32(a, b, shift) silk_ADD32((a), silk_RSHIFT32((b), (shift))) /* shift >= 0 */ [all...] |
MacroDebug.h | 53 #undef silk_ADD32 54 #define silk_ADD32(a,b) silk_ADD32_((a), (b), __FILE__, __LINE__) 61 fprintf (stderr, "silk_ADD32(%d, %d) in %s: line %d\n", a, b, file, line); 137 res = (opus_int16)silk_SAT16( silk_ADD32( (opus_int32)(a16), (b16) ) ); 336 ret = silk_ADD32( a32, silk_SMULWB( b32, c32 ) ); 337 if ( silk_ADD32( a32, silk_SMULWB( b32, c32 ) ) != silk_ADD_SAT32( a32, silk_SMULWB( b32, c32 ) ) ) 460 ret = silk_ADD32( tmp1, tmp2 ); 461 fail |= silk_ADD32( tmp1, tmp2 ) != silk_ADD_SAT32( tmp1, tmp2 ); 483 ret = silk_ADD32( a32, tmp ); [all...] |
MacroCount.h | 508 #undef silk_ADD32 509 static inline opus_int32 silk_ADD32(opus_int32 a, opus_int32 b){ 552 res = (opus_int16)silk_SAT16( silk_ADD32( (opus_int32)(a16), (b16) ) );
|
/external/chromium_org/third_party/opus/src/silk/fixed/ |
regularize_correlations_FIX.c | 44 matrix_ptr( &XX[ 0 ], i, i, D ) = silk_ADD32( matrix_ptr( &XX[ 0 ], i, i, D ), noise );
|
find_LTP_FIX.c | 151 LPC_res_nrg = silk_ADD32( LPC_res_nrg, silk_RSHIFT( silk_ADD32( silk_SMULWB( rr[ k ], Wght_Q15[ k ] ), 1 ), 1 + ( maxRshifts - corr_rshifts[ k ] ) ) ); /* Q( -maxRshifts ) */ 152 LPC_LTP_res_nrg = silk_ADD32( LPC_LTP_res_nrg, silk_RSHIFT( silk_ADD32( silk_SMULWB( nrg[ k ], Wght_Q15[ k ] ), 1 ), 1 + ( maxRshifts - corr_rshifts[ k ] ) ) ); /* Q( -maxRshifts ) */ 201 temp32 = silk_ADD32( temp32, silk_RSHIFT( w[ k ], maxRshifts_wxtra - corr_rshifts[ k ] ) ); /* Q( 18 - maxRshifts_wxtra ) */ 202 wd = silk_ADD32( wd, silk_LSHIFT( silk_SMULWW( silk_RSHIFT( w[ k ], maxRshifts_wxtra - corr_rshifts[ k ] ), d_Q14[ k ] ), 2 ) ); /* Q( 18 - maxRshifts_wxtra ) */
|
corrMatrix_FIX.c | 113 energy = silk_ADD32( energy, silk_RSHIFT32( silk_SMULBB( ptr1[ -j ], ptr1[ -j ] ), rshifts_local ) ); 132 energy = silk_ADD32( energy, silk_RSHIFT32( silk_SMULBB( ptr1[ -j ], ptr2[ -j ] ), rshifts_local ) );
|
solve_LS_FIX.c | 148 matrix_ptr( A, i, i, M ) = silk_ADD32( matrix_ptr( A, i, i, M ), tmp_32 ); 176 matrix_ptr( L_Q16, i, j, M ) = silk_ADD32( silk_SMMUL( tmp_32, one_div_diag_Q48 ), 203 T[ i ] = silk_ADD32( silk_SMMUL( tmp_32, one_div_diag_Q48 ), silk_RSHIFT( silk_SMULWW( tmp_32, one_div_diag_Q36 ), 4 ) );
|
find_LPC_FIX.c | 112 res_nrg_interp = silk_ADD32( res_nrg0, res_nrg1 );
|
pitch_analysis_core_FIX.c | 197 normalizer = silk_ADD32( normalizer, silk_inner_prod_aligned( basis_ptr, basis_ptr, SF_LENGTH_8KHZ ) ); 198 normalizer = silk_ADD32( normalizer, silk_SMULBB( SF_LENGTH_8KHZ, 4000 ) ); 214 normalizer = silk_ADD32( normalizer, 335 energy_target = silk_ADD32( silk_inner_prod_aligned( target_ptr, target_ptr, SF_LENGTH_8KHZ ), 1 ); 349 silk_ADD32( energy_target, 527 energy_target = silk_ADD32( silk_inner_prod_aligned( target_ptr, target_ptr, nb_subfr * sf_length ), 1 ); 533 cross_corr = silk_ADD32( cross_corr, 536 energy = silk_ADD32( energy, [all...] |
burg_modified_FIX.c | 174 nrg = silk_ADD32( CAb[ 0 ], CAf[ 0 ] ); /* Q( 1-rshifts ) */ 184 nrg = silk_ADD_LSHIFT32( nrg, silk_SMMUL( silk_ADD32( CAb[ k + 1 ], CAf[ k + 1 ] ), 189 num = silk_ADD32( num, tmp2 ); /* Q( -rshifts ) */
|
noise_shape_analysis_FIX.c | 249 BWExp2_Q16 = silk_ADD32( BWExp2_Q16, delta_Q16 ); 288 auto_corr[0] = silk_ADD32( auto_corr[0], silk_max_32( silk_SMULWB( silk_RSHIFT( auto_corr[ 0 ], 4 ),
|