/external/webrtc/webrtc/modules/audio_processing/vad/ |
pitch_internal.h | 17 double* gains,
|
pitch_internal.cc | 29 double* gains, 39 gains[n] = log(gains[n] + 1e-12); 41 // Interpolate lags and gains. 42 PitchInterpolation(*log_old_gain, gains, log_pitch_gain); 43 *log_old_gain = gains[num_in_frames - 1];
|
pitch_internal_unittest.cc | 23 double gains[] = {0.6, 0.2, 0.5, 0.4}; local 31 double expected_log_old_gain = log(gains[kNumInputParameters - 1]); 40 GetSubframesPitchParameters(kSamplingRateHz, gains, lags, kNumInputParameters,
|
/external/webrtc/webrtc/modules/audio_coding/codecs/isac/main/source/ |
pitch_estimator.h | 29 double *gains); 41 double *gains); 47 double *gains); 53 double *gains); 60 double *gains);
|
pitch_filter.c | 101 * damper_state_dg : state of damping filter for different trial gains. 214 /* Update filter parameters based on the pitch-gains and pitch-lags. */ 240 * Filter a frame of 30 millisecond, given pitch-lags and pitch-gains. 245 * gains : pointer to pitch-gians, 4 gains per frame. 252 * pitch-gains. 261 double* lags, double* gains, PitchFilterOperation mode, 294 gains[n] *= -kEnhancer; 305 old_gain = gains[0]; 319 gain_delta = (gains[m] - old_gain) / PITCH_GRAN_PER_SUBFRAME [all...] |
/cts/apps/CameraITS/tests/scene1/ |
test_3a.py | 28 sens, exp, gains, xform, focus = cam.do_3a(get_results=True) 30 print "AWB: gains", gains, "transform", xform 34 assert(len(gains) == 4)
|
test_capture_result.py | 49 # The camera HAL may not support different gains for two G channels. 64 "android.colorCorrection.gains": manual_gains, 116 gains = cap_res["android.colorCorrection.gains"] 123 print "Gains:", gains 137 assert(len(gains) == 4) 139 assert(all([g > 0 for g in gains])) 143 assert(any([not is_close_float(gains[i], manual_gains[i]) 163 gains = cap_res["android.colorCorrection.gains" [all...] |
test_auto_vs_manual.py | 47 sens, exp, gains, xform, focus = cam.do_3a(get_results=True) 50 print "AWB gains", gains 61 gains_a = cap_auto["metadata"]["android.colorCorrection.gains"] 62 print "Auto gains:", gains_a 68 req["android.colorCorrection.gains"] = gains 74 gains_m1 = cap_man1["metadata"]["android.colorCorrection.gains"] 75 print "Manual wb gains:", gains_m1 89 gains_m2 = cap_man2["metadata"]["android.colorCorrection.gains"] [all...] |
/external/libopus/silk/float/ |
residual_energy_FLP.c | 95 const silk_float gains[], /* I Quantization gains */ 109 nrgs[ 0 ] = ( silk_float )( gains[ 0 ] * gains[ 0 ] * silk_energy_FLP( LPC_res_ptr + 0 * shift, subfr_length ) ); 110 nrgs[ 1 ] = ( silk_float )( gains[ 1 ] * gains[ 1 ] * silk_energy_FLP( LPC_res_ptr + 1 * shift, subfr_length ) ); 114 nrgs[ 2 ] = ( silk_float )( gains[ 2 ] * gains[ 2 ] * silk_energy_FLP( LPC_res_ptr + 0 * shift, subfr_length ) ); 115 nrgs[ 3 ] = ( silk_float )( gains[ 3 ] * gains[ 3 ] * silk_energy_FLP( LPC_res_ptr + 1 * shift, subfr_length ) ) [all...] |
/cts/apps/CameraITS/pymodules/its/ |
dng.py | 23 def compute_cm_fm(illuminant, gains, ccm, cal): 27 standard A illuminant, the HAL will produce the WB gains and transform, 28 in the android.colorCorrection.gains and android.colorCorrection.transform 32 This function is used to take the per-unit gains, ccm, and calibration 41 gains: White balance gains, as a list of 4 floats. 87 # G is formed by taking the r,g,b gains and putting them into a 89 G = numpy.array([[gains[0],0,0], [0,gains[1],0], [0,0,gains[3]]] [all...] |
/frameworks/base/media/java/android/media/ |
AudioMixPort.java | 35 int[] formats, AudioGain[] gains) { 37 formats, gains);
|
AudioPort.java | 81 int[] formats, AudioGain[] gains) { 90 mGains = gains; 159 public AudioGain[] gains() { method in class:AudioPort
|
AudioDevicePort.java | 41 int[] formats, AudioGain[] gains, int type, String address) { 45 deviceName, samplingRates, channelMasks, channelIndexMasks, formats, gains);
|
/external/libopus/silk/fixed/ |
residual_energy_FIX.c | 42 const opus_int32 gains[ MAX_NB_SUBFR ], /* I Quantization gains */ 82 /* Apply the squared subframe gains */ 84 /* Fully upscale gains and energies */ 86 lz2 = silk_CLZ32( gains[ i ] ) - 1; 88 tmp32 = silk_LSHIFT32( gains[ i ], lz2 ); 90 /* Find squared gains */
|
/external/libopus/celt/mips/ |
celt_mipsr1.h | 68 static const opus_val16 gains[3][3] = { local 81 g00 = MULT16_16_P15(g0, gains[tapset0][0]); 82 g01 = MULT16_16_P15(g0, gains[tapset0][1]); 83 g02 = MULT16_16_P15(g0, gains[tapset0][2]); 84 g10 = MULT16_16_P15(g1, gains[tapset1][0]); 85 g11 = MULT16_16_P15(g1, gains[tapset1][1]); 86 g12 = MULT16_16_P15(g1, gains[tapset1][2]);
|
/external/webrtc/webrtc/modules/audio_processing/agc/legacy/ |
digital_agc.c | 30 // gains = round(2^16*10.^(0.05 * (MinGain + B * ( log(exp(-Knee*A)+exp(-Knee*B)) - log(1+exp(-Knee*B)) ) / log(1/(1+exp(Knee*B)))))); 31 // fprintf(1, '\t%i, %i, %i, %i,\n', gains); 33 // in = 10*log10(lvl); out = 20*log10(gains/65536); 300 // array for gains (one value per ms, incl start & end) 301 int32_t gains[11]; local 416 gains[0] = stt->gain; 457 gains[k + 1] = stt->gainTable[zeros] + (tmp32 >> 12); 507 if ((gains[k + 1] - stt->gainTable[0]) > 8388608) 510 tmp32 = (gains[k + 1] - stt->gainTable[0]) >> 8; 514 tmp32 = (gains[k+1] - stt->gainTable[0]) * (178 + gain_adj) [all...] |
/cts/tests/camera/src/android/hardware/camera2/cts/rs/ |
raw_converter.rs | 43 // Interpolate gain map to find per-channel gains at a given pixel 218 float4 gains = 1.f; 220 gains = getGain(i, j); 226 g = gains.x; 230 g = gains.y; 234 g = gains.z; 238 g = gains.w; 243 g = gains.y; 247 g = gains.x; 251 g = gains.w [all...] |
/cts/apps/CameraITS/tests/inprog/ |
test_3a_remote.py | 50 gains = cap["metadata"]["android.colorCorrection.gains"] 58 print "Gains:", gains
|
test_param_edge_mode.py | 39 sens, exp, gains, xform, focus = cam.do_3a(get_results=True)
|
/frameworks/av/services/audiopolicy/ |
audio_policy.conf | 51 # - one or more "gains" sections can be present in a device descriptor section. 57 # gains { 94 gains {
|
/cts/apps/CameraITS/tests/inprog/scene2/ |
test_dng_tags.py | 47 gains = cap["metadata"]["android.colorCorrection.gains"] 51 print "HAL reported gains:\n", numpy.array(gains) 60 # the HAL-reported WB gains, CCM, and calibration matrix. 61 cm, fm = its.dng.compute_cm_fm(dng_illum[i], gains, ccm, cal)
|
/cts/apps/CameraITS/tests/dng_noise_model/ |
dng_noise_model.py | 41 """Capture a set of raw images with increasing gains and measure the noise. 196 gains = np.asarray([s[0] for s in samples]) 203 digital_gains = np.maximum(gains/sens_max_analog, 1) 208 ad = gains*means 210 cd = gains*gains 216 # by the gains. 217 a /= (np.tile(gains, (a.shape[1], 1)).T) 218 b /= gains
|
/external/libopus/celt/ |
celt.c | 191 static const opus_val16 gains[3][3] = { local 203 g00 = MULT16_16_P15(g0, gains[tapset0][0]); 204 g01 = MULT16_16_P15(g0, gains[tapset0][1]); 205 g02 = MULT16_16_P15(g0, gains[tapset0][2]); 206 g10 = MULT16_16_P15(g1, gains[tapset1][0]); 207 g11 = MULT16_16_P15(g1, gains[tapset1][1]); 208 g12 = MULT16_16_P15(g1, gains[tapset1][2]);
|
/external/webrtc/webrtc/modules/audio_coding/codecs/ilbc/ |
cb_search.c | 56 int16_t gains[CB_NSTAGES+1]; local 184 gains[0] = 16384; 290 (int16_t)WEBRTC_SPL_ABS_W16(gains[stage]), stage, &gain_index[stage]); 346 gains[stage+1] = bestGain; 369 tmp = (int16_t)((gains[1] * gains[1]) >> 14); 373 tmpW32 = ((int32_t)(gains[1]-1))<<1;
|
/cts/apps/CameraITS/tests/ |
tutorial.py | 100 sens, exp, gains, xform, focus = cam.do_3a(get_results=True) 102 print "AWB: gains", gains, "transform", xform 110 # request. Here, the AWB gains and transform (CCM) are being used. 119 req["android.colorCorrection.gains"] = gains
|