/external/webrtc/webrtc/modules/audio_coding/neteq/ |
background_noise_unittest.cc | 20 size_t channels = 1; local 21 BackgroundNoise bgn(channels);
|
merge_unittest.cc | 28 size_t channels = 1; local 29 BackgroundNoise bgn(channels); 33 Expand expand(&bgn, &sync_buffer, &random_vector, &statistics, fs, channels); 34 Merge merge(fs, channels, &expand, &sync_buffer);
|
normal_unittest.cc | 36 size_t channels = 1; local 37 BackgroundNoise bgn(channels); 41 Expand expand(&bgn, &sync_buffer, &random_vector, &statistics, fs, channels); 50 size_t channels = 1; local 51 BackgroundNoise bgn(channels); 56 channels); 60 rtc::scoped_ptr<int16_t[]> mute_factor_array(new int16_t[channels]); 61 for (size_t i = 0; i < channels; ++i) { 64 AudioMultiVector output(channels); 96 size_t channels = 2 local [all...] |
audio_classifier.cc | 42 int channels) { 44 assert((input_length / channels) == kDefaultFrameSizeSamples); 47 assert(channels == 1 || channels == 2); 61 channels,
|
audio_classifier_unittest.cc | 40 size_t channels) { 42 rtc::scoped_ptr<int16_t[]> in(new int16_t[channels * kFrameSize]); 51 while (fread(in.get(), sizeof(int16_t), channels * kFrameSize, audio_file) == 52 channels * kFrameSize) { 54 classifier.Analysis(in.get(), channels * kFrameSize, channels);
|
/external/libpng/contrib/tools/ |
cvtcolor.c | 53 int channels = 0; local 92 ++channels; 95 ++channels; 98 ++channels; 101 ++channels; 108 int components = channels; 114 if (components < channels) 123 if ((channels & 1) == 0) 125 double alpha = c[channels-1]; 128 for (i=0; i<channels-1; ++i) c[i] /= alpha [all...] |
/external/webrtc/webrtc/modules/audio_coding/codecs/opus/ |
opus_inst.h | 20 size_t channels; member in struct:WebRtcOpusEncInst 33 size_t channels; member in struct:WebRtcOpusDecInst
|
/external/webrtc/webrtc/common_audio/ |
audio_ring_buffer.h | 25 // Specify the number of channels and maximum number of frames the buffer will 27 AudioRingBuffer(size_t channels, size_t max_frames); 30 // Copies |data| to the buffer and advances the write pointer. |channels| must 32 void Write(const float* const* data, size_t channels, size_t frames); 34 // Copies from the buffer to |data| and advances the read pointer. |channels| 36 void Read(float* const* data, size_t channels, size_t frames);
|
audio_ring_buffer.cc | 20 AudioRingBuffer::AudioRingBuffer(size_t channels, size_t max_frames) { 21 buffers_.reserve(channels); 22 for (size_t i = 0; i < channels; ++i) 31 void AudioRingBuffer::Write(const float* const* data, size_t channels, 33 RTC_DCHECK_EQ(buffers_.size(), channels); 34 for (size_t i = 0; i < channels; ++i) { 40 void AudioRingBuffer::Read(float* const* data, size_t channels, size_t frames) { 41 RTC_DCHECK_EQ(buffers_.size(), channels); 42 for (size_t i = 0; i < channels; ++i) {
|
/external/webrtc/webrtc/audio/ |
audio_sink.h | 33 size_t channels, 38 channels(channels), 44 size_t channels; // Number of channels in the audio data. member in struct:webrtc::AudioSinkInterface::Data
|
/external/adhd/cras/src/dsp/ |
dsp_util.h | 21 * input - The interleaved input buffer. Every "channels" samples is a frame. 22 * output - Pointers to output buffers. There are "channels" output buffers. 23 * channels - The number of samples per frame. 28 int dsp_util_deinterleave(uint8_t *input, float *const *output, int channels, 35 * input - Pointers to input buffers. There are "channels" input buffers. 36 * output - The interleaved output buffer. Every "channels" samples is a 38 * channels - The number of samples per frame. 43 int dsp_util_interleave(float *const *input, uint8_t *output, int channels,
|
dsp_util.c | 380 int channels, int frames) 382 float *output_ptr[channels]; 386 if (channels == 2) { 392 for (i = 0; i < channels; i++) 396 for (j = 0; j < channels; j++) 402 int channels, int frames) 404 float *output_ptr[channels]; 407 for (i = 0; i < channels; i++) 411 for (j = 0; j < channels; j++, input++) 417 int channels, int frames [all...] |
/external/webrtc/webrtc/test/fuzzers/ |
audio_decoder_opus_fuzzer.cc | 16 const size_t channels = (size % 2) + 1; // 1 or 2 channels. local 17 AudioDecoderOpus dec(channels);
|
/external/adhd/cras/src/server/ |
cras_audio_area.c | 42 if (!(src->channels[src_idx].ch_set & 43 dst->channels[dst_idx].ch_set)) 46 schan = src->channels[src_idx].buf + 47 src_offset * src->channels[src_idx].step_bytes; 48 dchan = dst->channels[dst_idx].buf + 49 dst_offset * dst->channels[dst_idx].step_bytes; 53 dst->channels[dst_idx].step_bytes, 54 src->channels[src_idx].step_bytes, 80 channel_area_set_channel(area->channels, CRAS_CH_FL); 81 channel_area_set_channel(area->channels, CRAS_CH_FR) [all...] |
/external/autotest/client/site_tests/audio_Microphone/ |
audio_Microphone.py | 23 self, filesize, duration, channels, rate, bits=16): 24 expected = duration * channels * (bits / 8) * rate 29 def verify_alsa_capture(self, channels, rate, device, bits=16): 32 recorded_file.name, duration=DURATION, channels=channels, 36 DURATION, channels, rate, bits) 39 def verify_cras_capture(self, channels, rate): 42 recorded_file.name, duration=DURATION, channels=channels, 46 DURATION, channels, rate [all...] |
/external/aac/libPCMutils/src/ |
pcm_utils.cpp | 110 const UINT channels, const UINT frameSize, 114 for (UINT ch = 0; ch < channels; ch++) { 122 const UINT channels, const UINT frameSize, 126 for (UINT ch = 0; ch < channels; ch++) { 134 const UINT channels, const UINT frameSize, 138 for (UINT ch = 0; ch < channels; ch++) { 146 const UINT channels, const UINT frameSize, 148 for (UINT ch = 0; ch < channels; ch++) { 153 In += channels; 159 const UINT channels, const UINT frameSize [all...] |
/external/aac/libPCMutils/include/ |
pcm_utils.h | 110 const UINT channels, const UINT frameSize, 113 const UINT channels, const UINT frameSize, 116 const UINT channels, const UINT frameSize, 120 const UINT channels, const UINT frameSize, 123 const UINT channels, const UINT frameSize, 126 const UINT channels, const UINT frameSize, 129 const UINT channels, const UINT frameSize,
|
/external/tensorflow/tensorflow/lite/examples/label_image/ |
bitmap_helpers.cc | 32 int height, int channels, bool top_down) { 33 std::vector<uint8_t> output(height * width * channels); 40 src_pos = ((height - 1 - i) * row_size) + j * channels; 42 src_pos = i * row_size + j * channels; 45 dst_pos = (i * width + j) * channels; 47 switch (channels) { 65 LOG(FATAL) << "Unexpected number of channels: " << channels; 74 int* height, int* channels, Settings* s) { 99 *channels = bpp / 8 103 << ", " << *channels << "\\n"; local [all...] |
/device/linaro/hikey/hifi/xaf/host-apf/playback/ |
xa_playback.h | 31 unsigned int channels,
|
/external/drrickorang/LoopbackApp/app/src/main/cpp/ |
byte_buffer.h | 34 const char *srcBuffer, size_t frameCount, int channels);
|
/external/mp4parser/isoparser/src/main/java/com/googlecode/mp4parser/authoring/adaptivestreaming/ |
AudioQuality.java | 24 int channels; field in class:AudioQuality
|
/external/wpa_supplicant_8/src/p2p/ |
p2p_utils.c | 87 for (i = 0; i < a->channels; i++) { 88 for (j = 0; j < b->channels; j++) { 91 res->channel[res->channels] = a->channel[i]; 92 res->channels++; 93 if (res->channels == P2P_MAX_REG_CLASS_CHANNELS) 102 * @a: First set of supported channels 103 * @b: Second set of supported channels 104 * @res: Data structure for returning the intersection of support channels 106 * This function can be used to find a common set of supported channels. Both 107 * input channels sets are assumed to use the same country code. If differen [all...] |
/external/tensorflow/tensorflow/core/kernels/ |
resize_nearest_neighbor_op_gpu.cu.cc | 38 const int in_width, const int channels, const int out_height, 43 int c = n % channels; 44 n /= channels; 50 const T* bottom_data_n = bottom_data + n * channels * in_height * in_width; 61 const int idx = (in_y * in_width + in_x) * channels + c; 69 const int in_width, const int channels, const int out_height, 74 int c = n % channels; 75 n /= channels; 81 const T* bottom_data_n = bottom_data + n * channels * in_height * in_width; 90 const int idx = (in_y * in_width + in_x) * channels + c 168 const int channels = input.dimension(3); local 215 const int channels = input.dimension(3); local [all...] |
resize_bilinear_op_gpu.cu.cc | 39 int channels, int out_height, 42 // out_idx = c + channels * (x + out_width * (y + out_height * b)) 44 const int c = idx % channels; 45 idx /= channels; 66 channels + 70 channels + 74 channels + 78 channels + 91 int channels, int resized_height, int resized_width, T* output_grad) { 93 // in_idx = c + channels * (x + resized_width * (y + resized_height * b) 274 const int channels = images.dimension(3); local 310 const int channels = output_grad.dimension(3); local [all...] |
avgpooling_op_gpu.cu.cc | 46 const int channels, const int pooled_height, 54 const int c = index % channels; 55 const int w = index / channels % width + pad_l; 56 const int h = (index / channels / width) % height + pad_t; 57 const int n = index / channels / width / height; 64 top_diff + n * pooled_height * pooled_width * channels + c; 75 gradient += top_diff_slice[(ph * pooled_width + pw) * channels] / 86 const int channels, const int pooled_height, 92 int x_size = num * height * width * channels; 97 channels, pooled_height, pooled_width, kernel_h, kernel_w, stride_h [all...] |