/external/tensorflow/tensorflow/contrib/lite/examples/ios/simple/ |
ios_image_load.h | 21 int* out_height, int* out_channels);
|
/external/tensorflow/tensorflow/examples/ios/benchmark/ |
ios_image_load.h | 25 int* out_channels);
|
/external/tensorflow/tensorflow/examples/ios/camera/ |
ios_image_load.h | 25 int* out_channels);
|
/external/tensorflow/tensorflow/examples/ios/simple/ |
ios_image_load.h | 25 int* out_channels);
|
/external/tensorflow/tensorflow/tools/graph_transforms/ |
flatten_atrous.cc | 73 const int32 out_channels = filter.dim_size(3); 82 in_channels, out_channels})); 91 for (int c_out = 0; c_out < out_channels; ++c_out) {
|
/external/webrtc/webrtc/common_audio/ |
lapped_transform_unittest.cc | 30 size_t out_channels, 32 RTC_CHECK_EQ(in_channels, out_channels); 33 for (size_t i = 0; i < out_channels; ++i) { 54 size_t out_channels, 56 RTC_CHECK_EQ(in_channels, out_channels);
|
/frameworks/av/media/libstagefright/codecs/mp3dec/test/ |
mp3reader.cpp | 42 uint32_t *out_sampling_rate = NULL, uint32_t *out_channels = NULL , 50 if (out_channels) { 51 *out_channels = 0; 181 if (out_channels) { 184 *out_channels = (channel_mode == 3) ? 1 : 2; 392 header, &frame_size, &sample_rate, NULL /*out_channels*/,
|
/external/webrtc/webrtc/modules/audio_coding/test/ |
TestStereo.h | 78 void Run(TestPackStereo* channel, int in_channels, int out_channels,
|
TestStereo.cc | 727 void TestStereo::Run(TestPackStereo* channel, int in_channels, int out_channels, 782 if ((rec_size != pack_size_bytes_ * out_channels)
|
/hardware/google/av/media/codecs/mp3/ |
C2SoftMp3Dec.h | 27 uint32_t *out_channels = nullptr,
|
C2SoftMp3Dec.cpp | 167 uint32_t *out_channels = nullptr, 172 if (out_channels) *out_channels = 0; 267 if (out_channels) { 270 *out_channels = (channel_mode == 3) ? 1 : 2;
|
/external/tensorflow/tensorflow/contrib/kfac/examples/ |
convnet.py | 57 def conv_layer(layer_id, inputs, kernel_size, out_channels): 66 out_channels: int. Number of output features per pixel. 69 preactivations: Tensor of shape [num_examples, width, height, out_channels]. 71 activations: Tensor of shape [num_examples, width, height, out_channels]. 77 out_channels, 101 Tensor of shape [num_examples, width/stride, height/stride, out_channels]. 150 layer_id=0, inputs=examples, kernel_size=5, out_channels=16) 153 layer_id=2, inputs=act1, kernel_size=5, out_channels=16)
|
/frameworks/av/media/libstagefright/foundation/include/media/stagefright/foundation/ |
avc_utils.h | 100 int *out_sampling_rate = NULL, int *out_channels = NULL,
|
/external/webrtc/webrtc/modules/audio_processing/intelligibility/ |
intelligibility_enhancer.h | 90 size_t out_channels,
|
/frameworks/av/media/libstagefright/foundation/ |
avc_utils.cpp | 622 int *out_sampling_rate, int *out_channels, 630 if (out_channels) { 631 *out_channels = 0; 763 if (out_channels) { 766 *out_channels = (channel_mode == 3) ? 1 : 2;
|
/external/webrtc/webrtc/modules/audio_processing/test/ |
audioproc_float.cc | 44 DEFINE_int32(out_channels, 1, "Number of output channels.");
|
debug_dump_test.cc | 216 const int out_channels = config.num_channels(); local 223 RTC_CHECK_LE(out_channels, channels); 224 for (int channel = 0; channel < out_channels; ++channel) {
|
/external/tensorflow/tensorflow/python/kernel_tests/ |
depthtospace_op_test.py | 230 def compareToTranspose(self, batch_size, in_height, in_width, out_channels, 232 in_channels = out_channels * block_size * block_size
|
/external/tensorflow/tensorflow/contrib/kfac/examples/tests/ |
convnet_test.py | 36 out_channels=5)
|
/external/tensorflow/tensorflow/contrib/kfac/python/ops/ |
fisher_factors.py | 922 outputs_grads: Tensor of shape [batch_size, height, width, out_channels]. 926 out_channels). Represents shape of kernel used in this layer. [all...] |
/hardware/qcom/audio/hal/msm8974/ |
platform.c | 4038 unsigned int out_channels = audio_channel_count_from_out_mask(out->channel_mask); local [all...] |
/external/tensorflow/tensorflow/go/op/ |
wrappers.go | [all...] |