| /external/webrtc/webrtc/common_audio/ |
| audio_util_unittest.cc | 182 const float input_data[kNumChannels][kNumFrames] = {{1.f, 2.f, -1.f, -3.f}}; local 185 input[i] = input_data[i]; 192 EXPECT_THAT(downmixed, ElementsAreArray(input_data[0])); 197 const float input_data[kNumChannels][kNumFrames] = {{1.f, 2.f, -1.f}, local 201 input[i] = input_data[i]; 214 const int16_t input_data[kNumChannels][kNumFrames] = { local 218 input[i] = input_data[i];
|
| /external/protobuf/src/google/protobuf/compiler/ |
| subprocess.cc | 157 string input_data = input.SerializeAsString(); local 191 input_data.data() + input_pos, 192 input_data.size() - input_pos, 196 input_pos = input_data.size(); 201 if (input_pos == input_data.size()) { 360 string input_data = input.SerializeAsString(); 388 int n = write(child_stdin_, input_data.data() + input_pos, 389 input_data.size() - input_pos); 393 input_pos = input_data.size(); 398 if (input_pos == input_data.size()) [all...] |
| /external/protobuf/src/google/protobuf/util/ |
| json_util_test.cc | 242 string input_data = "0123456789"; local 243 for (int input_pattern = 0; input_pattern < (1 << (input_data.size() - 1)); 250 for (int j = 0; j < input_data.length() - 1; ++j) { 252 byte_sink.Append(&input_data[start], j - start + 1); 256 byte_sink.Append(&input_data[start], input_data.length() - start); 258 EXPECT_EQ(input_data, string(buffer, input_data.length())); 262 input_data = "012345678"; 263 for (int input_pattern = 0; input_pattern < (1 << (input_data.size() - 1)) [all...] |
| /external/tensorflow/tensorflow/contrib/tensor_forest/hybrid/core/ops/ |
| k_feature_gradient_op.cc | 41 .Input("input_data: float") 56 input_data: The training batch's features as a 2-d tensor; 57 `input_data[i][j]` gives the j-th feature of the i-th input. 137 const auto input_data = input_data_tensor.tensor<float, 2>(); variable 177 weights_grad(i, j, k) = input_data(i, feature_set[k]);
|
| stochastic_hard_routing_gradient_op.cc | 46 .Input("input_data: float") 76 input_data: The training batch's features as a 2-d tensor; `input_data[i][j]` 119 const Tensor& input_data = context->input(0); variable 126 const int32 num_data = static_cast<int32>(input_data.shape().dim_size(0)); 128 static_cast<int32>(input_data.shape().dim_size(1)); 171 const auto data = input_data.tensor<float, 2>(); 178 const Tensor point = input_data.Slice(i, i + 1);
|
| /external/tensorflow/tensorflow/core/kernels/ |
| adjust_hue_op.cc | 211 auto input_data = input->shaped<float, 2>({channel_count, kChannelSize}); variable 219 [&input_data, &output_data, delta_h](int64 start_channel, 221 const float* p = input_data.data() + start_channel * kChannelSize; 269 const T* input_data = input->flat<T>().data(); variable 272 functor::AdjustHueGPU<T>()(&device, number_of_elements, input_data,
|
| pooling_ops_common.cc | 252 auto input_data = local 266 ->ThenPoolForward(pooling_desc, input_desc, input_data,
|
| adjust_contrast_op.cc | 216 auto input_data = input->shaped<float, 3>({batch, image_size, channels}); variable 221 ReduceMeanAcrossImage(input_data, mean_data, output_data); 226 IncrementWithScaling(input_data, factor->scalar<float>(), output_data);
|
| adjust_saturation_op.cc | 188 auto input_data = input->shaped<float, 2>({channel_count, kChannelSize}); variable 196 [&input_data, &output_data, scale_h](int64 start_channel, 198 const float* p = input_data.data() + start_channel * kChannelSize; 235 const T* input_data = input->flat<T>().data(); variable 238 functor::AdjustSaturationGPU<T>()(&device, number_of_elements, input_data,
|
| lrn_op.cc | 207 auto input_data = StreamExecutorUtil::AsDeviceMemory<T>(in); local 216 input_data, &output_data)
|
| where_op_gpu.cu.h | 145 const T* input_data = input.data(); local 152 /*d_in*/ input_data, 171 /*d_in*/ input_data,
|
| /external/tensorflow/tensorflow/lite/kernels/ |
| audio_spectrogram.cc | 111 const float* input_data = GetTensorData<float>(input); local 125 input_for_channel[i] = input_data[i * channel_count + channel];
|
| mirror_pad.cc | 56 const T* input_data = nullptr; member in struct:tflite::ops::builtin::mirror_pad::__anon46067::EvalData 95 eval_data->output_data[output_index] = eval_data->input_data[flat_index]; 194 eval_data.input_data = GetTensorData<type>(input_tensor); \
|
| /external/tensorflow/tensorflow/lite/toco/graph_transformations/ |
| resolve_constant_transpose.cc | 32 const std::vector<DataType<Type>>& input_data = local 80 input_data.data() + i3 * input_stride_3;
|
| /external/perf_data_converter/src/quipper/ |
| perf_reader_test.cc | 1724 std::stringstream input_data; local [all...] |
| /external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
| image_resize_ops.cc | 266 xla::XlaOp input_data = input; local 287 input_data, {0, in_size[0] - 1, 0, 0}, 290 input_data = xla::ConcatInDim(builder, {input_data, slice}, 1); 296 input_data, {0, 0, in_size[1] - 1, 0}, 300 input_data = xla::ConcatInDim(builder, {input_data, slice}, 2); 320 xla::ConvGeneralDilated(input_data, kernel, dims.stride, 331 input_data, kernel0, {dims.stride[0], 1},
|
| /external/tensorflow/tensorflow/compiler/xla/tests/ |
| dynamic_ops_test.cc | 545 std::unique_ptr<GlobalData> input_data = local 557 {input_data.get(), update_data.get()}, [all...] |
| reduce_test.cc | 100 std::vector<float> input_data(element_count); 102 input_data[i] = rand_r(&seed_) % 3; 104 input_data[i] *= -1; 107 Literal input_literal = LiteralUtil::CreateR1(AsSlice(input_data)); 112 for (float item : input_data) { 119 void RunR1ToR0PredTest(bool and_reduce, absl::Span<const int> input_data) { 120 const int element_count = input_data.size(); 138 Literal input_literal = LiteralUtil::CreateR1(input_data); 143 for (bool item : input_data) { 177 Array2D<uint8> input_data(rows, cols) 821 std::unique_ptr<GlobalData> input_data = local 898 auto input_data = local [all...] |
| reduce_window_test.cc | 345 auto input_data = CreateParameterAndTransferLiteral( local 355 {input_data.get()}, DefaultErrorSpec()); 403 auto input_data = CreateParameterAndTransferLiteral( local 415 {input_data.get()}, DefaultErrorSpec()); 424 auto input_data = CreateParameterAndTransferLiteral( local 436 {input_data.get()}, DefaultErrorSpec()); 445 auto input_data = CreateParameterAndTransferLiteral( local 457 {input_data.get()}, DefaultErrorSpec()); [all...] |
| /external/tensorflow/tensorflow/contrib/tensor_forest/kernels/ |
| model_ops.cc | 20 #include "tensorflow/contrib/tensor_forest/kernels/v4/input_data.h" 176 const Tensor& input_data = context->input(1); variable 182 data_set->set_input_tensors(input_data, sparse_input_indices, 277 const Tensor& input_data = context->input(1); variable 283 data_set->set_input_tensors(input_data, sparse_input_indices,
|
| stats_ops.cc | 20 #include "tensorflow/contrib/tensor_forest/kernels/v4/input_data.h" 241 const Tensor& input_data = context->input(2); variable 251 data_set->set_input_tensors(input_data, sparse_input_indices,
|
| /external/tensorflow/tensorflow/lite/experimental/micro/ |
| micro_interpreter_test.cc | 37 const int32_t* input_data = input->data.i32; local 42 output_data[0] = input_data[0] + weight_data[0];
|
| /external/u-boot/drivers/ata/ |
| sata_sil3114.c | 48 static int input_data (struct sata_ioports *ioaddr, u16 * sect_buf, int words) function 130 input_data (&port[num].ioaddr, iobuf, ATA_SECTOR_WORDS); 501 input_data (&port[device].ioaddr, buff, ATA_SECTOR_WORDS);
|
| /external/tensorflow/tensorflow/python/lib/core/ |
| py_func.cc | 378 PyObject** input_data = reinterpret_cast<PyObject**>(PyArray_DATA(input)); local 380 TF_RETURN_IF_ERROR(PyObjectToString(input_data[i], &tflat(i))); 389 char* input_data = PyArray_BYTES(input); local 392 tflat(i) = string(input_data + i * el_size, el_size);
|
| /external/webrtc/webrtc/modules/audio_coding/acm2/ |
| audio_coding_module_impl.cc | 128 int32_t AudioCodingModuleImpl::Encode(const InputData& input_data) { 139 first_frame_ ? input_data.input_timestamp 142 input_data.input_timestamp - last_timestamp_, 146 last_timestamp_ = input_data.input_timestamp; 153 input_data.audio, input_data.audio_channel * 154 input_data.length_per_channel), 275 InputData input_data; local 277 int r = Add10MsDataInternal(audio_frame, &input_data); 278 return r < 0 ? r : Encode(input_data); [all...] |