/external/tensorflow/tensorflow/core/kernels/ |
ops_testutil.cc | 44 Tensor* OpsTestBase::GetOutput(int output_index) { 45 CHECK_LT(output_index, context_->num_outputs()); 46 Tensor* output = context_->mutable_output(output_index); 51 if (!managed_outputs_[output_index]) { 59 managed_outputs_[output_index] = managed_output; 61 output = managed_outputs_[output_index];
|
dynamic_stitch_op_gpu.cu.cc | 39 CUDA_1D_KERNEL_LOOP(output_index, output_size) { 40 const int32 slice_id = output_index / slice_size; 41 const int32 slice_offset = output_index % slice_size; 44 output[output_index] = ldg(data_ptrs[input_index] + slice_offset);
|
reduce_join_op.cc | 50 int64 output_index, const gtl::InlinedVector<int32, 8>& dim_list, 54 int64 quotient = output_index; 164 for (int64 output_index = 0; output_index < output_shape.num_elements(); 165 ++output_index) { 167 output_index, unreduced_indices, input_shape, strides); 175 output_flat(output_index) =
|
segment_reduction_ops_gpu.cu.cc | 77 const Index output_index = local 82 CudaAtomicAdd(output + output_index, sum); 84 *(output + output_index) = sum; 95 const Index output_index = local 97 CudaAtomicAdd(output + output_index, sum); 119 const Index output_index = local 121 KernelReductionFunctor()(output + output_index, ldg(input + input_index));
|
dynamic_partition_op.cc | 100 gtl::InlinedVector<int, 32> output_index(num_partitions_); 117 auto oi = output_index[p]; 121 " is not LTE output_index[", p, "] : ", oi)); 123 output_index[p]++; 140 // outputs[p][output_index[p]++] = data[i] 147 auto oi = output_index[p]; 149 errors::InvalidArgument("Size of output_index: ", oi, 155 output_index[p]++;
|
conv_ops_gpu_3.cu.cc | 184 int output_index = index; local 186 Index<3> output_tensor_index = FlatToTensorIndex(output_index, output_dims); 195 output[output_index] = 211 int output_index = index; local 212 Index<3> output_tensor_index = FlatToTensorIndex(output_index, output_dims); 221 output[output_index] = 351 int output_index = output_origin_flat_index + ti * output_dims[2] + tj; local 357 output[output_index] = shared_memory_tile[tj][i_loc]; 358 output_index += output_increment; 363 output[output_index] = shared_memory_tile[tj][i_loc] 379 int output_index = index; local 408 int output_index = index; local [all...] |
deep_conv2d_test.cc | 31 const int output_index = k * cols * cols + l; local 32 matrix_out[output_index_base + output_index] =
|
/external/tensorflow/tensorflow/java/src/main/native/ |
operation_jni.cc | 89 jint output_index) { 96 if (output_index < 0 || output_index >= num_outputs) { 100 output_index, num_outputs); 104 TF_Output output{op, output_index}; 142 jint output_index) { 149 if (output_index < 0 || output_index >= num_outputs) { 153 output_index, num_outputs); 157 return static_cast<jint>(TF_OperationOutputType(TF_Output{op, output_index})); [all...] |
/external/tensorflow/tensorflow/tools/graph_transforms/ |
insert_logging.cc | 82 int32 output_index; local 83 if (!strings::safe_strto32(output_index_string, &output_index)) { 87 node_outputs[name].insert(output_index); 131 for (int output_index : node_outputs[node.name()]) { 132 print_node->add_input(strings::StrCat(node.name(), ":", output_index));
|
/external/tensorflow/tensorflow/contrib/periodic_resample/kernels/ |
periodic_resample_op.h | 33 IndexVecT* target_dimensions, const IndexT& output_index, 42 auto last_reduced_i = output_index; 182 for (tensorflow::int64 output_index = 0; output_index < new_size; 183 ++output_index) { 184 output(output_index) = input(compute_input_index( 185 &target_dimensions, output_index, original_dimensions,
|
/external/tensorflow/tensorflow/contrib/lite/kernels/ |
gather.cc | 70 int output_index = 0; local 72 output_shape->data[output_index++] = input->dims->data[i]; 75 output_shape->data[output_index++] = positions->dims->data[i]; 78 output_shape->data[output_index++] = input->dims->data[i];
|
/external/tensorflow/tensorflow/contrib/tensorrt/ |
tensorrt_test.cc | 105 const int output_index = engine.getBindingIndex(kOutputTensor); local 110 ASSERT_EQ(0, cudaMalloc(&buffers[output_index], sizeof(float))); 122 ASSERT_EQ(0, cudaMemcpyAsync(output, buffers[output_index], sizeof(float), 129 ASSERT_EQ(0, cudaFree(buffers[output_index]));
|
/external/tensorflow/tensorflow/contrib/lite/ |
graph_info.cc | 95 for (int output_index : info_->outputs()) { 96 int output_epoch = tensor_epochs_[output_index]; 98 output_subgraph.output_tensors.push_back(output_index);
|
/external/tensorflow/tensorflow/core/common_runtime/ |
memory_types.cc | 31 int output_index; member in struct:tensorflow::Endpoint 37 x.output_index); 43 return (x.node_id == y.node_id) && (x.output_index == y.output_index);
|
/external/tensorflow/tensorflow/compiler/tf2xla/ |
tf2xla_util.cc | 45 if (id.output_index() < 0) { 46 return errors::InvalidArgument("TensorId output_index must be positive"); 115 "aot_feed_", feed->id().output_index(), "/", feed->id().node_name()); 155 BaseType(feed_node->output_type(info.feed->id().output_index())); 196 std::make_pair(feed.id().node_name(), feed.id().output_index())); 253 return strings::StrCat(id.node_name(), ":", id.output_index());
|
tf2xla.cc | 70 const int output_index = 0; local 92 .Attr("T", BaseType(feed_node->output_type(output_index))) 106 if (edge->src_output() == output_index) { 130 if (id.output_index() >= fetch_node->num_outputs()) { 139 .Input(fetch_node, id.output_index()) 140 .Attr("T", BaseType(fetch_node->output_type(id.output_index())))
|
/external/tensorflow/tensorflow/core/kernels/data/ |
scan_dataset_op.cc | 192 const size_t output_index = i - dataset()->state_types_.size(); variable 193 if (state_and_output[i].dtype() != output_dtypes()[output_index]) { 197 DataTypeString(dataset()->state_types_[output_index]), 200 if (!output_shapes()[output_index].IsCompatibleWith( 204 " (expected ", output_shapes()[output_index].DebugString(),
|
/external/tensorflow/tensorflow/contrib/lite/toco/graph_transformations/ |
quantize.cc | 310 std::size_t output_index, ArrayDataType* quantized_data_type, 312 const auto& output = op.outputs[output_index]; 343 if (output_index == LstmCellOperator::STATE_OUTPUT || 344 output_index == LstmCellOperator::ACTIV_TEMP) { 498 for (std::size_t output_index = 0; output_index < op.outputs.size(); 499 output_index++) { 502 if (ChooseQuantizationForOperatorOutput(this, model, op, output_index, 506 const auto& output = op.outputs[output_index];
|
/external/valgrind/coregrind/m_gdbserver/ |
remote-utils.c | 701 int input_index, output_index; local 703 output_index = 0; 709 if (output_index + 2 > out_maxlen) 711 out_buf[output_index++] = '}'; 712 out_buf[output_index++] = b ^ 0x20; 714 if (output_index + 1 > out_maxlen) 716 out_buf[output_index++] = b; 721 return output_index; 737 int input_index, output_index; local 740 output_index = 0 [all...] |
/external/tensorflow/tensorflow/contrib/copy_graph/python/util/ |
copy_elements.py | 170 output_index = op.outputs.index(org_instance) 171 new_tensor = new_op.outputs[output_index]
|
/external/tensorflow/tensorflow/compiler/xla/service/llvm_ir/ |
ops.cc | 55 // Calculate output_index, where we'll write the value from update. For 58 // output_index[dim] = (start_index[dim] + update_index[dim]) % dim_size. 60 IrArray::Index output_index(rank); 66 output_index[i] = ir_builder->CreateURem( 70 // Do output[output_index] = update[update_index]. 73 output_array.EmitWriteArrayElement(output_index, update_data, ir_builder);
|
/external/tensorflow/tensorflow/contrib/lite/python/ |
op_hint.py | 291 for output_index, tensor in call.outputs.items(): 292 output_names[output_index] = _tensor_name_base(tensor) 293 output_dtypes[output_index] = tensor.dtype.as_datatype_enum
|
/external/tensorflow/tensorflow/python/eager/ |
execution_callbacks.py | 42 output_index, 52 output_index: The 0-based output index of the tensor that contains 59 self._output_index = output_index 94 def output_index(self): member in class:InfOrNanError
|
/external/tensorflow/tensorflow/python/framework/ |
importer.py | 118 # Expected format: 'operation_name:output_index'. 120 output_index = int(components[1]) 123 return components[0], output_index 135 input_op_name, output_index = _ParseTensorName(input_name) 136 return '%s:%d' % (input_op_name, output_index) 649 operation_name, output_index = _ParseTensorName(input_name) 652 source_tensor = list(source_op.values())[output_index] 761 operation_name, output_index = _ParseTensorName(tensor_name) 763 return output_index < len(name_to_op[operation_name].outputs) 782 operation_name, output_index = _ParseTensorName(name [all...] |
/external/tensorflow/tensorflow/core/framework/ |
op_kernel.cc | 403 int output_index) { 407 set_output_ref(output_index, (*params_->inputs)[input_index].mutex_if_ref, 412 int input_index, int output_index, const TensorShape& output_shape, 416 : output_alloc_attr(output_index); 418 input_index, expected_output_dtype(output_index), output_shape, 419 output_memory_type(output_index), output_attr); 422 outputs_[output_index] = TensorValue(new_tensor.release()); 423 *output = outputs_[output_index].tensor; 433 int input_index, output_index, stop; local 443 params_->op_kernel->OutputRange(output_name, &output_index, &stop)) [all...] |