/external/tensorflow/tensorflow/compiler/xla/service/cpu/ |
ir_emission_utils.cc | 105 dnums.input_feature_dimension() == input_shape.dimensions_size() - 1 && 108 output_shape.dimensions_size() - 1 && 110 kernel_shape.dimensions_size() - 2 && 112 kernel_shape.dimensions_size() - 1;
|
cpu_instruction_fusion.cc | 49 hlo->opcode() == HloOpcode::kDot && hlo_shape.dimensions_size() == 2 && 138 if (output_shape.dimensions_size() == 2) {
|
cpu_layout_assignment.cc | 82 std::vector<int64> dimension_order(new_shape.dimensions_size()); 90 std::vector<int64> dimension_order(new_shape.dimensions_size());
|
parallel_loop_emitter.cc | 41 const int64 num_dims = shape_.dimensions_size();
|
/external/tensorflow/tensorflow/compiler/xla/client/lib/ |
sorting.cc | 28 int last_dim = input_shape.dimensions_size() - 1; 41 std::vector<int64> start_indices(input_shape.dimensions_size(), 0); 44 std::vector<int64> strides(input_shape.dimensions_size(), 1);
|
quantize.h | 124 std::vector<int64> shift_transpose_dimensions(shape.dimensions_size()); 128 shape.dimensions_size()); 159 std::vector<int64> transpose_dimensions(shape.dimensions_size()); 176 std::vector<int64> result_dimensions(shape.dimensions_size());
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
gather_expander.cc | 33 if (start_indices_shape.dimensions_size() == index_vector_dim) { 37 if (index_vector_dim == (start_indices_shape.dimensions_size() - 1)) { 42 permutation.reserve(start_indices_shape.dimensions_size()); 43 for (int64 i = 0, e = start_indices_shape.dimensions_size(); i < e; i++) { 63 index_vector_dim == start_indices->shape().dimensions_size(); 73 if (shape.dimensions_size() == index_dims_in_start_indices) { 80 shape.dimensions_size() - index_dims_in_start_indices); 90 batch_dim_bounds.reserve(start_indices_shape.dimensions_size()); 91 for (int64 i = 0, e = start_indices_shape.dimensions_size(); i < e; i++) { 159 bool has_scalar_indices = start_indices->shape().dimensions_size() == 1 [all...] |
scatter_expander.cc | 35 if (scatter_indices_shape.dimensions_size() == index_vector_dim) { 39 if (index_vector_dim == (scatter_indices_shape.dimensions_size() - 1)) { 44 permutation.reserve(scatter_indices_shape.dimensions_size()); 45 for (int64 i = 0, e = scatter_indices_shape.dimensions_size(); i < e; i++) { 70 index_vector_dim == scatter_indices->shape().dimensions_size(); 80 if (shape.dimensions_size() == index_dims_in_scatter_indices) { 87 shape.dimensions_size() - index_dims_in_scatter_indices); 117 int64 num_scatter_dims = scatter_indices_shape.dimensions_size(); 118 if (index_vector_dim < scatter_indices_shape.dimensions_size()) { 229 bool has_scalar_indices = scatter_indices->shape().dimensions_size() == 1 [all...] |
hlo_creation_utils.cc | 315 CHECK_GE(operand_shape.dimensions_size(), n); 322 new_shape_dims.reserve(operand_shape.dimensions_size() - n + 1); 340 new_shape_dims.reserve(n + operand_shape.dimensions_size()); 348 CHECK_GT(operand->shape().dimensions_size(), 0); 353 operand->shape().dimensions_size() - 1); 370 new_shape_dim_bounds.reserve(input_shape.dimensions_size() - 373 for (int64 i = input_shape.dimensions_size() - 1; i >= 0; i--) { 394 operand_shape.dimensions_size() + dims_to_insert.size(); 424 CHECK_EQ(operand->shape().dimensions_size(), 1);
|
shape_inference.cc | 163 if (window.dimensions_size() != base_shape.rank()) { 166 window.dimensions_size(), base_shape.rank()); 169 std::vector<int64> output_dimensions(window.dimensions_size()); 170 std::vector<bool> output_is_dynamic(window.dimensions_size()); 171 for (int64 i = 0; i < window.dimensions_size(); ++i) { 495 if (operand_shape.rank() != padding_config.dimensions_size()) { 521 for (int64 i = 0; i < operand_shape.dimensions_size(); ++i) { [all...] |
convolution_group_converter.cc | 106 int64 num_dims = shape.dimensions_size(); 286 for (int64 i = 0; i < new_convolution->shape().dimensions_size(); ++i) { 363 for (int64 i = 0; i < filter->shape().dimensions_size(); ++i) { 476 std::vector<int64> slice_strides(filter->shape().dimensions_size(), 1); 477 std::vector<int64> filter_slice_starts(filter->shape().dimensions_size(), 483 activation->shape().dimensions_size(), 0);
|
indexed_array_analysis.cc | 203 std::vector<IndexComponent> simulated_index(a->shape().dimensions_size(), 256 if (dim_numbers.index_vector_dim() != indices->shape().dimensions_size()) { 278 for (int64 i = 0, e = source->shape().dimensions_size(); i < e; i++) { 292 for (int64 i = 0, e = shape.dimensions_size(); i < e; i++) { 478 for (int64 i = 0, e = source_shape.dimensions_size(); i < e; i++) { 500 for (int64 i = 0, e = shape.dimensions_size(); i < e; i++) { 536 operand->shape().dimensions_size()); 623 for (int64 i = 0, e = shape.dimensions_size(); i < e; i++) { [all...] |
hlo_evaluator.cc | 783 int64 output_rank = output_shape.dimensions_size(); 834 for (int64 i = 0; i < output_shape.dimensions_size(); i++) { 839 for (int64 i = 0; i < input_shape.dimensions_size(); i++) { 852 index_vector_index_.resize(start_indices_.shape().dimensions_size()); 853 input_index_.resize(input_shape.dimensions_size()); [all...] |
hlo_element_type_converter_test.cc | 165 ASSERT_EQ(shape.dimensions_size(), 3);
|
hlo_sharding.cc | 168 return std::vector<int64>(shape.dimensions_size(), 0); 171 CHECK_EQ(shape.dimensions_size(), tile_assignment_.num_dimensions()); 190 CHECK_EQ(shape.dimensions_size(), tile_assignment_.num_dimensions()); 439 for (int64 i = 0; i < shape.dimensions_size(); ++i) {
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
cudnn_conv_padding_legalization.cc | 63 MakeNoPaddingConfig(input->shape().dimensions_size()); 94 std::vector<int64> start_indices(input->shape().dimensions_size(), 0); 97 std::vector<int64> strides(input->shape().dimensions_size(), 1); 132 for (size_t i = 0; i < kernel->shape().dimensions_size(); ++i) { 167 for (size_t i = 0; i < new_conv_window.dimensions_size(); ++i) { 225 for (size_t i = 0; i < backward_conv->window().dimensions_size(); ++i) { 290 for (size_t i = 0; i < backward_conv->window().dimensions_size(); ++i) { 347 std::vector<int64> start_indices(new_backward_conv->shape().dimensions_size(), 352 std::vector<int64> strides(new_backward_conv->shape().dimensions_size(), 1LL); 353 for (size_t i = 0; i < backward_conv->window().dimensions_size(); ++i) [all...] |
cudnn_conv_pad_for_tensor_cores.cc | 101 std::vector<int64> start_indices(result_shape.dimensions_size(), 0); 104 std::vector<int64> strides(result_shape.dimensions_size(), 1);
|
fft_thunk.cc | 125 for (int i = 0; i < input_shape_.dimensions_size() - fft_rank; ++i) { 137 auto dim_offset = input_shape_.dimensions_size() - fft_rank + i;
|
/external/tensorflow/tensorflow/compiler/xla/ |
index_util.cc | 30 DCHECK_EQ(shape.dimensions_size(), multi_index.size()); 104 std::vector<int64> multi_index(shape.dimensions_size());
|
text_literal_reader.cc | 113 if (coordinate_values.size() != shape.dimensions_size()) { 117 shape.dimensions_size(), coordinate_values.size(), line);
|
shape.cc | 26 dimensions_.reserve(shape_proto.dimensions_size()); 31 // dimensions_size. Since C++ is evil, and we have no good way of bailing out 35 if (shape_proto.dimensions_size() != 45 shape_proto.dimensions_size(), shape_proto.is_dynamic_dimension_size()); 61 proto.mutable_dimensions()->Reserve(dimensions_size());
|
literal_comparison.cc | 167 if (dimension == expected.shape().dimensions_size()) { 677 std::vector<int64> multi_index(expected.shape().dimensions_size(), 0); [all...] |
/external/tensorflow/tensorflow/compiler/tf2xla/lib/ |
scatter.cc | 140 ? indices_shape.dimensions_size() - 1 141 : indices_shape.dimensions_size());
|
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
conv_op_helpers.cc | 47 int num_dims = shape.dimensions_size(); 101 filter_shape.dimensions(filter_shape.dimensions_size() - 1); 130 int64 input_feature_dim = filter_shape.dimensions_size() - 2; 131 int64 output_feature_dim = filter_shape.dimensions_size() - 1; 161 {filter_shape.dimensions_size() - 2}), 258 if (input_shape.dimensions_size() != num_dims) { 262 if (filter_shape.dimensions_size() != num_dims) {
|
/external/tensorflow/tensorflow/c/eager/ |
c_api_debug.cc | 122 int rank = padded_shape.dimensions_size();
|