/external/tensorflow/tensorflow/python/tpu/ |
tpu_infeed_test.py | 36 self.assertEqual(i.tuple_shapes, None) 43 self.assertEqual(i.tuple_shapes, None) 45 i = tpu_feed.InfeedQueue(tuple_shapes=[[1], [2, 3]]) 48 self.assertEqual(i.tuple_shapes, [[1], [2, 3]]) 53 self.assertEqual(i.tuple_shapes, None) 62 i = tpu_feed.InfeedQueue(number_of_tuple_elements=2, tuple_shapes=[[1]]) 66 i = tpu_feed.InfeedQueue(tuple_shapes=[[1], [2, 3]], shard_dimensions=[1]) 78 self.assertEqual(i.tuple_shapes, [[1], [2, 3]]) 80 self.assertEqual(i.tuple_shapes, [[1, 2], [3, 4]]) 90 self.assertEqual(i.tuple_shapes, [[6], [3, 18]] [all...] |
tpu_feed.py | 131 tuple_shapes=None, 144 tuple_shapes: if not None, a list of shapes of the elements of the queue. 152 number_of_tuple_arguments, tuple_types, tuple_shapes, and 154 tuple_shapes, or shard_dimensions is not equal to 157 TypeError: if any element of tuple_types or tuple_shapes can't 167 elif tuple_shapes is not None: 168 number_of_tuple_elements = len(tuple_shapes) 187 if tuple_shapes is not None: 188 self.set_tuple_shapes(tuple_shapes) 201 if self.tuple_shapes is not None 251 def tuple_shapes(self): member in class:InfeedQueue [all...] |
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
infeed_thunk.cc | 43 TF_RET_CHECK(infeed_shape.tuple_shapes().size() == 2) 45 TF_RET_CHECK(infeed_shape.tuple_shapes(1).IsToken()) 48 ShapeUtil::Equal(infeed_buffers.shape(), infeed_shape.tuple_shapes(0))) 50 << ShapeUtil::HumanStringWithLayout(infeed_shape.tuple_shapes(0))
|
cudnn_batchnorm_thunk.cc | 156 CHECK(LayoutUtil::LayoutsInShapesEqual(hlo->shape().tuple_shapes(0), 158 for (const auto& tuple_shape : hlo->shape().tuple_shapes()) { 172 hlo_instruction()->shape().tuple_shapes(0), feature_index_); 238 CHECK(LayoutUtil::LayoutsInShapesEqual(hlo->shape().tuple_shapes(0), 240 CHECK(LayoutUtil::LayoutsInShapesEqual(hlo->shape().tuple_shapes(0), 242 for (const auto& tuple_shape : hlo->shape().tuple_shapes()) { 257 hlo_instruction()->shape().tuple_shapes(0), feature_index_);
|
cudnn_conv_pad_for_tensor_cores.cc | 77 CHECK_EQ(0, conv->shape().tuple_shapes(1).dimensions(0)) 85 const Shape& result_shape = conv->shape().tuple_shapes(0); 126 const Shape& result_shape = conv->shape().tuple_shapes(0); 141 Shape new_result_shape = conv->shape().tuple_shapes(0);
|
cudnn_batchnorm_rewriter.cc | 128 libcall->shape().tuple_shapes(2), libcall, 2)); 147 libcall->shape().tuple_shapes(0), libcall, 0)), 149 libcall->shape().tuple_shapes(1), libcall, 1)),
|
cudnn_conv_algorithm_picker.cc | 143 instr->shape().tuple_shapes(0).element_type() == xla::F16; 222 &stream, ShapeUtil::ByteSizeOf(instr->shape().tuple_shapes(0)))); 380 {instr->shape().tuple_shapes(0), 401 new_call_shape.tuple_shapes(0), new_call, 0)),
|
cudnn_fused_conv_rewriter.cc | 210 int64 num_output_feature = conv->shape().tuple_shapes(0).dimensions( 240 return HloInstruction::CreateGetTupleElement(conv->shape().tuple_shapes(0),
|
gpu_layout_assignment_test.cc | 246 batchnorm->shape().tuple_shapes(0).layout())) 328 batchnorm->shape().tuple_shapes(0).layout())) 332 batchnorm->shape().tuple_shapes(0).layout()))
|
cudnn_conv_padding_legalization.cc | 287 Shape backward_conv_shape = backward_conv->shape().tuple_shapes(0); 341 new_backward_conv_call->shape().tuple_shapes(1),
|
cudnn_conv_runner.cc | 331 const auto& conv_result_shape = conv->shape().tuple_shapes(0); 417 conv->shape().tuple_shapes(0).element_type();
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
tuple_util.cc | 34 input_shape.tuple_shapes(i), input_tuple, i))); 53 input_shape.tuple_shapes(i), input_tuple, i)));
|
hlo_module_dce.cc | 74 while_body_param->shape().tuple_shapes(i), while_body_param,
|
while_loop_simplifier.cc | 174 while_init->shape().tuple_shapes(old_idx)); 257 while_init->shape().tuple_shapes(old_idx), while_init, old_idx))); 293 new_while_op->shape().tuple_shapes(gte_idx), new_while_op, 298 while_init->shape().tuple_shapes(old_idx), while_init, old_idx))); 348 new_while_shape_elems.push_back(while_shape.tuple_shapes(i)); 371 while_shape.tuple_shapes(i), instr, i))); 388 while_shape.tuple_shapes(i), instr, j))); 593 const Shape& subshape = desired_shape.tuple_shapes(i); 611 desired_shape.tuple_shapes(i), new_instrs); 631 const Shape& subshape = shape.tuple_shapes(i) [all...] |
/external/tensorflow/tensorflow/compiler/xla/ |
layout_util_test.cc | 270 EXPECT_TRUE(shape.tuple_shapes(0).has_layout()); 271 EXPECT_TRUE(shape.tuple_shapes(2).tuple_shapes(1).has_layout()); 276 EXPECT_FALSE(shape.tuple_shapes(0).has_layout()); 277 EXPECT_FALSE(shape.tuple_shapes(2).tuple_shapes(1).has_layout()); 297 EXPECT_FALSE(LayoutUtil::Equal(shape.tuple_shapes(0).layout(), 298 shape.tuple_shapes(1).layout())); 300 EXPECT_TRUE(LayoutUtil::Equal(shape.tuple_shapes(0).layout(), 301 shape.tuple_shapes(1).layout())) [all...] |
shape.cc | 50 for (const ShapeProto& element_shape : shape_proto.tuple_shapes()) { 69 for (const Shape& shape : tuple_shapes()) { 123 lhs.tuple_shapes(), rhs.tuple_shapes(),
|
shape_util.cc | 330 absl::c_any_of(shape.tuple_shapes(), 347 TF_DCHECK_OK(ValidateShapeWithOptionalLayout(shape.tuple_shapes(index))); 348 return shape.tuple_shapes(index); 365 std::vector<Shape> new_elements(tuple.tuple_shapes().begin() + start, 366 tuple.tuple_shapes().begin() + limit); 395 for (const Shape& element_shape : shape.tuple_shapes()) { 406 for (const Shape& element_shape : shape.tuple_shapes()) { 427 for (const Shape& elem_shape : shape.tuple_shapes()) { 451 for (const Shape& elem_shape : shape.tuple_shapes()) { 632 for (auto& element_shape : shape.tuple_shapes()) { [all...] |
layout_util.cc | 180 for (auto& element_shape : shape.tuple_shapes()) { 309 return absl::c_all_of(shape.tuple_shapes(), 388 TF_RETURN_IF_ERROR(CopyLayoutInternal(src.tuple_shapes(i), 421 if (!LayoutsInShapesEqual(lhs.tuple_shapes(i), rhs.tuple_shapes(i))) {
|
shape.h | 113 const Shape& tuple_shapes(int index) const { return tuple_shapes_.at(index); } function in class:xla::Shape 120 const std::vector<Shape>& tuple_shapes() const { return tuple_shapes_; } function in class:xla::Shape
|
shape_tree.h | 401 current_count += CountSubshapes(shape.tuple_shapes(i)); 434 InitChildren(shape.tuple_shapes(i), init_value, &nodes_.back(), 470 InitChildren(shape.tuple_shapes(i), &nodes_.back(),
|
xla_data.proto | 73 // tuple_shapes fields. 77 // primitive type will have empty dimensions and tuple_shapes fields. 180 // Tuples are a special case in that they have rank zero and have tuple_shapes 204 repeated ShapeProto tuple_shapes = 4;
|
/external/tensorflow/tensorflow/compiler/xla/client/lib/ |
testing.cc | 42 for (const Shape& s : shape.tuple_shapes()) { 56 for (const Shape& s : shape.tuple_shapes()) {
|
/external/tensorflow/tensorflow/compiler/xla/python_api/ |
xla_shape.py | 62 component_message = self.message.tuple_shapes.add() 79 raise ValueError('Tuple shape has no dimensions. Try tuple_shapes()?') 82 def tuple_shapes(self): member in class:Shape 92 raise ValueError('tuple_shapes() called on a non-tuple shape')
|
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
while_op.cc | 162 body.xla_output_shape.tuple_shapes(i); 168 body.xla_output_shape.tuple_shapes(ctx->num_inputs()); 395 arg.shape = body_output_shape.tuple_shapes(i); 491 GetTensorListBufferShape(body_input_shape.tuple_shapes(i),
|
/external/tensorflow/tensorflow/compiler/xla/python/ |
xla_client.py | 478 def tuple_shape(tuple_shapes): 480 if (not isinstance(tuple_shapes, (tuple, list)) or 481 not all(isinstance(t, Shape) for t in tuple_shapes)): 482 raise TypeError('tuple_shapes must be a tuple of Shapes') 483 return Shape(tuple_shapes, tuple) 536 def tuple_shapes(self): member in class:Shape 578 children = tuple(child.map_leaves(f) for child in self.tuple_shapes()) 616 for shape in self.tuple_shapes(): 617 shape.serialize(proto.tuple_shapes.add()) [all...] |