/external/tensorflow/tensorflow/compiler/xla/ |
layout_util.cc | 392 if (ShapeUtil::TupleElementCount(src) != 393 ShapeUtil::TupleElementCount(*dst)) { 397 for (int64 i = 0; i < ShapeUtil::TupleElementCount(src); ++i) { 430 if (ShapeUtil::TupleElementCount(lhs) != 431 ShapeUtil::TupleElementCount(rhs)) { 434 for (int i = 0; i < ShapeUtil::TupleElementCount(lhs); ++i) {
|
shape_util.cc | 326 return IsTuple(shape) && TupleElementCount(shape) == 0; 333 /* static */ int64 ShapeUtil::TupleElementCount(const Shape& shape) { 341 CHECK_GT(TupleElementCount(shape), index); 350 CHECK_LE(start, TupleElementCount(tuple)); 351 CHECK_LE(limit, TupleElementCount(tuple)); 904 for (int64 i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { 921 for (int64 i = 0; i < ShapeUtil::TupleElementCount(*shape); ++i) { [all...] |
shape_tree.h | 474 for (int i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { 485 for (int i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) {
|
shape_util.h | 412 static int64 TupleElementCount(const Shape& shape); 415 // Precondition: IsTuple(shape) && TupleElementCount(shape) > index
|
literal_util.cc | 254 for (int i = 0; i < ShapeUtil::TupleElementCount(shape()); ++i) { [all...] |
literal_util_test.cc | [all...] |
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
gpu_transfer_manager.cc | 70 buffers.reserve(ShapeUtil::TupleElementCount(shape)); 77 for (int64 i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) {
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
hlo_element_type_converter.cc | 69 for (int64 i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { 90 for (int64 i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) {
|
generic_transfer_manager.cc | 58 TF_RET_CHECK(elements.size() == ShapeUtil::TupleElementCount(shape));
|
allocation_tracker.cc | 117 i < ShapeUtil::TupleElementCount(shaped_buffer->on_device_shape());
|
hlo_sharding.h | 90 CHECK_EQ(flattened_list.size(), ShapeUtil::TupleElementCount(tuple_shape));
|
transfer_manager.cc | 131 for (int64 i = 0; i < ShapeUtil::TupleElementCount(device_subshape);
|
while_loop_simplifier.cc | 332 const int64 tuple_size = ShapeUtil::TupleElementCount(while_init->shape());
|
hlo_computation.cc | 483 for (int64 i = 0; i < ShapeUtil::TupleElementCount(instruction->shape());
|
algebraic_simplifier.cc | 500 elems.reserve(ShapeUtil::TupleElementCount(literal.shape())); 501 for (int i = 0; i < ShapeUtil::TupleElementCount(literal.shape()); ++i) { [all...] |
/external/tensorflow/tensorflow/compiler/xla/tests/ |
literal_test_util.cc | 50 if (ShapeUtil::TupleElementCount(expected) != 51 ShapeUtil::TupleElementCount(actual)) { 54 << ShapeUtil::TupleElementCount(expected) 56 << ShapeUtil::TupleElementCount(actual); 318 for (int i = 0; i < ShapeUtil::TupleElementCount(expected.shape()); ++i) { 589 const int64 tuple_elements = ShapeUtil::TupleElementCount(shape); 697 for (int64 i = 0; i < ShapeUtil::TupleElementCount(expected.shape()); ++i) {
|
client_test.cc | 98 EXPECT_EQ(2, ShapeUtil::TupleElementCount(result->shape()));
|
local_client_execute_test.cc | 217 EXPECT_EQ(3, ShapeUtil::TupleElementCount(result->on_host_shape())); 246 EXPECT_EQ(2, ShapeUtil::TupleElementCount(result->on_host_shape())); 327 EXPECT_EQ(2, ShapeUtil::TupleElementCount(result->on_host_shape())); [all...] |
/external/tensorflow/tensorflow/compiler/xla/python/ |
numpy_bridge.cc | 119 int num_elements = ShapeUtil::TupleElementCount(shape); 120 dimensions = PyTuple_New(ShapeUtil::TupleElementCount(shape)); 352 int num_elements = ShapeUtil::TupleElementCount(literal.shape());
|
/external/tensorflow/tensorflow/compiler/tf2xla/ |
xla_jit_compiled_cpu_function_test.cc | 127 ASSERT_EQ(ShapeUtil::TupleElementCount(result), 1);
|
/external/tensorflow/tensorflow/compiler/xla/service/llvm_ir/ |
fused_ir_emitter.cc | 143 for (size_t i = 0; i < ShapeUtil::TupleElementCount(tuple->shape()); ++i) {
|
tuple_ops.cc | 49 for (int i = 0; i < ShapeUtil::TupleElementCount(select.GetShape()); ++i) {
|
/external/tensorflow/tensorflow/compiler/xla/service/cpu/ |
cpu_transfer_manager.cc | 114 buffers.reserve(ShapeUtil::TupleElementCount(shape)); 121 for (int64 i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) {
|
parallel_cpu_executable.cc | 91 for (int i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) {
|
/external/tensorflow/tensorflow/compiler/aot/tests/ |
tfcompile_test.cc | 460 ASSERT_EQ(ShapeUtil::TupleElementCount(muladd_result), 2);
|