/external/tensorflow/tensorflow/compiler/xla/client/ |
executable_build_options.cc | 43 const Shape& shape_with_layout) { 45 result_layout_ = shape_with_layout;
|
client.cc | 40 const GlobalData& data, const Shape* shape_with_layout) { 43 if (shape_with_layout != nullptr) { 44 *request.mutable_shape_with_layout() = *shape_with_layout; 117 const Shape* shape_with_layout, int64 replica_id, 124 if (shape_with_layout != nullptr) { 125 *request.mutable_shape_with_layout() = *shape_with_layout;
|
client.h | 96 // If shape_with_layout is not nullptr, it points to a shape whose layout will 99 const GlobalData& data, const Shape* shape_with_layout = nullptr); 125 const Shape* shape_with_layout, int64 replica_id = 0,
|
executable_build_options.h | 42 ExecutableBuildOptions& set_result_layout(const Shape& shape_with_layout);
|
computation_builder.h | 429 // shape_with_layout communicates the laid out shape that we want to outfeed 430 // -- if !ShapeUtil::Compatible(GetShape(operand), shape_with_layout) an error 433 const Shape& shape_with_layout, const string& outfeed_config); [all...] |
/external/tensorflow/tensorflow/compiler/xla/tests/ |
client_library_test_base.cc | 148 const Shape* shape_with_layout) { 150 shape_with_layout)); 156 const Shape* shape_with_layout) { 158 error, shape_with_layout)); 255 const Shape* shape_with_layout) { 282 if (shape_with_layout != nullptr) { 283 layout_shape = *shape_with_layout; 290 shape_with_layout = &layout_shape; 302 computation, *expected_ptr, arguments, expect_equal, shape_with_layout); 305 shape_with_layout)); [all...] |
client_library_test_base.h | 180 // literal. shape_with_layout indicates the result layout to request when 185 const Shape* shape_with_layout = nullptr); 189 const Shape* shape_with_layout = nullptr); 195 const Shape* shape_with_layout = nullptr); 199 const Shape* shape_with_layout = nullptr);
|
literal_test_util.cc | 772 Shape shape_with_layout = new_literal->shape(); local 773 *shape_with_layout.mutable_layout() = LayoutUtil::MakeLayout(minor_to_major); 780 IndexUtil::LinearIndexToMultidimensionalIndex(shape_with_layout, i);
|
local_client_execute_test.cc | 273 Shape shape_with_layout = ShapeUtil::MakeTupleShape( local 278 options.set_result_layout(shape_with_layout); [all...] |
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
gpu_layout_assignment.cc | 213 Shape shape_with_layout = buf.shape(); local 214 *shape_with_layout.mutable_layout() = buffer_constraint.layout(); 223 shape_with_layout, instruction, /*operand_no=*/0)); 231 shape_with_layout, instruction, /*operand_no=*/0)); 239 shape_with_layout, instruction, /*operand_no=*/0)); 241 shape_with_layout, instruction, /*operand_no=*/4));
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
layout_assignment.cc | 64 const Shape& shape_with_layout, HloInstruction* instruction) { 65 TF_RET_CHECK(LayoutUtil::HasLayout(shape_with_layout)); 66 DCHECK(ShapeUtil::Compatible(shape_with_layout, instruction->shape())) 67 << ShapeUtil::HumanString(shape_with_layout) << " " 85 ShapeUtil::GetSubshape(shape_with_layout, {i}), gte)); 93 shape_with_layout, tuple_copy->mutable_shape())); 101 shape_with_layout, copy->mutable_shape())); 267 Status LayoutConstraints::SetOperandLayout(const Shape& shape_with_layout, 273 << ShapeUtil::HumanStringWithLayout(shape_with_layout); 279 shape_with_layout)) { 465 Shape const* shape_with_layout = nullptr; local [all...] |
hlo_cost_analysis_test.cc | 397 Shape shape_with_layout = ShapeUtil::MakeShape(F32, {2, 3, 4, 5}); local 399 Shape shape_without_layout = shape_with_layout; 411 shape_with_layout, HloOpcode::kAdd, c1, broadcast));
|
layout_assignment.h | 168 Status SetOperandLayout(const Shape& shape_with_layout, 171 Status SetResultLayout(const Shape& shape_with_layout, bool dfs = true); 183 Status SetInstructionLayout(const Shape& shape_with_layout,
|
service.cc | 219 const Shape& shape_with_layout, const Shape& result_shape) const { 220 if (!ShapeUtil::Compatible(shape_with_layout, result_shape)) { 224 ShapeUtil::HumanStringWithLayout(shape_with_layout).c_str(), 227 if (!LayoutUtil::HasLayout(shape_with_layout)) { 230 ShapeUtil::HumanStringWithLayout(shape_with_layout).c_str()); 232 return ShapeUtil::ValidateShape(shape_with_layout); [all...] |
service.h | 339 // Convenience function which checks whether the given shape_with_layout 343 const Shape& shape_with_layout, const Shape& result_shape) const;
|
/external/tensorflow/tensorflow/compiler/xla/python/ |
local_computation_builder.cc | 110 const tensorflow::gtl::optional<Shape>& shape_with_layout) { 113 if (shape_with_layout) { 115 argument.Relayout(shape_with_layout.value()); 165 const tensorflow::gtl::optional<Shape>& shape_with_layout = 169 if (shape_with_layout) { 171 argument.Relayout(shape_with_layout.value());
|
local_computation_builder.h | 64 const tensorflow::gtl::optional<Shape>& shape_with_layout);
|
/external/tensorflow/tensorflow/compiler/xla/ |
layout_util_test.cc | 261 Shape shape_with_layout = ShapeUtil::MakeShape(F32, {2, 3, 4}); local 262 LayoutUtil::SetToDefaultLayout(&shape_with_layout); 263 EXPECT_FALSE(LayoutUtil::IsPadded(shape_with_layout));
|
xla.proto | 251 Shape shape_with_layout = 2; 279 Shape shape_with_layout = 1;
|
literal_util.cc | 718 const Shape& shape_with_layout) const { 719 CHECK(ShapeUtil::Compatible(shape_with_layout, shape())) 720 << "Given shape_with_layout " << ShapeUtil::HumanString(shape_with_layout) 723 std::unique_ptr<Literal> result = CreateFromShape(shape_with_layout); [all...] |
literal_util.h | 295 std::unique_ptr<Literal> Relayout(const Shape& shape_with_layout) const; [all...] |