/external/tensorflow/tensorflow/compiler/xla/client/ |
computation.h | 30 // Wraps a ComputationHandle protobuf with a lifetime. Computation is 33 class Computation { 35 // Creates a null Computation. 36 Computation(); 38 // parent: stub for the service on which we will deallocate the computation 40 // handle: the computation handle protobuf from the service. 41 Computation(ServiceInterface* parent, const ComputationHandle& handle); 43 Computation(Computation&& computation); [all...] |
computation.cc | 16 #include "tensorflow/compiler/xla/client/computation.h" 24 Computation::Computation() : parent_(nullptr) {} 26 Computation::Computation(ServiceInterface* parent, 30 Computation::Computation(Computation&& computation) 31 : handle_(std::move(computation.handle_)), parent_(computation.parent_) [all...] |
client.h | 22 #include "tensorflow/compiler/xla/client/computation.h" 43 // Executes the computation with the given arguments and returns the global 46 // service to affect how it compiles our computation. (The pointer does not 48 // * If execution_options.device_handles is not empty, the computation is 50 // computation based on the attached sharding attributes. Otherwise, a 55 const Computation& computation, 60 // A struct to represent a computation instance to be executed. 61 // * If execution_options.device_handles is not empty, the computation is 63 // computation based on the attached sharding attributes. Otherwise, 66 const Computation& computation; member in struct:xla::Client::ComputationInstance [all...] |
compile_only_client.h | 20 #include "tensorflow/compiler/xla/client/computation.h" 40 // A description of a computation to compile using CompileAheadOfTime. 42 const Computation* computation; member in struct:xla::CompileOnlyClient::AotComputationInstance
|
/external/tensorflow/tensorflow/compiler/xla/client/lib/ |
arithmetic.h | 21 #include "tensorflow/compiler/xla/client/computation.h" 27 // Creates a scalar add computation and returns it. 28 Computation CreateScalarAddComputation(PrimitiveType type, 31 // Creates a scalar multiply computation and returns it. 32 Computation CreateScalarMultiplyComputation(PrimitiveType type, 35 // Creates a scalar ge computation and returns it. 36 Computation CreateScalarGeComputation(PrimitiveType type, 39 // Creates a scalar max computation and returns it. 40 Computation CreateScalarMaxComputation(PrimitiveType type, 43 // Creates a scalar min computation and returns it [all...] |
arithmetic.cc | 33 Computation CreateScalarComputation(const string& name, PrimitiveType type, 52 Computation CreateScalarAddComputation(PrimitiveType type, 60 Computation CreateScalarMultiplyComputation(PrimitiveType type, 68 Computation CreateScalarGeComputation(PrimitiveType type, 76 Computation CreateScalarMaxComputation(PrimitiveType type, 84 Computation CreateScalarMinComputation(PrimitiveType type, 92 Computation CreateScalarAndComputation(ComputationBuilder* builder) { 99 Computation CreateScalarOrComputation(ComputationBuilder* builder) { 109 Computation logical_or = CreateScalarOrComputation(builder);
|
testing.h | 23 #include "tensorflow/compiler/xla/client/computation.h" 30 // is created by performing a computation on the device rather than transferring 37 // computation. 39 const Computation& computation, Client* client);
|
testing.cc | 18 #include "tensorflow/compiler/xla/client/computation.h" 71 Computation computation = b.Build().ConsumeValueOrDie(); local 75 return client->Execute(computation, /*arguments=*/{}, &execution_options) 87 // an on-device computation. 100 const Computation& computation, Client* client) { 102 client->GetComputationShape(computation).ConsumeValueOrDie(); 104 // For every (unbound) parameter that the computation wants, we manufacture 105 // some arbitrary data so that we can invoke the computation [all...] |
/external/tensorflow/tensorflow/compiler/tf2xla/ |
tf2xla.h | 21 #include "tensorflow/compiler/xla/client/computation.h" 26 // Converts a tensorflow::GraphDef into an xla::Computation. The given `config` 28 // feed is a positional input argument for the generated computation, while each 31 // The computation is built in the context of the given `client`, which may 32 // subsequently be used to compile or execute the computation. 35 xla::Computation* computation);
|
xla_context.h | 25 #include "tensorflow/compiler/xla/client/computation.h" 98 // separate specialization of the computation for each DataType. 99 const xla::Computation* GetOrCreateMax(const DataType type); 103 // separate specialization of the computation for each DataType. 104 const xla::Computation* GetOrCreateMin(const DataType type); 108 // separate specialization of the computation for each DataType. 109 const xla::Computation* GetOrCreateAdd(const DataType type); 113 // separate specialization of the computation for each DataType. 114 const xla::Computation* GetOrCreateMul(const DataType type); 130 // run-time computation outputs [all...] |
xla_op_kernel.h | 207 // separate specialization of the computation for each DataType. 208 const xla::Computation* GetOrCreateMax(const DataType type); 212 // separate specialization of the computation for each DataType. 213 const xla::Computation* GetOrCreateMin(const DataType type); 217 // separate specialization of the computation for each DataType. 218 const xla::Computation* GetOrCreateAdd(const DataType type); 222 // separate specialization of the computation for each DataType. 223 const xla::Computation* GetOrCreateMul(const DataType type);
|
xla_context.cc | 82 VLOG(1) << "Added retval index " << retval_index << " to XLA computation"; 93 << " with non-data-dependent tensor to XLA computation"; 126 const xla::Computation* XlaContext::GetOrCreateMax(const DataType type) { 140 const xla::Computation* XlaContext::GetOrCreateMin(const DataType type) { 154 const xla::Computation* XlaContext::GetOrCreateAdd(const DataType type) { 168 const xla::Computation* XlaContext::GetOrCreateMul(const DataType type) { 182 const xla::Computation* XlaContext::LookupOrCreate( 184 const std::function<xla::Computation()>& create) {
|
/external/tensorflow/tensorflow/compiler/xla/tests/ |
call_test.cc | 19 #include "tensorflow/compiler/xla/client/computation.h" 35 Computation CreateR0F32IdentityComputation() { 43 Computation CreateR1S0F32AdditionComputation() { 53 Computation CreateR1S2F32AdditionComputation() { 63 Computation CreateR0F32TupleComputation() { 78 Computation callee = CreateR0F32IdentityComputation(); 87 Computation callee = CreateR1S0F32AdditionComputation(); 97 Computation callee = CreateR1S2F32AdditionComputation(); 111 TF_ASSERT_OK_AND_ASSIGN(Computation inner, builder.Build()); 120 TF_ASSERT_OK_AND_ASSIGN(Computation outer, builder2.Build()) [all...] |
bad_rng_shape_validation_test.cc | 21 #include "tensorflow/compiler/xla/client/computation.h" 43 StatusOr<Computation> computation = builder.Build(); local 44 EXPECT_FALSE(computation.ok()); 45 LOG(INFO) << "status received: " << computation.status(); 46 EXPECT_THAT(computation.status().error_message(), 60 StatusOr<Computation> computation = builder.Build(); local 61 ASSERT_TRUE(computation.ok()); 62 LOG(INFO) << computation.status() [all...] |
replay_test.cc | 18 #include "tensorflow/compiler/xla/client/computation.h" 40 // Make 2+2 computation. 44 Computation computation = builder.Build().ConsumeValueOrDie(); local 48 computation.Snapshot().ConsumeValueOrDie(); 51 Computation replayed = client_->LoadSnapshot(*module).ConsumeValueOrDie(); 55 client_->GetComputationShape(computation).ConsumeValueOrDie(); 71 // Make computation. 76 Computation computation = builder.Build().ConsumeValueOrDie() local 122 Computation computation = mapper_builder.Build().ConsumeValueOrDie(); local [all...] |
local_client_test_base.h | 24 #include "tensorflow/compiler/xla/client/computation.h" 96 // Execute the given computation on the local client. With and without 99 const Computation& computation, 102 const Computation& computation, 108 const Computation& computation, 111 const Computation& computation, [all...] |
compilation_cache_test.cc | 20 #include "tensorflow/compiler/xla/client/computation.h" 42 const Computation& computation, 48 ->ExecuteAndTransfer(computation, arguments, 58 const Computation& computation, 64 ->Execute(computation, arguments, 80 Computation computation = builder.Build().ConsumeValueOrDie(); local 82 ExecuteComputationR0F32(computation, {}, -42.0, /*expect_cache_hit=*/false) 100 Computation computation = builder.Build().ConsumeValueOrDie(); local 153 Computation computation = builder.Build().ConsumeValueOrDie(); local 179 Computation computation = builder.Build().ConsumeValueOrDie(); local [all...] |
while_test.cc | 21 #include "tensorflow/compiler/xla/client/computation.h" 56 // Create a computation for the condition: repeat for 5 iterations. 57 Computation condition; 65 // Create a computation for the body: add 1 to the result variable. 66 Computation body; 93 // Create a computation for the condition: repeat for 5 iterations. 94 Computation condition; 102 // Create a computation for the body: add 1 to the result variable. 103 Computation body; 125 // Create a computation for the condition: repeat for 5 iterations 1307 auto computation = builder.Build().ConsumeValueOrDie(); local [all...] |
conditional_test.cc | 26 Computation CreateR0ConstantComputation(float value) { 35 Computation CreateR0IdentityComputation() { 43 Computation CreateCeilComputation(const Shape& shape) { 52 Computation CreateR0CeilComputation() { 56 Computation CreateR1CeilComputation() { 60 Computation CreateFloorComputation(const Shape& shape) { 69 Computation CreateR0FloorComputation() { 73 Computation CreateR1FloorComputation() { 77 Computation CreateTupleCeilComputation(const string& computation_name, 91 Computation CreateR0TupleCeilComputation() [all...] |
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
cwise_ops.h | 32 // description of the operation; and Computation adds the 47 // applied to the inputs. The desired computation should be added to 58 virtual xla::ComputationDataHandle Computation(
|
l2loss_op.cc | 39 const xla::Computation& add = *ctx->GetOrCreateAdd(dtype);
|
quantize_and_dequantize_op.cc | 58 const xla::Computation* fmax = ctx->GetOrCreateMax(data_type); 59 const xla::Computation* fmin = ctx->GetOrCreateMin(data_type);
|
/external/tensorflow/tensorflow/compiler/xla/tools/ |
show_signature.cc | 34 #include "tensorflow/compiler/xla/client/computation.h" 55 Computation computation = client->LoadSnapshot(module).ConsumeValueOrDie(); local 57 client->GetComputationShape(computation).ConsumeValueOrDie();
|
/external/tensorflow/tensorflow/compiler/xla/service/cpu/ |
sample_harness.cc | 22 #include "tensorflow/compiler/xla/client/computation.h" 50 // Build computation. 56 xla::StatusOr<xla::Computation> computation_status = builder.Build(); 57 xla::Computation computation = computation_status.ConsumeValueOrDie(); local 59 // Execute and transfer result of computation. 63 computation, 69 LOG(INFO) << tensorflow::strings::Printf("computation took %lldns",
|
/external/tensorflow/tensorflow/compiler/aot/ |
compile.cc | 44 // Compiles the XLA computation into executable code. 46 const xla::Computation& computation, 49 // Retrieves arg and result layouts from the computation. 52 client->GetComputationShape(computation); 65 instance.computation = &computation; 87 // Converts the graph into an XLA computation, and compiles the 88 // computation. 96 xla::Computation computation local [all...] |