/external/tensorflow/tensorflow/compiler/xla/service/ |
device_memory_allocator.cc | 29 absl::Span<se::StreamExecutor* const> stream_executors) 31 stream_executors_(stream_executors.begin(), stream_executors.end()) {}
|
backend.cc | 109 auto stream_executors, 116 new Backend(platform, compiler, stream_executors, transfer_manager, 144 absl::Span<se::StreamExecutor* const> stream_executors, 153 for (se::StreamExecutor* exec : stream_executors) { 160 platform, stream_executors);
|
platform_util.cc | 228 std::vector<se::StreamExecutor*> stream_executors(device_count, nullptr); 245 thread_pool.Schedule([platform, i, &stream_executors]() { 253 stream_executors[i] = executor; 266 if (absl::c_all_of(stream_executors, 271 return stream_executors;
|
backend.h | 106 const std::vector<se::StreamExecutor*>& stream_executors() const { function in class:xla::Backend 160 absl::Span<se::StreamExecutor* const> stream_executors,
|
device_memory_allocator.h | 83 absl::Span<se::StreamExecutor* const> stream_executors);
|
service.cc | 171 auto stream_executors = execute_backend_->stream_executors(); local 173 se::StreamExecutor* executor = stream_executors.at(i); 232 absl::Span<se::StreamExecutor* const> stream_executors) const { 233 CHECK_EQ(options_.number_of_replicas(), stream_executors.size()); 247 int replica_device_ordinal = stream_executors[replica]->device_ordinal(); [all...] |
service.h | 219 absl::Span<se::StreamExecutor* const> stream_executors) const;
|
/external/tensorflow/tensorflow/compiler/jit/ |
xla_compilation_cache.cc | 50 for (auto* executor : client_->backend().stream_executors()) {
|
xla_device.cc | 106 backend->stream_executors()[device_ordinal]);
|