/external/tensorflow/tensorflow/compiler/xla/service/interpreter/ |
platform_id.cc | 18 namespace gputools { namespace in namespace:perftools 24 } // namespace gputools
|
platform_id.h | 22 namespace gputools { namespace in namespace:perftools 28 } // namespace gputools
|
/external/tensorflow/tensorflow/stream_executor/cuda/ |
cuda_platform_id.cc | 19 namespace gputools { namespace in namespace:perftools 25 } // namespace gputools
|
cuda_platform_id.h | 22 namespace gputools { namespace in namespace:perftools 33 } // namespace gputools
|
/external/tensorflow/tensorflow/stream_executor/host/ |
host_platform_id.cc | 19 namespace gputools { namespace in namespace:perftools 25 } // namespace gputools
|
host_platform_id.h | 22 namespace gputools { namespace in namespace:perftools 33 } // namespace gputools
|
/external/tensorflow/tensorflow/core/common_runtime/gpu/ |
gpu_init.h | 22 namespace gputools { namespace in namespace:perftools 24 } // namespace gputools 37 perftools::gputools::Platform* GPUMachineManager();
|
/external/tensorflow/tensorflow/core/platform/default/ |
from_stream_executor_status.h | 26 namespace gpu = ::perftools::gputools; 31 const perftools::gputools::port::Status& s) {
|
/external/tensorflow/tensorflow/stream_executor/lib/ |
demangle.h | 22 namespace gputools { namespace in namespace:perftools 28 } // namespace gputools
|
error.h | 16 // IWYU pragma: private, include "perftools/gputools/executor/stream_executor.h" 24 namespace gputools { namespace in namespace:perftools 30 } // namespace gputools
|
inlined_vector.h | 22 namespace gputools { namespace in namespace:perftools 28 } // namespace gputools
|
notification.h | 22 namespace gputools { namespace in namespace:perftools 28 } // namespace gputools
|
stacktrace.h | 23 namespace gputools { namespace in namespace:perftools 29 } // namespace gputools
|
stringpiece.h | 23 namespace gputools { namespace in namespace:perftools 29 } // namespace gputools
|
thread_options.h | 22 namespace gputools { namespace in namespace:perftools 28 } // namespace gputools
|
threadpool.h | 25 namespace gputools { namespace in namespace:perftools 31 } // namespace gputools
|
numbers.h | 22 namespace gputools { namespace in namespace:perftools 31 } // namespace gputools
|
process_state.h | 22 namespace gputools { namespace in namespace:perftools 29 } // namespace gputools
|
/external/tensorflow/tensorflow/stream_executor/ |
shared_memory_config.h | 23 namespace gputools { namespace in namespace:perftools 33 } // namespace gputools
|
kernel_cache_config.h | 22 namespace gputools { namespace in namespace:perftools 41 } // namespace gputools
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
platform_util.h | 37 static StatusOr<std::vector<perftools::gputools::Platform*>> 45 static StatusOr<perftools::gputools::Platform*> GetDefaultPlatform(); 50 static StatusOr<perftools::gputools::Platform*> GetSolePlatform(); 54 static StatusOr<perftools::gputools::Platform*> GetPlatform( 59 static StatusOr<perftools::gputools::Platform*> GetPlatformExceptFor( 67 static StatusOr<std::vector<perftools::gputools::StreamExecutor*>> 68 GetStreamExecutors(perftools::gputools::Platform* platform);
|
transfer_manager.h | 45 virtual perftools::gputools::Platform::Id PlatformId() const = 0; 61 perftools::gputools::StreamExecutor* executor, 69 perftools::gputools::StreamExecutor* executor, const Literal& literal, 76 perftools::gputools::StreamExecutor* executor, const Literal& literal, 77 const perftools::gputools::DeviceMemoryBase& dest); 79 perftools::gputools::StreamExecutor* executor, const Shape& shape, 80 const perftools::gputools::DeviceMemoryBase& source); 85 perftools::gputools::StreamExecutor* executor, 91 perftools::gputools::StreamExecutor* executor, const Shape& literal_shape, 96 tensorflow::gtl::ArraySlice<perftools::gputools::StreamExecutor* [all...] |
generic_transfer_manager.h | 39 GenericTransferManager(perftools::gputools::Platform::Id platform_id, 43 perftools::gputools::Platform::Id PlatformId() const override; 46 perftools::gputools::StreamExecutor* executor, 49 Status TransferLiteralToDevice(perftools::gputools::StreamExecutor* executor, 53 Status TransferLiteralToInfeed(perftools::gputools::StreamExecutor* executor, 56 perftools::gputools::StreamExecutor* executor, const Shape& literal_shape, 60 tensorflow::gtl::ArraySlice<perftools::gputools::StreamExecutor*> 66 Status TransferBufferToInfeed(perftools::gputools::StreamExecutor* executor, 70 perftools::gputools::StreamExecutor* executor, 71 tensorflow::gtl::ArraySlice<perftools::gputools::DeviceMemoryBase [all...] |
device_memory_allocator.h | 36 explicit DeviceMemoryAllocator(const perftools::gputools::Platform* platform) 46 virtual StatusOr<perftools::gputools::DeviceMemoryBase> Allocate( 49 int device_ordinal, perftools::gputools::DeviceMemoryBase* mem) = 0; 52 const perftools::gputools::Platform* platform() const { return platform_; } 59 const perftools::gputools::Platform* platform_; 67 const perftools::gputools::Platform* platform, 68 tensorflow::gtl::ArraySlice<perftools::gputools::StreamExecutor*> 71 StatusOr<perftools::gputools::DeviceMemoryBase> Allocate( 74 int device_ordinal, perftools::gputools::DeviceMemoryBase* mem) override; 79 StatusOr<perftools::gputools::StreamExecutor*> GetStreamExecutor [all...] |
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
cudnn_convolution_runner.h | 75 const Shape& output_shape, perftools::gputools::DeviceMemoryBase input_buf, 76 perftools::gputools::DeviceMemoryBase filter_buf, 77 perftools::gputools::DeviceMemoryBase output_buf, 78 perftools::gputools::DeviceMemoryBase scratch_buf, const Window& window, 80 perftools::gputools::dnn::AlgorithmConfig algorithm, 81 perftools::gputools::Stream* stream, 82 perftools::gputools::dnn::ProfileResult* profile_result = nullptr); 86 const Shape& output_shape, perftools::gputools::DeviceMemoryBase input_buf, 87 perftools::gputools::DeviceMemoryBase filter_buf, 88 perftools::gputools::DeviceMemoryBase output_buf [all...] |