/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
cudnn_conv_runner.cc | 16 #include "tensorflow/compiler/xla/service/gpu/cudnn_conv_runner.h" 19 #include "tensorflow/compiler/xla/service/gpu/backend_configs.pb.h" 20 #include "tensorflow/compiler/xla/service/gpu/ir_emission_utils.h" 21 #include "tensorflow/compiler/xla/service/gpu/stream_executor_util.h" 27 namespace gpu { namespace in namespace:xla 433 } // namespace gpu
|
cudnn_fused_conv_rewriter.cc | 16 #include "tensorflow/compiler/xla/service/gpu/cudnn_fused_conv_rewriter.h" 19 #include "tensorflow/compiler/xla/service/gpu/backend_configs.pb.h" 20 #include "tensorflow/compiler/xla/service/gpu/ir_emission_utils.h" 26 namespace gpu { namespace in namespace:xla 279 } // namespace gpu
|
fusion_merger.cc | 16 #include "tensorflow/compiler/xla/service/gpu/fusion_merger.h" 23 #include "tensorflow/compiler/xla/service/gpu/gpu_fusible.h" 24 #include "tensorflow/compiler/xla/service/gpu/instruction_fusion.h" 32 namespace gpu { namespace in namespace:xla 240 // overhead from a GPU kernel launch). 308 } // namespace gpu
|
gemm_thunk.cc | 16 #include "tensorflow/compiler/xla/service/gpu/gemm_thunk.h" 21 #include "tensorflow/compiler/xla/service/gpu/ir_emission_utils.h" 28 namespace gpu { namespace in namespace:xla 107 // unless you've already checked that it works for this particular GPU + input 153 // fail if the program was built with < CUDA 8 or if we're using a gpu older 462 // Autotune will fail when CUDA 8 and GPU sm_50 or older are used. 489 } // namespace gpu
|
gpu_executable.cc | 16 #include "tensorflow/compiler/xla/service/gpu/gpu_executable.h" 25 #include "tensorflow/compiler/xla/service/gpu/buffer_allocations.h" 26 #include "tensorflow/compiler/xla/service/gpu/hlo_execution_profiler.h" 41 namespace gpu { namespace in namespace:xla 48 // Implementation note: HLO profiling is always enabled for GPU executables, 106 ScopedAnnotation top_level_annotation(hlo_module_->name(), "XLA GPU module"); 354 "Asynchronous execution on stream is not yet supported on GPU."); 362 } // namespace gpu
|
gpu_hlo_schedule_test.cc | 16 #include "tensorflow/compiler/xla/service/gpu/gpu_hlo_schedule.h" 22 #include "tensorflow/compiler/xla/service/gpu/stream_assignment.h" 32 namespace gpu { namespace in namespace:xla 404 } // namespace gpu
|
gpu_layout_assignment.cc | 16 #include "tensorflow/compiler/xla/service/gpu/gpu_layout_assignment.h" 21 #include "tensorflow/compiler/xla/service/gpu/ir_emission_utils.h" 22 #include "tensorflow/compiler/xla/service/gpu/stream_executor_util.h" 33 namespace gpu { namespace in namespace:xla 368 } // namespace gpu
|
gpu_layout_assignment_test.cc | 16 #include "tensorflow/compiler/xla/service/gpu/gpu_layout_assignment.h" 21 #include "tensorflow/compiler/xla/service/gpu/ir_emission_utils.h" 34 namespace gpu { namespace in namespace:xla 406 } // namespace gpu
|
ir_emission_utils.cc | 16 #include "tensorflow/compiler/xla/service/gpu/ir_emission_utils.h" 36 namespace gpu { namespace in namespace:xla 290 } // namespace gpu
|
ir_emitter_unnested.h | 20 #include "tensorflow/compiler/xla/service/gpu/ir_emitter.h" 21 #include "tensorflow/compiler/xla/service/gpu/sequential_thunk.h" 22 #include "tensorflow/compiler/xla/service/gpu/thunk.h" 27 namespace gpu { namespace in namespace:xla 379 } // namespace gpu
|
nccl_all_reduce_thunk.cc | 16 #include "tensorflow/compiler/xla/service/gpu/nccl_all_reduce_thunk.h" 29 namespace gpu { namespace in namespace:xla 42 // GPU-replica-driving host threads (i.e. the threads that call 90 // but this is a reasonable stopgap measure to get multi-GPU-replica up and 355 } // namespace gpu
|
thunk.h | 22 #include "tensorflow/compiler/xla/service/gpu/buffer_allocations.h" 23 #include "tensorflow/compiler/xla/service/gpu/hlo_execution_profiler.h" 29 namespace gpu { namespace in namespace:xla 112 } // namespace gpu
|
/external/tensorflow/tensorflow/stream_executor/cuda/ |
cuda_fft.cc | 35 namespace gpu { namespace in namespace:stream_executor 534 } // namespace gpu 539 cuda::kCudaPlatformId, gpu::kCuFftPlugin, "cuFFT", 541 gpu::GpuExecutor *cuda_executor = 542 dynamic_cast<gpu::GpuExecutor *>(parent); 549 return new gpu::CUDAFft(cuda_executor); 557 cuda::kCudaPlatformId, PluginKind::kFft, gpu::kCuFftPlugin);
|
cuda_platform.cc | 28 namespace gpu { namespace in namespace:stream_executor 72 // at module initialization time, because non-GPU programs still include this 137 if (!gpu::GpuDriver::Init().ok()) { 194 } // namespace gpu 200 std::unique_ptr<gpu::CudaPlatform> platform(new gpu::CudaPlatform);
|
/external/tensorflow/tensorflow/stream_executor/rocm/ |
rocm_diagnostics.cc | 102 namespace gpu { namespace in namespace:stream_executor 239 } // namespace gpu
|
rocm_fft.cc | 21 #include "tensorflow/stream_executor/gpu/gpu_activation.h" 22 #include "tensorflow/stream_executor/gpu/gpu_executor.h" 23 #include "tensorflow/stream_executor/gpu/gpu_helpers.h" 24 #include "tensorflow/stream_executor/gpu/gpu_stream.h" 36 namespace gpu { namespace in namespace:stream_executor 52 gpu::ScopedActivateExecutorContext sac{parent}; \ 81 gpu::ScopedActivateExecutorContext sac{parent}; \ 583 } // namespace gpu 587 rocm::kROCmPlatformId, PluginKind::kFft, gpu::kRocFftPlugin); 592 rocm::kROCmPlatformId, gpu::kRocFftPlugin, "rocFFT" [all...] |
/cts/tests/tests/graphics/jni/ |
VulkanPreTransformTestHelpers.h | 39 VkPhysicalDevice gpu() const { return mGpu; } function in class:DeviceInfo
|
VulkanTestHelpers.h | 37 VkPhysicalDevice gpu() { return mGpu; } function in class:VkInit
|
/external/libchrome/base/threading/ |
thread_restrictions.h | 71 namespace gpu { namespace 495 friend class gpu::GpuChannelHost; // http://crbug.com/125264
|
/external/mesa3d/src/gallium/drivers/radeonsi/ |
si_get.c | 751 const char *gpu; local 759 gpu = ac_get_llvm_processor_name(sscreen->info.family); 761 sprintf(ret, "%s-%s", gpu, triple); 764 return (strlen(triple) + strlen(gpu) + 2) * sizeof(char);
|
/external/skia/src/gpu/ |
GrDrawingManager.cpp | 224 GrGpu* gpu = direct->priv().getGpu(); local 225 if (!gpu) { 247 // expect each pool will use a CPU buffer as a staging buffer before uploading to a GPU 253 GrOpFlushState flushState(gpu, resourceProvider, &fTokenTracker, fCpuBufferCache); 348 GrSemaphoresSubmitted result = gpu->finishFlush(proxy, access, flags, numSemaphores, 421 // Upload all data to the GPU 427 // put a cap on the number of oplists we will execute before flushing to the GPU to relieve some 440 flushState->gpu()->finishFlush(nullptr, SkSurface::BackendSurfaceAccess::kNoAccess, 458 flushState->gpu()->finishFlush(nullptr, SkSurface::BackendSurfaceAccess::kNoAccess, 493 GrGpu* gpu = direct->priv().getGpu() local [all...] |
/external/skia/src/gpu/gl/builders/ |
GrGLProgramBuilder.cpp | 29 #define GL_CALL(X) GR_GL_CALL(this->gpu()->glInterface(), X) 30 #define GL_CALL_RET(R, X) GR_GL_CALL_RET(this->gpu()->glInterface(), R, X) 37 GrGLGpu* gpu) { 45 GrGLProgramBuilder builder(gpu, renderTarget, origin, 48 auto persistentCache = gpu->getContext()->priv().getPersistentCache(); 64 GrGLProgramBuilder::GrGLProgramBuilder(GrGLGpu* gpu, 72 , fGpu(gpu) 91 GrGLGpu* gpu = this->gpu(); local 92 GrGLuint shaderId = GrGLCompileAndAttachShader(gpu->glContext() [all...] |
/external/skia/src/utils/ |
SkCanvasStateUtils.cpp | 74 } gpu; member in union:SkCanvasLayerState::__anon38290
|
/external/skia/tests/ |
DrawOpAtlasTest.cpp | 185 auto gpu = context->priv().getGpu(); local 214 GrOpFlushState flushState(gpu, resourceProvider, uploadTarget.writeableTokenTracker());
|
GLProgramsTest.cpp | 8 // This is a GPU-backend specific test. It relies on static intializers to work 357 GrGLGpu* gpu = static_cast<GrGLGpu*>(context->priv().getGpu()); local 359 if (kGLES_GrGLStandard == gpu->glStandard()) { 364 if (kARM_GrGLVendor != gpu->ctxInfo().vendor()) {
|