HomeSort by relevance Sort by last modified time
    Searched defs:gpu (Results 226 - 250 of 321) sorted by null

1 2 3 4 5 6 7 8 91011>>

  /external/tensorflow/tensorflow/compiler/xla/service/gpu/
cudnn_conv_runner.cc 16 #include "tensorflow/compiler/xla/service/gpu/cudnn_conv_runner.h"
19 #include "tensorflow/compiler/xla/service/gpu/backend_configs.pb.h"
20 #include "tensorflow/compiler/xla/service/gpu/ir_emission_utils.h"
21 #include "tensorflow/compiler/xla/service/gpu/stream_executor_util.h"
27 namespace gpu { namespace in namespace:xla
433 } // namespace gpu
cudnn_fused_conv_rewriter.cc 16 #include "tensorflow/compiler/xla/service/gpu/cudnn_fused_conv_rewriter.h"
19 #include "tensorflow/compiler/xla/service/gpu/backend_configs.pb.h"
20 #include "tensorflow/compiler/xla/service/gpu/ir_emission_utils.h"
26 namespace gpu { namespace in namespace:xla
279 } // namespace gpu
fusion_merger.cc 16 #include "tensorflow/compiler/xla/service/gpu/fusion_merger.h"
23 #include "tensorflow/compiler/xla/service/gpu/gpu_fusible.h"
24 #include "tensorflow/compiler/xla/service/gpu/instruction_fusion.h"
32 namespace gpu { namespace in namespace:xla
240 // overhead from a GPU kernel launch).
308 } // namespace gpu
gemm_thunk.cc 16 #include "tensorflow/compiler/xla/service/gpu/gemm_thunk.h"
21 #include "tensorflow/compiler/xla/service/gpu/ir_emission_utils.h"
28 namespace gpu { namespace in namespace:xla
107 // unless you've already checked that it works for this particular GPU + input
153 // fail if the program was built with < CUDA 8 or if we're using a gpu older
462 // Autotune will fail when CUDA 8 and GPU sm_50 or older are used.
489 } // namespace gpu
gpu_executable.cc 16 #include "tensorflow/compiler/xla/service/gpu/gpu_executable.h"
25 #include "tensorflow/compiler/xla/service/gpu/buffer_allocations.h"
26 #include "tensorflow/compiler/xla/service/gpu/hlo_execution_profiler.h"
41 namespace gpu { namespace in namespace:xla
48 // Implementation note: HLO profiling is always enabled for GPU executables,
106 ScopedAnnotation top_level_annotation(hlo_module_->name(), "XLA GPU module");
354 "Asynchronous execution on stream is not yet supported on GPU.");
362 } // namespace gpu
gpu_hlo_schedule_test.cc 16 #include "tensorflow/compiler/xla/service/gpu/gpu_hlo_schedule.h"
22 #include "tensorflow/compiler/xla/service/gpu/stream_assignment.h"
32 namespace gpu { namespace in namespace:xla
404 } // namespace gpu
gpu_layout_assignment.cc 16 #include "tensorflow/compiler/xla/service/gpu/gpu_layout_assignment.h"
21 #include "tensorflow/compiler/xla/service/gpu/ir_emission_utils.h"
22 #include "tensorflow/compiler/xla/service/gpu/stream_executor_util.h"
33 namespace gpu { namespace in namespace:xla
368 } // namespace gpu
gpu_layout_assignment_test.cc 16 #include "tensorflow/compiler/xla/service/gpu/gpu_layout_assignment.h"
21 #include "tensorflow/compiler/xla/service/gpu/ir_emission_utils.h"
34 namespace gpu { namespace in namespace:xla
406 } // namespace gpu
ir_emission_utils.cc 16 #include "tensorflow/compiler/xla/service/gpu/ir_emission_utils.h"
36 namespace gpu { namespace in namespace:xla
290 } // namespace gpu
ir_emitter_unnested.h 20 #include "tensorflow/compiler/xla/service/gpu/ir_emitter.h"
21 #include "tensorflow/compiler/xla/service/gpu/sequential_thunk.h"
22 #include "tensorflow/compiler/xla/service/gpu/thunk.h"
27 namespace gpu { namespace in namespace:xla
379 } // namespace gpu
nccl_all_reduce_thunk.cc 16 #include "tensorflow/compiler/xla/service/gpu/nccl_all_reduce_thunk.h"
29 namespace gpu { namespace in namespace:xla
42 // GPU-replica-driving host threads (i.e. the threads that call
90 // but this is a reasonable stopgap measure to get multi-GPU-replica up and
355 } // namespace gpu
thunk.h 22 #include "tensorflow/compiler/xla/service/gpu/buffer_allocations.h"
23 #include "tensorflow/compiler/xla/service/gpu/hlo_execution_profiler.h"
29 namespace gpu { namespace in namespace:xla
112 } // namespace gpu
  /external/tensorflow/tensorflow/stream_executor/cuda/
cuda_fft.cc 35 namespace gpu { namespace in namespace:stream_executor
534 } // namespace gpu
539 cuda::kCudaPlatformId, gpu::kCuFftPlugin, "cuFFT",
541 gpu::GpuExecutor *cuda_executor =
542 dynamic_cast<gpu::GpuExecutor *>(parent);
549 return new gpu::CUDAFft(cuda_executor);
557 cuda::kCudaPlatformId, PluginKind::kFft, gpu::kCuFftPlugin);
cuda_platform.cc 28 namespace gpu { namespace in namespace:stream_executor
72 // at module initialization time, because non-GPU programs still include this
137 if (!gpu::GpuDriver::Init().ok()) {
194 } // namespace gpu
200 std::unique_ptr<gpu::CudaPlatform> platform(new gpu::CudaPlatform);
  /external/tensorflow/tensorflow/stream_executor/rocm/
rocm_diagnostics.cc 102 namespace gpu { namespace in namespace:stream_executor
239 } // namespace gpu
rocm_fft.cc 21 #include "tensorflow/stream_executor/gpu/gpu_activation.h"
22 #include "tensorflow/stream_executor/gpu/gpu_executor.h"
23 #include "tensorflow/stream_executor/gpu/gpu_helpers.h"
24 #include "tensorflow/stream_executor/gpu/gpu_stream.h"
36 namespace gpu { namespace in namespace:stream_executor
52 gpu::ScopedActivateExecutorContext sac{parent}; \
81 gpu::ScopedActivateExecutorContext sac{parent}; \
583 } // namespace gpu
587 rocm::kROCmPlatformId, PluginKind::kFft, gpu::kRocFftPlugin);
592 rocm::kROCmPlatformId, gpu::kRocFftPlugin, "rocFFT"
    [all...]
  /cts/tests/tests/graphics/jni/
VulkanPreTransformTestHelpers.h 39 VkPhysicalDevice gpu() const { return mGpu; } function in class:DeviceInfo
VulkanTestHelpers.h 37 VkPhysicalDevice gpu() { return mGpu; } function in class:VkInit
  /external/libchrome/base/threading/
thread_restrictions.h 71 namespace gpu { namespace
495 friend class gpu::GpuChannelHost; // http://crbug.com/125264
  /external/mesa3d/src/gallium/drivers/radeonsi/
si_get.c 751 const char *gpu; local
759 gpu = ac_get_llvm_processor_name(sscreen->info.family);
761 sprintf(ret, "%s-%s", gpu, triple);
764 return (strlen(triple) + strlen(gpu) + 2) * sizeof(char);
  /external/skia/src/gpu/
GrDrawingManager.cpp 224 GrGpu* gpu = direct->priv().getGpu(); local
225 if (!gpu) {
247 // expect each pool will use a CPU buffer as a staging buffer before uploading to a GPU
253 GrOpFlushState flushState(gpu, resourceProvider, &fTokenTracker, fCpuBufferCache);
348 GrSemaphoresSubmitted result = gpu->finishFlush(proxy, access, flags, numSemaphores,
421 // Upload all data to the GPU
427 // put a cap on the number of oplists we will execute before flushing to the GPU to relieve some
440 flushState->gpu()->finishFlush(nullptr, SkSurface::BackendSurfaceAccess::kNoAccess,
458 flushState->gpu()->finishFlush(nullptr, SkSurface::BackendSurfaceAccess::kNoAccess,
493 GrGpu* gpu = direct->priv().getGpu() local
    [all...]
  /external/skia/src/gpu/gl/builders/
GrGLProgramBuilder.cpp 29 #define GL_CALL(X) GR_GL_CALL(this->gpu()->glInterface(), X)
30 #define GL_CALL_RET(R, X) GR_GL_CALL_RET(this->gpu()->glInterface(), R, X)
37 GrGLGpu* gpu) {
45 GrGLProgramBuilder builder(gpu, renderTarget, origin,
48 auto persistentCache = gpu->getContext()->priv().getPersistentCache();
64 GrGLProgramBuilder::GrGLProgramBuilder(GrGLGpu* gpu,
72 , fGpu(gpu)
91 GrGLGpu* gpu = this->gpu(); local
92 GrGLuint shaderId = GrGLCompileAndAttachShader(gpu->glContext()
    [all...]
  /external/skia/src/utils/
SkCanvasStateUtils.cpp 74 } gpu; member in union:SkCanvasLayerState::__anon38290
  /external/skia/tests/
DrawOpAtlasTest.cpp 185 auto gpu = context->priv().getGpu(); local
214 GrOpFlushState flushState(gpu, resourceProvider, uploadTarget.writeableTokenTracker());
GLProgramsTest.cpp 8 // This is a GPU-backend specific test. It relies on static intializers to work
357 GrGLGpu* gpu = static_cast<GrGLGpu*>(context->priv().getGpu()); local
359 if (kGLES_GrGLStandard == gpu->glStandard()) {
364 if (kARM_GrGLVendor != gpu->ctxInfo().vendor()) {

Completed in 3045 milliseconds

1 2 3 4 5 6 7 8 91011>>