HomeSort by relevance Sort by last modified time
    Searched refs:gpu (Results 351 - 375 of 554) sorted by null

<<11121314151617181920>>

  /external/tensorflow/tensorflow/compiler/xla/service/gpu/
fft_thunk.h 20 #include "tensorflow/compiler/xla/service/gpu/buffer_allocations.h"
21 #include "tensorflow/compiler/xla/service/gpu/gpu_executable.h"
22 #include "tensorflow/compiler/xla/service/gpu/thunk.h"
31 namespace gpu { namespace in namespace:xla
95 } // namespace gpu
gpu_executable.h 25 #include "tensorflow/compiler/xla/service/gpu/buffer_allocations.h"
26 #include "tensorflow/compiler/xla/service/gpu/stream_assignment.h"
27 #include "tensorflow/compiler/xla/service/gpu/thunk.h"
28 #include "tensorflow/compiler/xla/service/gpu/thunk_schedule.h"
41 namespace gpu { namespace in namespace:xla
43 // GPU-targeting implementation of the XLA Executable interface.
51 // compilation up to the GPU driver.
72 // empty, in which case compilation is left up to the GPU driver.
87 // TODO(b/62952745) Implement equality test on GPU executable.
88 return Unimplemented("Equality test on GPU executable is not implemented.")
    [all...]
infeed_manager.h 17 // buffer that are used by the GPU runtime to transfer buffers into an
18 // executing GPU computation, e.g., to feed data into a while loop.
31 namespace gpu { namespace in namespace:xla
33 // TODO(b/30467474) Once GPU infeed implementation settles, consider
34 // folding back the cpu and gpu infeed implementations into a generic
40 // * Buffer space on GPU is allocated on every infeed enqueue request,
130 // Singleton creator-or-accessor: Returns the GPU infeed manager.
133 } // namespace gpu
ir_emission_utils.h 29 namespace gpu { namespace in namespace:xla
136 } // namespace gpu
partition_assignment.cc 16 #include "tensorflow/compiler/xla/service/gpu/partition_assignment.h"
35 namespace gpu { namespace in namespace:xla
55 // smallest block size that allows the GPU to reach full occupancy (assuming
73 LOG(WARNING) << "Attempting to calculate launch dimensions for GPU "
100 } // namespace gpu
thunk_schedule.h 24 #include "tensorflow/compiler/xla/service/gpu/stream_assignment.h"
25 #include "tensorflow/compiler/xla/service/gpu/thunk.h"
30 namespace gpu { namespace in namespace:xla
90 } // namespace gpu
while_thunk.cc 16 #include "tensorflow/compiler/xla/service/gpu/while_thunk.h"
23 namespace gpu { namespace in namespace:xla
74 } // namespace gpu
  /external/tensorflow/tensorflow/compiler/xla/service/llvm_ir/
ops.cc 17 #include "tensorflow/compiler/xla/service/gpu/parallel_loop_emitter.h"
18 #include "tensorflow/compiler/xla/service/gpu/partition_assignment.h"
42 const gpu::LaunchDimensions* launch_dimensions,
78 return gpu::ParallelLoopEmitter(loop_body_emitter, update_shape,
119 const gpu::LaunchDimensions* launch_dimensions,
173 const gpu::LaunchDimensions& launch_dimensions,
ops.h 21 #include "tensorflow/compiler/xla/service/gpu/partition_assignment.h"
87 const gpu::LaunchDimensions& launch_dimensions,
  /external/tensorflow/tensorflow/core/common_runtime/gpu/
gpu_util.h 30 namespace gpu = ::perftools::gputools;
34 // "tensor" is GPU-local. "dev" is the hosting GPU.
35 // "device_context" should be the context of the GPU "_Send" op
38 // bytes from GPU to CPU RAM. "is_dead" indicates that the
60 // corresponding GPU device at the time of call have completed.
66 // (up to a limit). "device" can be either a CPU or a GPU device.
104 // Deep-copying of GPU tensor on the same device.
gpu_util.cc 16 #include "tensorflow/core/common_runtime/gpu/gpu_util.h"
21 #include "tensorflow/core/common_runtime/gpu/gpu_event_mgr.h"
22 #include "tensorflow/core/common_runtime/gpu/process_state.h"
63 namespace gpu = ::perftools::gputools;
68 gpu::Stream** stream) {
82 return errors::Internal("No gpu stream is available.");
104 return errors::Internal("GPU copy from non-DMA ",
123 gpu::Stream* send_stream = nullptr;
135 done(errors::Internal("No send gpu copy-out-stream is available."));
141 // Tensor values need to be copied from GPU to CPU ram so tha
    [all...]
  /external/tensorflow/tensorflow/core/kernels/
adjust_contrast_op_benchmark_test.cc 52 // The GPU test ran with -c opt --config=cuda --copt=-mavx, CPU ran without
55 // GPU: Tesla K40m
60 BM_AdjustContrastDev(gpu, 32, 299, 299);
multinomial_op_test.cc 52 BM_MultinomialDev(gpu, B, C, S);
  /frameworks/native/vulkan/tools/
vkinfo.cpp 106 void EnumerateDeviceExtensions(VkPhysicalDevice gpu,
112 vkEnumerateDeviceExtensionProperties(gpu, layer_name, &count, nullptr);
117 result = vkEnumerateDeviceExtensionProperties(gpu, layer_name, &count,
124 void GatherGpuInfo(VkPhysicalDevice gpu,
130 vkGetPhysicalDeviceProperties(gpu, &info.properties);
131 vkGetPhysicalDeviceMemoryProperties(gpu, &info.memory);
132 vkGetPhysicalDeviceFeatures(gpu, &info.features);
134 vkGetPhysicalDeviceQueueFamilyProperties(gpu, &count, nullptr);
136 vkGetPhysicalDeviceQueueFamilyProperties(gpu, &count,
139 result = vkEnumerateDeviceLayerProperties(gpu, &count, nullptr)
    [all...]
  /external/skia/src/gpu/
GrBufferAllocPool.cpp 34 TRACE_EVENT_INSTANT1("skia.gpu", \
42 GrBufferAllocPool::GrBufferAllocPool(GrGpu* gpu, GrBufferType bufferType, size_t blockSize)
45 fGpu = SkRef(gpu);
53 fBufferMapThreshold = gpu->caps()->bufferMapThreshold();
381 GrVertexBufferAllocPool::GrVertexBufferAllocPool(GrGpu* gpu)
382 : GrBufferAllocPool(gpu, kVertex_GrBufferType, MIN_VERTEX_BUFFER_SIZE) {
436 GrIndexBufferAllocPool::GrIndexBufferAllocPool(GrGpu* gpu)
437 : GrBufferAllocPool(gpu, kIndex_GrBufferType, MIN_INDEX_BUFFER_SIZE) {
GrDrawingManager.cpp 158 GrGpu* gpu = fContext->contextPriv().getGpu(); local
160 GrOpFlushState flushState(gpu, fContext->contextPriv().resourceProvider(),
232 GrSemaphoresSubmitted result = gpu->finishFlush(numSemaphores, backendSemaphores);
279 // Upload all data to the GPU
342 GrGpu* gpu = fContext->contextPriv().getGpu(); local
345 if (gpu && surface->asRenderTarget()) {
346 gpu->resolveRenderTarget(surface->asRenderTarget());
  /external/skqp/src/gpu/
GrBufferAllocPool.cpp 34 TRACE_EVENT_INSTANT1("skia.gpu", \
42 GrBufferAllocPool::GrBufferAllocPool(GrGpu* gpu, GrBufferType bufferType, size_t blockSize)
45 fGpu = SkRef(gpu);
53 fBufferMapThreshold = gpu->caps()->bufferMapThreshold();
381 GrVertexBufferAllocPool::GrVertexBufferAllocPool(GrGpu* gpu)
382 : GrBufferAllocPool(gpu, kVertex_GrBufferType, MIN_VERTEX_BUFFER_SIZE) {
436 GrIndexBufferAllocPool::GrIndexBufferAllocPool(GrGpu* gpu)
437 : GrBufferAllocPool(gpu, kIndex_GrBufferType, MIN_INDEX_BUFFER_SIZE) {
GrDrawingManager.cpp 126 GrGpu* gpu = fContext->contextPriv().getGpu(); local
128 GrOpFlushState flushState(gpu, fContext->contextPriv().resourceProvider(),
207 GrSemaphoresSubmitted result = gpu->finishFlush(numSemaphores, backendSemaphores);
249 // Upload all data to the GPU
312 GrGpu* gpu = fContext->contextPriv().getGpu(); local
315 if (gpu && surface->asRenderTarget()) {
316 gpu->resolveRenderTarget(surface->asRenderTarget(), proxy->origin());
  /external/skia/src/gpu/vk/
GrVkResourceProvider.cpp 25 GrVkResourceProvider::GrVkResourceProvider(GrVkGpu* gpu)
26 : fGpu(gpu)
28 fPipelineStateCache = new PipelineStateCache(gpu);
445 const GrVkGpu* gpu,
449 fRenderPasses[0]->initSimple(gpu, target);
461 const GrVkGpu* gpu,
472 renderPass->init(gpu, *this->getCompatibleRenderPass(), colorOps, stencilOps);
477 void GrVkResourceProvider::CompatibleRenderPassSet::releaseResources(const GrVkGpu* gpu) {
480 fRenderPasses[i]->unref(gpu);
  /external/skqp/src/gpu/vk/
GrVkResourceProvider.cpp 25 GrVkResourceProvider::GrVkResourceProvider(GrVkGpu* gpu)
26 : fGpu(gpu)
28 fPipelineStateCache = new PipelineStateCache(gpu);
445 const GrVkGpu* gpu,
449 fRenderPasses[0]->initSimple(gpu, target);
461 const GrVkGpu* gpu,
472 renderPass->init(gpu, *this->getCompatibleRenderPass(), colorOps, stencilOps);
477 void GrVkResourceProvider::CompatibleRenderPassSet::releaseResources(const GrVkGpu* gpu) {
480 fRenderPasses[i]->unref(gpu);
  /external/skia/tests/
VkHeapTests.cpp 8 // This is a GPU-backend specific test. It relies on static intializers to work
23 GrVkGpu* gpu = static_cast<GrVkGpu*>(context->contextPriv().getGpu()); local
26 GrVkSubHeap heap(gpu, 0, 0, 64 * 1024, 32);
120 GrVkGpu* gpu = static_cast<GrVkGpu*>(context->contextPriv().getGpu()); local
123 GrVkHeap heap(gpu, GrVkHeap::kSubAlloc_Strategy, 64 * 1024);
179 GrVkGpu* gpu = static_cast<GrVkGpu*>(context->contextPriv().getGpu()); local
182 GrVkHeap heap(gpu, GrVkHeap::kSingleAlloc_Strategy, 64 * 1024);
  /external/skqp/tests/
VkHeapTests.cpp 8 // This is a GPU-backend specific test. It relies on static intializers to work
23 GrVkGpu* gpu = static_cast<GrVkGpu*>(context->contextPriv().getGpu()); local
26 GrVkSubHeap heap(gpu, 0, 0, 64 * 1024, 32);
120 GrVkGpu* gpu = static_cast<GrVkGpu*>(context->contextPriv().getGpu()); local
123 GrVkHeap heap(gpu, GrVkHeap::kSubAlloc_Strategy, 64 * 1024);
179 GrVkGpu* gpu = static_cast<GrVkGpu*>(context->contextPriv().getGpu()); local
182 GrVkHeap heap(gpu, GrVkHeap::kSingleAlloc_Strategy, 64 * 1024);
  /hardware/qcom/display/msm8084/libgralloc/
gpu.cpp 25 #include "gpu.h"
359 gpu_context_t* gpu = reinterpret_cast<gpu_context_t*>(dev); local
360 return gpu->alloc_impl(w, h, format, usage, pHandle, pStride, 0);
370 gpu_context_t* gpu = reinterpret_cast<gpu_context_t*>(dev); local
371 return gpu->alloc_impl(w, h, format, usage, pHandle, pStride, bufferSize);
382 gpu_context_t* gpu = reinterpret_cast<gpu_context_t*>(dev); local
383 return gpu->free_impl(hnd);
  /hardware/qcom/display/msm8226/libgralloc/
gpu.cpp 25 #include "gpu.h"
362 gpu_context_t* gpu = reinterpret_cast<gpu_context_t*>(dev); local
363 return gpu->alloc_impl(w, h, format, usage, pHandle, pStride, 0);
373 gpu_context_t* gpu = reinterpret_cast<gpu_context_t*>(dev); local
374 return gpu->alloc_impl(w, h, format, usage, pHandle, pStride, bufferSize);
385 gpu_context_t* gpu = reinterpret_cast<gpu_context_t*>(dev); local
386 return gpu->free_impl(hnd);
  /hardware/qcom/display/msm8909/libgralloc/
gpu.cpp 28 #include "gpu.h"
225 // As GPU needs ION FD, the private handle is created
383 gpu_context_t* gpu = reinterpret_cast<gpu_context_t*>(dev); local
384 return gpu->alloc_impl(w, h, format, usage, pHandle, pStride, 0);
394 gpu_context_t* gpu = reinterpret_cast<gpu_context_t*>(dev); local
395 return gpu->alloc_impl(w, h, format, usage, pHandle, pStride, bufferSize);
406 gpu_context_t* gpu = reinterpret_cast<gpu_context_t*>(dev); local
407 return gpu->free_impl(hnd);

Completed in 2046 milliseconds

<<11121314151617181920>>