/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
gpu_constants.cc | 16 #include "tensorflow/compiler/xla/service/gpu/gpu_constants.h" 19 namespace gpu { namespace in namespace:xla 24 } // namespace gpu
|
gpu_constants.h | 22 namespace gpu { namespace in namespace:xla 28 } // namespace gpu
|
/external/libdrm/etnaviv/ |
etnaviv_gpu.c | 53 struct etna_gpu *gpu; local 55 gpu = calloc(1, sizeof(*gpu)); 56 if (!gpu) { 61 gpu->dev = dev; 62 gpu->core = core; 65 gpu->specs.model = get_param(dev, core, ETNAVIV_PARAM_GPU_MODEL); 66 gpu->specs.revision = get_param(dev, core, ETNAVIV_PARAM_GPU_REVISION); 67 gpu->specs.features[0] = get_param(dev, core, ETNAVIV_PARAM_GPU_FEATURES_0); 68 gpu->specs.features[1] = get_param(dev, core, ETNAVIV_PARAM_GPU_FEATURES_1) [all...] |
/external/skia/src/gpu/ |
GrSemaphore.h | 22 void resetGpu(const GrGpu* gpu) { fGpu = gpu; } 30 explicit GrSemaphore(const GrGpu* gpu) : fGpu(gpu) {}
|
/external/skqp/src/gpu/ |
GrSemaphore.h | 22 void resetGpu(const GrGpu* gpu) { fGpu = gpu; } 30 explicit GrSemaphore(const GrGpu* gpu) : fGpu(gpu) {}
|
/external/skia/src/gpu/vk/ |
GrVkUniformBuffer.h | 18 static GrVkUniformBuffer* Create(GrVkGpu* gpu, size_t size); 19 static const GrVkResource* CreateResource(GrVkGpu* gpu, size_t size); 22 void* map(GrVkGpu* gpu) { 23 return this->vkMap(gpu); 25 void unmap(GrVkGpu* gpu) { 26 this->vkUnmap(gpu); 30 bool updateData(GrVkGpu* gpu, const void* src, size_t srcSizeInBytes, 32 return this->vkUpdateData(gpu, src, srcSizeInBytes, createdNewBuffer); 34 void release(const GrVkGpu* gpu) { this->vkRelease(gpu); } [all...] |
GrVkDescriptorSet.cpp | 23 void GrVkDescriptorSet::freeGPUData(const GrVkGpu* gpu) const { 24 fPool->unref(gpu); 27 void GrVkDescriptorSet::onRecycle(GrVkGpu* gpu) const { 28 gpu->resourceProvider().recycleDescriptorSet(this, fHandle);
|
GrVkSemaphore.cpp | 21 sk_sp<GrVkSemaphore> GrVkSemaphore::Make(const GrVkGpu* gpu, bool isOwned) { 28 GR_VK_CALL_ERRCHECK(gpu->vkInterface(), 29 CreateSemaphore(gpu->device(), &createInfo, nullptr, &semaphore)); 31 return sk_sp<GrVkSemaphore>(new GrVkSemaphore(gpu, semaphore, false, false, isOwned)); 34 sk_sp<GrVkSemaphore> GrVkSemaphore::MakeWrapped(const GrVkGpu* gpu, 43 return sk_sp<GrVkSemaphore>(new GrVkSemaphore(gpu, semaphore, prohibitSignal, prohibitWait, 47 GrVkSemaphore::GrVkSemaphore(const GrVkGpu* gpu, VkSemaphore semaphore, bool prohibitSignal, 49 : INHERITED(gpu) { 61 void GrVkSemaphore::Resource::freeGPUData(const GrVkGpu* gpu) const { 63 GR_VK_CALL(gpu->vkInterface() [all...] |
GrVkUniformBuffer.cpp | 11 #define VK_CALL(GPU, X) GR_VK_CALL(GPU->vkInterface(), X) 13 GrVkUniformBuffer* GrVkUniformBuffer::Create(GrVkGpu* gpu, size_t size) { 19 resource = gpu->resourceProvider().findOrCreateStandardUniformBufferResource(); 21 resource = CreateResource(gpu, size); 31 GrVkUniformBuffer* buffer = new GrVkUniformBuffer(gpu, desc, 36 resource->unref(gpu); 42 const GrVkResource* GrVkUniformBuffer::CreateResource(GrVkGpu* gpu, size_t size) { 62 err = VK_CALL(gpu, CreateBuffer(gpu->device(), &bufInfo, nullptr, &buffer)) [all...] |
GrVkCommandBuffer.h | 41 void pipelineBarrier(const GrVkGpu* gpu, 48 void bindInputBuffer(GrVkGpu* gpu, uint32_t binding, const GrVkVertexBuffer* vbuffer); 50 void bindIndexBuffer(GrVkGpu* gpu, const GrVkIndexBuffer* ibuffer); 52 void bindPipeline(const GrVkGpu* gpu, const GrVkPipeline* pipeline); 54 void bindDescriptorSets(const GrVkGpu* gpu, 63 void bindDescriptorSets(const GrVkGpu* gpu, 73 void setViewport(const GrVkGpu* gpu, 78 void setScissor(const GrVkGpu* gpu, 83 void setBlendConstants(const GrVkGpu* gpu, const float blendConstants[4]); 86 void clearAttachments(const GrVkGpu* gpu, [all...] |
GrVkDescriptorSetManager.h | 29 static GrVkDescriptorSetManager* CreateUniformManager(GrVkGpu* gpu); 30 static GrVkDescriptorSetManager* CreateSamplerManager(GrVkGpu* gpu, VkDescriptorType type, 32 static GrVkDescriptorSetManager* CreateSamplerManager(GrVkGpu* gpu, VkDescriptorType type, 38 void release(const GrVkGpu* gpu); 42 const GrVkDescriptorSet* getDescriptorSet(GrVkGpu* gpu, const Handle& handle); 52 DescriptorPoolManager(VkDescriptorType type, GrVkGpu* gpu, 61 void getNewDescriptorSet(GrVkGpu* gpu, VkDescriptorSet* ds); 63 void freeGPUResources(const GrVkGpu* gpu); 80 void getNewPool(GrVkGpu* gpu); 83 GrVkDescriptorSetManager(GrVkGpu* gpu, [all...] |
GrVkBufferView.cpp | 12 const GrVkBufferView* GrVkBufferView::Create(const GrVkGpu* gpu, VkBuffer buffer, VkFormat format, 27 VkResult err = GR_VK_CALL(gpu->vkInterface(), CreateBufferView(gpu->device(), &viewInfo, 36 void GrVkBufferView::freeGPUData(const GrVkGpu* gpu) const { 37 GR_VK_CALL(gpu->vkInterface(), DestroyBufferView(gpu->device(), fBufferView, nullptr));
|
/external/skqp/src/gpu/vk/ |
GrVkUniformBuffer.h | 18 static GrVkUniformBuffer* Create(GrVkGpu* gpu, size_t size); 19 static const GrVkResource* CreateResource(GrVkGpu* gpu, size_t size); 22 void* map(GrVkGpu* gpu) { 23 return this->vkMap(gpu); 25 void unmap(GrVkGpu* gpu) { 26 this->vkUnmap(gpu); 30 bool updateData(GrVkGpu* gpu, const void* src, size_t srcSizeInBytes, 32 return this->vkUpdateData(gpu, src, srcSizeInBytes, createdNewBuffer); 34 void release(const GrVkGpu* gpu) { this->vkRelease(gpu); } [all...] |
GrVkDescriptorSet.cpp | 23 void GrVkDescriptorSet::freeGPUData(const GrVkGpu* gpu) const { 24 fPool->unref(gpu); 27 void GrVkDescriptorSet::onRecycle(GrVkGpu* gpu) const { 28 gpu->resourceProvider().recycleDescriptorSet(this, fHandle);
|
GrVkSemaphore.cpp | 21 sk_sp<GrVkSemaphore> GrVkSemaphore::Make(const GrVkGpu* gpu, bool isOwned) { 28 GR_VK_CALL_ERRCHECK(gpu->vkInterface(), 29 CreateSemaphore(gpu->device(), &createInfo, nullptr, &semaphore)); 31 return sk_sp<GrVkSemaphore>(new GrVkSemaphore(gpu, semaphore, false, false, isOwned)); 34 sk_sp<GrVkSemaphore> GrVkSemaphore::MakeWrapped(const GrVkGpu* gpu, 43 return sk_sp<GrVkSemaphore>(new GrVkSemaphore(gpu, semaphore, prohibitSignal, prohibitWait, 47 GrVkSemaphore::GrVkSemaphore(const GrVkGpu* gpu, VkSemaphore semaphore, bool prohibitSignal, 49 : INHERITED(gpu) { 61 void GrVkSemaphore::Resource::freeGPUData(const GrVkGpu* gpu) const { 63 GR_VK_CALL(gpu->vkInterface() [all...] |
GrVkUniformBuffer.cpp | 11 #define VK_CALL(GPU, X) GR_VK_CALL(GPU->vkInterface(), X) 13 GrVkUniformBuffer* GrVkUniformBuffer::Create(GrVkGpu* gpu, size_t size) { 19 resource = gpu->resourceProvider().findOrCreateStandardUniformBufferResource(); 21 resource = CreateResource(gpu, size); 31 GrVkUniformBuffer* buffer = new GrVkUniformBuffer(gpu, desc, 36 resource->unref(gpu); 42 const GrVkResource* GrVkUniformBuffer::CreateResource(GrVkGpu* gpu, size_t size) { 62 err = VK_CALL(gpu, CreateBuffer(gpu->device(), &bufInfo, nullptr, &buffer)) [all...] |
GrVkCommandBuffer.h | 41 void pipelineBarrier(const GrVkGpu* gpu, 48 void bindInputBuffer(GrVkGpu* gpu, uint32_t binding, const GrVkVertexBuffer* vbuffer); 50 void bindIndexBuffer(GrVkGpu* gpu, const GrVkIndexBuffer* ibuffer); 52 void bindPipeline(const GrVkGpu* gpu, const GrVkPipeline* pipeline); 54 void bindDescriptorSets(const GrVkGpu* gpu, 63 void bindDescriptorSets(const GrVkGpu* gpu, 73 void setViewport(const GrVkGpu* gpu, 78 void setScissor(const GrVkGpu* gpu, 83 void setBlendConstants(const GrVkGpu* gpu, const float blendConstants[4]); 86 void clearAttachments(const GrVkGpu* gpu, [all...] |
GrVkDescriptorSetManager.h | 29 static GrVkDescriptorSetManager* CreateUniformManager(GrVkGpu* gpu); 30 static GrVkDescriptorSetManager* CreateSamplerManager(GrVkGpu* gpu, VkDescriptorType type, 32 static GrVkDescriptorSetManager* CreateSamplerManager(GrVkGpu* gpu, VkDescriptorType type, 38 void release(const GrVkGpu* gpu); 42 const GrVkDescriptorSet* getDescriptorSet(GrVkGpu* gpu, const Handle& handle); 52 DescriptorPoolManager(VkDescriptorType type, GrVkGpu* gpu, 61 void getNewDescriptorSet(GrVkGpu* gpu, VkDescriptorSet* ds); 63 void freeGPUResources(const GrVkGpu* gpu); 80 void getNewPool(GrVkGpu* gpu); 83 GrVkDescriptorSetManager(GrVkGpu* gpu, [all...] |
/external/tensorflow/tensorflow/core/common_runtime/ |
gpu_device_context.h | 30 namespace gpu = ::perftools::gputools; 35 GPUDeviceContext(int stream_id, gpu::Stream* stream, 36 gpu::Stream* host_to_device_stream, 37 gpu::Stream* device_to_host_stream, 38 gpu::Stream* device_to_device_stream) 47 gpu::Stream* stream() const override { return stream_; } 48 gpu::Stream* host_to_device_stream() const { return host_to_device_stream_; } 49 gpu::Stream* device_to_host_stream() const { return device_to_host_stream_; } 50 gpu::Stream* device_to_device_stream() const { 70 gpu::Stream* stream_ [all...] |
/external/skia/src/gpu/gl/ |
GrGLSemaphore.cpp | 12 GrGLSemaphore::GrGLSemaphore(const GrGLGpu* gpu, bool isOwned) 13 : INHERITED(gpu), fSync(0), fIsOwned(isOwned) {
|
GrGLTextureRenderTarget.cpp | 15 GrGLTextureRenderTarget::GrGLTextureRenderTarget(GrGLGpu* gpu, 21 : GrSurface(gpu, desc) 22 , GrGLTexture(gpu, desc, texIDDesc, mipMapsStatus) 23 , GrGLRenderTarget(gpu, desc, rtIDDesc) { 27 GrGLTextureRenderTarget::GrGLTextureRenderTarget(GrGLGpu* gpu, 32 : GrSurface(gpu, desc) 33 , GrGLTexture(gpu, desc, texIDDesc, mipMapsStatus) 34 , GrGLRenderTarget(gpu, desc, rtIDDesc) { 59 GrGLGpu* gpu, const GrSurfaceDesc& desc, const GrGLTexture::IDDesc& texIDDesc, 63 new GrGLTextureRenderTarget(gpu, desc, texIDDesc, rtIDDesc, mipMapsStatus)) [all...] |
/external/skqp/src/gpu/gl/ |
GrGLSemaphore.cpp | 12 GrGLSemaphore::GrGLSemaphore(const GrGLGpu* gpu, bool isOwned) 13 : INHERITED(gpu), fSync(0), fIsOwned(isOwned) {
|
/external/tensorflow/tensorflow/core/common_runtime/gpu/ |
gpu_id_utils.h | 19 #include "tensorflow/core/common_runtime/gpu/gpu_id.h" 20 #include "tensorflow/core/common_runtime/gpu/gpu_id_manager.h" 21 #include "tensorflow/core/common_runtime/gpu/gpu_init.h" 26 namespace gpu = ::perftools::gputools; 28 // Utility methods for translation between Tensorflow GPU ids and CUDA GPU ids. 33 static gpu::port::StatusOr<gpu::StreamExecutor*> ExecutorForCudaGpuId( 34 gpu::Platform* gpu_manager, CudaGpuId cuda_gpu_id) { 37 static gpu::port::StatusOr<gpu::StreamExecutor*> ExecutorForCudaGpuId [all...] |
gpu_init.cc | 16 #include "tensorflow/core/common_runtime/gpu/gpu_init.h" 29 namespace gpu = ::perftools::gputools; 34 auto result = gpu::MultiPlatformManager::PlatformWithName("CUDA"); 42 gpu::Platform* GPUMachineManager() { 43 auto result = gpu::MultiPlatformManager::PlatformWithName("CUDA");
|
/external/autotest/client/site_tests/graphics_dEQP/scripts/ |
failure_matrix.py | 64 def get_problem_count(dict, gpu): 65 if gpu in dict: 66 if not dict[gpu]: 69 for status in dict[gpu]: 71 count = count + len((dict[gpu])[status]) 78 if status in dict[gpu]: 79 count = count + len((dict[gpu])[status]) 81 print 'Warning: %s not found in dict!' % gpu 87 for gpu in dict: 89 if status in dict[gpu] [all...] |