/external/skia/src/gpu/vk/ |
GrVkImageView.cpp | 12 const GrVkImageView* GrVkImageView::Create(const GrVkGpu* gpu, VkImage image, VkFormat format, 34 VkResult err = GR_VK_CALL(gpu->vkInterface(), CreateImageView(gpu->device(), &viewInfo, 43 void GrVkImageView::freeGPUData(const GrVkGpu* gpu) const { 44 GR_VK_CALL(gpu->vkInterface(), DestroyImageView(gpu->device(), fImageView, nullptr));
|
GrVkDescriptorPool.cpp | 14 GrVkDescriptorPool::GrVkDescriptorPool(const GrVkGpu* gpu, VkDescriptorType type, uint32_t count) 33 GR_VK_CALL_ERRCHECK(gpu->vkInterface(), CreateDescriptorPool(gpu->device(), 43 void GrVkDescriptorPool::reset(const GrVkGpu* gpu) { 44 GR_VK_CALL_ERRCHECK(gpu->vkInterface(), ResetDescriptorPool(gpu->device(), fDescPool, 0)); 47 void GrVkDescriptorPool::freeGPUData(const GrVkGpu* gpu) const { 50 GR_VK_CALL(gpu->vkInterface(), DestroyDescriptorPool(gpu->device(), fDescPool, nullptr));
|
GrVkBuffer.cpp | 14 #define VK_CALL(GPU, X) GR_VK_CALL(GPU->vkInterface(), X) 22 const GrVkBuffer::Resource* GrVkBuffer::Create(const GrVkGpu* gpu, const Desc& desc) { 60 err = VK_CALL(gpu, CreateBuffer(gpu->device(), &bufInfo, nullptr, &buffer)); 65 if (!GrVkMemory::AllocAndBindBufferMemory(gpu, 75 VK_CALL(gpu, DestroyBuffer(gpu->device(), buffer, nullptr)); 76 GrVkMemory::FreeBufferMemory(gpu, desc.fType, alloc); 83 void GrVkBuffer::addMemoryBarrier(const GrVkGpu* gpu, [all...] |
GrVkBuffer.h | 35 void addMemoryBarrier(const GrVkGpu* gpu, 73 void freeGPUData(const GrVkGpu* gpu) const override; 75 void onRecycle(GrVkGpu* gpu) const override { this->unref(gpu); } 81 static const Resource* Create(const GrVkGpu* gpu, 88 void* vkMap(GrVkGpu* gpu) { 89 this->internalMap(gpu, fDesc.fSizeInBytes); 92 void vkUnmap(GrVkGpu* gpu) { this->internalUnmap(gpu, this->size()); } 96 bool vkUpdateData(GrVkGpu* gpu, const void* src, size_t srcSizeInBytes [all...] |
GrVkTextureRenderTarget.cpp | 19 #define VK_CALL(GPU, X) GR_VK_CALL(GPU->vkInterface(), X) 21 GrVkTextureRenderTarget::GrVkTextureRenderTarget(GrVkGpu* gpu, 31 : GrSurface(gpu, desc) 33 , GrVkTexture(gpu, desc, info, texView, mipMapsStatus, ownership) 34 , GrVkRenderTarget(gpu, desc, info, msaaInfo, colorAttachmentView, 39 GrVkTextureRenderTarget::GrVkTextureRenderTarget(GrVkGpu* gpu, 47 : GrSurface(gpu, desc) 49 , GrVkTexture(gpu, desc, info, texView, mipMapsStatus, ownership) 50 , GrVkRenderTarget(gpu, desc, info, colorAttachmentView, GrBackendObjectOwnership::kOwned) [all...] |
GrVkRenderTarget.cpp | 21 #define VK_CALL(GPU, X) GR_VK_CALL(GPU->vkInterface(), X) 25 GrVkRenderTarget::GrVkRenderTarget(GrVkGpu* gpu, 33 : GrSurface(gpu, desc) 36 , GrRenderTarget(gpu, desc) 43 this->createFramebuffer(gpu); 49 GrVkRenderTarget::GrVkRenderTarget(GrVkGpu* gpu, 56 : GrSurface(gpu, desc) 59 , GrRenderTarget(gpu, desc) 66 this->createFramebuffer(gpu); 281 GrVkGpu* gpu = this->getVkGpu(); local [all...] |
GrVkResource.h | 114 Any GPU data associated with this resource will be freed before it's deleted. 116 void unref(const GrVkGpu* gpu) const { 118 SkASSERT(gpu); 123 this->internal_dispose(gpu); 127 /** Unref without freeing GPU data. Used only when we're abandoning the resource */ 154 virtual void freeGPUData(const GrVkGpu* gpu) const = 0; 166 void internal_dispose(const GrVkGpu* gpu) const { 167 this->freeGPUData(gpu); 203 void recycle(GrVkGpu* gpu) const { 205 this->onRecycle(gpu); [all...] |
/external/skia/tests/ |
VkMakeCopyPipelineTest.cpp | 8 // This is a GPU-backend specific test. It relies on static intializers to work 33 void test(GrVkGpu* gpu, skiatest::Reporter* reporter) { 67 if (!GrCompileVkShaderModule(gpu, vertShaderText, VK_SHADER_STAGE_VERTEX_BIT, 69 this->destroyResources(gpu); 75 if (!GrCompileVkShaderModule(gpu, fragShaderText, VK_SHADER_STAGE_FRAGMENT_BIT, 77 this->destroyResources(gpu); 84 GrVkResourceProvider& resourceProvider = gpu->resourceProvider(); 107 VkResult err = GR_VK_CALL(gpu->vkInterface(), CreatePipelineLayout(gpu->device(), 112 this->destroyResources(gpu); 181 GrVkGpu* gpu = static_cast<GrVkGpu*>(context->contextPriv().getGpu()); local [all...] |
/external/skqp/src/gpu/vk/ |
GrVkBufferView.cpp | 12 const GrVkBufferView* GrVkBufferView::Create(const GrVkGpu* gpu, VkBuffer buffer, VkFormat format, 27 VkResult err = GR_VK_CALL(gpu->vkInterface(), CreateBufferView(gpu->device(), &viewInfo, 36 void GrVkBufferView::freeGPUData(const GrVkGpu* gpu) const { 37 GR_VK_CALL(gpu->vkInterface(), DestroyBufferView(gpu->device(), fBufferView, nullptr));
|
GrVkImageView.cpp | 12 const GrVkImageView* GrVkImageView::Create(const GrVkGpu* gpu, VkImage image, VkFormat format, 34 VkResult err = GR_VK_CALL(gpu->vkInterface(), CreateImageView(gpu->device(), &viewInfo, 43 void GrVkImageView::freeGPUData(const GrVkGpu* gpu) const { 44 GR_VK_CALL(gpu->vkInterface(), DestroyImageView(gpu->device(), fImageView, nullptr));
|
GrVkDescriptorPool.cpp | 14 GrVkDescriptorPool::GrVkDescriptorPool(const GrVkGpu* gpu, VkDescriptorType type, uint32_t count) 33 GR_VK_CALL_ERRCHECK(gpu->vkInterface(), CreateDescriptorPool(gpu->device(), 43 void GrVkDescriptorPool::reset(const GrVkGpu* gpu) { 44 GR_VK_CALL_ERRCHECK(gpu->vkInterface(), ResetDescriptorPool(gpu->device(), fDescPool, 0)); 47 void GrVkDescriptorPool::freeGPUData(const GrVkGpu* gpu) const { 50 GR_VK_CALL(gpu->vkInterface(), DestroyDescriptorPool(gpu->device(), fDescPool, nullptr));
|
GrVkBuffer.cpp | 14 #define VK_CALL(GPU, X) GR_VK_CALL(GPU->vkInterface(), X) 22 const GrVkBuffer::Resource* GrVkBuffer::Create(const GrVkGpu* gpu, const Desc& desc) { 60 err = VK_CALL(gpu, CreateBuffer(gpu->device(), &bufInfo, nullptr, &buffer)); 65 if (!GrVkMemory::AllocAndBindBufferMemory(gpu, 75 VK_CALL(gpu, DestroyBuffer(gpu->device(), buffer, nullptr)); 76 GrVkMemory::FreeBufferMemory(gpu, desc.fType, alloc); 83 void GrVkBuffer::addMemoryBarrier(const GrVkGpu* gpu, [all...] |
GrVkBuffer.h | 35 void addMemoryBarrier(const GrVkGpu* gpu, 73 void freeGPUData(const GrVkGpu* gpu) const override; 75 void onRecycle(GrVkGpu* gpu) const override { this->unref(gpu); } 81 static const Resource* Create(const GrVkGpu* gpu, 88 void* vkMap(GrVkGpu* gpu) { 89 this->internalMap(gpu, fDesc.fSizeInBytes); 92 void vkUnmap(GrVkGpu* gpu) { this->internalUnmap(gpu, this->size()); } 96 bool vkUpdateData(GrVkGpu* gpu, const void* src, size_t srcSizeInBytes [all...] |
GrVkTextureRenderTarget.cpp | 19 #define VK_CALL(GPU, X) GR_VK_CALL(GPU->vkInterface(), X) 21 GrVkTextureRenderTarget::GrVkTextureRenderTarget(GrVkGpu* gpu, 31 : GrSurface(gpu, desc) 33 , GrVkTexture(gpu, desc, info, texView, mipMapsStatus, ownership) 34 , GrVkRenderTarget(gpu, desc, info, msaaInfo, colorAttachmentView, 39 GrVkTextureRenderTarget::GrVkTextureRenderTarget(GrVkGpu* gpu, 47 : GrSurface(gpu, desc) 49 , GrVkTexture(gpu, desc, info, texView, mipMapsStatus, ownership) 50 , GrVkRenderTarget(gpu, desc, info, colorAttachmentView, GrBackendObjectOwnership::kOwned) [all...] |
GrVkRenderTarget.cpp | 21 #define VK_CALL(GPU, X) GR_VK_CALL(GPU->vkInterface(), X) 25 GrVkRenderTarget::GrVkRenderTarget(GrVkGpu* gpu, 33 : GrSurface(gpu, desc) 36 , GrRenderTarget(gpu, desc) 43 this->createFramebuffer(gpu); 49 GrVkRenderTarget::GrVkRenderTarget(GrVkGpu* gpu, 56 : GrSurface(gpu, desc) 59 , GrRenderTarget(gpu, desc) 66 this->createFramebuffer(gpu); 281 GrVkGpu* gpu = this->getVkGpu(); local [all...] |
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
while_transformer.h | 23 namespace gpu { namespace in namespace:xla 40 } // namespace gpu
|
/external/tensorflow/tensorflow/core/platform/default/ |
from_stream_executor_status.h | 26 namespace gpu = ::perftools::gputools;
|
/external/tensorflow/tensorflow/core/kernels/ |
xent_op_test.cc | 43 /// The representative tests for ptb_word on GPU 44 BM_XentDev(16, 10000, gpu); 45 BM_XentDev(16, 30000, gpu); 46 BM_XentDev(16, 100000, gpu); 48 BM_XentDev(32, 10000, gpu); 49 BM_XentDev(32, 30000, gpu); 50 BM_XentDev(32, 100000, gpu); 52 BM_XentDev(64, 10000, gpu); 53 BM_XentDev(64, 30000, gpu); 54 BM_XentDev(64, 100000, gpu); [all...] |
/external/skqp/tests/ |
VkMakeCopyPipelineTest.cpp | 8 // This is a GPU-backend specific test. It relies on static intializers to work 33 void test(GrVkGpu* gpu, skiatest::Reporter* reporter) { 66 if (!GrCompileVkShaderModule(gpu, vertShaderText, VK_SHADER_STAGE_VERTEX_BIT, 68 this->destroyResources(gpu); 74 if (!GrCompileVkShaderModule(gpu, fragShaderText, VK_SHADER_STAGE_FRAGMENT_BIT, 76 this->destroyResources(gpu); 83 GrVkResourceProvider& resourceProvider = gpu->resourceProvider(); 106 VkResult err = GR_VK_CALL(gpu->vkInterface(), CreatePipelineLayout(gpu->device(), 111 this->destroyResources(gpu); 180 GrVkGpu* gpu = static_cast<GrVkGpu*>(context->contextPriv().getGpu()); local [all...] |
/external/skia/src/gpu/ |
GrResourceProviderPriv.h | 18 GrGpu* gpu() { return fResourceProvider->gpu(); } function in class:GrResourceProviderPriv
|
/external/skia/src/gpu/mock/ |
GrMockStencilAttachment.h | 16 GrMockStencilAttachment(GrMockGpu* gpu, int width, int height, int bits, int sampleCnt) 17 : INHERITED(gpu, width, height, bits, sampleCnt) {
|
/external/skqp/src/gpu/ |
GrResourceProviderPriv.h | 18 GrGpu* gpu() { return fResourceProvider->gpu(); } function in class:GrResourceProviderPriv
|
/external/skqp/src/gpu/mock/ |
GrMockStencilAttachment.h | 16 GrMockStencilAttachment(GrMockGpu* gpu, int width, int height, int bits, int sampleCnt) 17 : INHERITED(gpu, width, height, bits, sampleCnt) {
|
/external/skia/src/gpu/gl/ |
GrGLSemaphore.h | 19 static sk_sp<GrGLSemaphore> Make(const GrGLGpu* gpu, bool isOwned) { 20 return sk_sp<GrGLSemaphore>(new GrGLSemaphore(gpu, isOwned)); 23 static sk_sp<GrGLSemaphore> MakeWrapped(const GrGLGpu* gpu, 26 auto sema = sk_sp<GrGLSemaphore>(new GrGLSemaphore(gpu, 38 GrGLSemaphore(const GrGLGpu* gpu, bool isOwned);
|
/external/skqp/src/gpu/gl/ |
GrGLSemaphore.h | 19 static sk_sp<GrGLSemaphore> Make(const GrGLGpu* gpu, bool isOwned) { 20 return sk_sp<GrGLSemaphore>(new GrGLSemaphore(gpu, isOwned)); 23 static sk_sp<GrGLSemaphore> MakeWrapped(const GrGLGpu* gpu, 26 auto sema = sk_sp<GrGLSemaphore>(new GrGLSemaphore(gpu, 38 GrGLSemaphore(const GrGLGpu* gpu, bool isOwned);
|