/external/deqp/external/vulkancts/modules/vulkan/compute/ |
vktComputeIndirectComputeDispatchTests.cpp | 182 const vk::Allocation& alloc = indirectBuffer.getAllocation(); local 183 deUint8* indirectDataPtr = reinterpret_cast<deUint8*>(alloc.getHostPtr()); 198 vk::flushMappedMemoryRange(m_device_interface, m_device, alloc.getMemory(), alloc.getOffset(), m_bufferSize); 228 const vk::Allocation& alloc = resultBuffer.getAllocation(); local 229 deUint8* resultDataPtr = reinterpret_cast<deUint8*>(alloc.getHostPtr()); 241 vk::flushMappedMemoryRange(m_device_interface, m_device, alloc.getMemory(), alloc.getOffset(), resultBufferSize); 334 const vk::Allocation& alloc = resultBuffer.getAllocation(); local 335 vk::invalidateMappedMemoryRange(m_device_interface, m_device, alloc.getMemory(), alloc.getOffset(), resultBufferSize) [all...] |
vktComputeShaderBuiltinVarTests.cpp | 417 const Allocation& alloc = uniformBuffer.getAllocation(); local 418 memcpy(alloc.getHostPtr(), &stride, sizeOfUniformBuffer); 419 flushMappedMemoryRange(m_vki, m_device, alloc.getMemory(), alloc.getOffset(), sizeOfUniformBuffer);
|
/external/deqp/external/vulkancts/modules/vulkan/geometry/ |
vktGeometryTestsUtil.cpp | 697 de::MovePtr<Allocation> alloc = allocator.allocate(getImageMemoryRequirements(vk, device, image), requirement); local 698 VK_CHECK(vk.bindImageMemory(device, image, alloc->getMemory(), alloc->getOffset())); 699 return alloc; 704 de::MovePtr<Allocation> alloc(allocator.allocate(getBufferMemoryRequirements(vk, device, buffer), requirement)); 705 VK_CHECK(vk.bindBufferMemory(device, buffer, alloc->getMemory(), alloc->getOffset())); 706 return alloc; 709 void zeroBuffer (const DeviceInterface& vk, const VkDevice device, const Allocation& alloc, const VkDeviceSize size) 711 deMemset(alloc.getHostPtr(), 0, static_cast<std::size_t>(size)) [all...] |
/external/deqp/external/vulkancts/modules/vulkan/image/ |
vktImageMultisampleLoadStoreTests.cpp | 272 const Allocation& alloc = constantsBuffer->getAllocation(); local 273 deUint8* const basePtr = static_cast<deUint8*>(alloc.getHostPtr()); 275 deMemset(alloc.getHostPtr(), 0, static_cast<size_t>(constantsBufferSizeBytes)); 283 flushMappedMemoryRange(vk, device, alloc.getMemory(), alloc.getOffset(), constantsBufferSizeBytes); 291 const Allocation& alloc = resultBuffer->getAllocation(); local 292 deMemset(alloc.getHostPtr(), 0, static_cast<size_t>(resultBufferSizeBytes)); 293 flushMappedMemoryRange(vk, device, alloc.getMemory(), alloc.getOffset(), resultBufferSizeBytes); 436 const Allocation& alloc = resultBuffer->getAllocation() local [all...] |
/external/deqp/external/vulkancts/modules/vulkan/spirv_assembly/ |
vktSpvAsmComputeShaderCase.cpp | 360 AllocationMp alloc; local 367 BufferHandleUp* buffer = new BufferHandleUp(createBufferAndBindMemory(vkdi, device, descriptorTypes.back(), allocator, numBytes, &alloc)); 369 setMemory(vkdi, device, &*alloc, numBytes, &inputBytes.front()); 372 inputAllocs.push_back(de::SharedPtr<Allocation>(alloc.release())); 379 AllocationMp alloc; local 386 BufferHandleUp* buffer = new BufferHandleUp(createBufferAndBindMemory(vkdi, device, descriptorTypes.back(), allocator, numBytes, &alloc)); 388 fillMemoryWithValue(vkdi, device, &*alloc, numBytes, 0xff); 391 outputAllocs.push_back(de::SharedPtr<Allocation>(alloc.release()));
|
/external/deqp/external/vulkancts/modules/vulkan/tessellation/ |
vktTessellationGeometryGridRenderTests.cpp | 627 const Allocation& alloc = colorBuffer.getAllocation(); local 628 invalidateMappedMemoryRange(vk, device, alloc.getMemory(), alloc.getOffset(), colorBufferSizeBytes); 630 const tcu::ConstPixelBufferAccess imageAllLayers(mapVkFormat(colorFormat), renderSize.x(), renderSize.y(), m_params.numLayers, alloc.getHostPtr());
|
vktTessellationGeometryPassthroughTests.cpp | 355 const Allocation& alloc = colorBuffer.getAllocation(); local 356 invalidateMappedMemoryRange(vk, device, alloc.getMemory(), alloc.getOffset(), colorBufferSizeBytes); 357 return tcu::ConstPixelBufferAccess(mapVkFormat(colorFormat), renderSize.x(), renderSize.y(), 1, alloc.getHostPtr()); 434 const Allocation& alloc = tessLevelsBuffer.getAllocation(); local 435 TessLevels* const bufferTessLevels = static_cast<TessLevels*>(alloc.getHostPtr()); 437 flushMappedMemoryRange(vk, device, alloc.getMemory(), alloc.getOffset(), sizeof(TessLevels)); 447 const Allocation& alloc = vertexBuffer.getAllocation(); local 448 deMemcpy(alloc.getHostPtr(), &m_params.vertices[0], static_cast<std::size_t>(vertexDataSizeBytes)) [all...] |
vktTessellationShaderInputOutputTests.cpp | 86 const Allocation& alloc = vertexBuffer.getAllocation(); local 87 deMemcpy(alloc.getHostPtr(), vertexData, static_cast<std::size_t>(vertexDataSizeBytes)); 89 flushMappedMemoryRange(vk, device, alloc.getMemory(), alloc.getOffset(), vertexDataSizeBytes);
|
/external/deqp/framework/delibs/depool/ |
deMemPool.c | 187 /* Alloc pool from initial page. */ 344 DebugAlloc* alloc = pool->debugAllocListHead; local 347 while (alloc) 349 next = alloc->next; 350 deAlignedFree(alloc->memPtr); 351 deFree(alloc); 352 alloc = next;
|
/external/lzma/C/ |
Lzma2Enc.c | 63 ISzAlloc *alloc, ISzAlloc *allocBig);
65 UInt32 keepWindowSize, ISzAlloc *alloc, ISzAlloc *allocBig);
280 ISzAlloc *alloc;
member in struct:__anon28720 302 mainEncoder->outBuf = (Byte *)IAlloc_Alloc(mainEncoder->alloc, LZMA2_CHUNK_SIZE_COMPRESSED_MAX);
309 mainEncoder->alloc, mainEncoder->allocBig));
363 mainEncoder->alloc, mainEncoder->allocBig));
406 CLzma2EncHandle Lzma2Enc_Create(ISzAlloc *alloc, ISzAlloc *allocBig)
408 CLzma2Enc *p = (CLzma2Enc *)alloc->Alloc(alloc, sizeof(CLzma2Enc)); [all...] |
XzEnc.c | 10 #include "Alloc.h"
130 static SRes Xz_AddIndexRecord(CXzStream *p, UInt64 unpackSize, UInt64 totalSize, ISzAlloc *alloc)
139 blocks = (CXzBlockSizes *)alloc->Alloc(alloc, newSize);
145 alloc->Free(alloc, p->blocks);
270 SRes BraState_SetFromMethod(IStateCoder *p, UInt64 id, int encodeMode, ISzAlloc *alloc);
276 p->buf = g_Alloc.Alloc(&g_Alloc, FILTER_BUF_SIZE);
329 void SbEncInStream_Construct(CSbEncInStream *p, ISzAlloc *alloc)
355 ISzAlloc *alloc; member in struct:__anon28761 [all...] |
/external/mesa3d/src/vulkan/wsi/ |
wsi_common_wayland.c | 61 const VkAllocationCallbacks *alloc; member in struct:wsi_wayland 253 vk_free(wsi->alloc, display); 260 vk_alloc(wsi->alloc, sizeof(*display), 8, 352 const VkAllocationCallbacks *alloc, 775 const VkAllocationCallbacks *alloc, 782 wsi = vk_alloc(alloc, sizeof(*wsi), 8, 790 wsi->alloc = alloc; 825 vk_free(alloc, wsi); 834 const VkAllocationCallbacks *alloc) [all...] |
/external/python/cpython2/Modules/_io/ |
stringio.c | 58 size_t alloc = self->buf_size; local 70 if (size < alloc / 2) { 72 alloc = size + 1; 74 else if (size < alloc) { 78 else if (size <= alloc * 1.125) { 80 alloc = size + (size >> 3) + (size < 9 ? 3 : 6); 84 alloc = size + 1; 87 if (alloc > ((size_t)-1) / sizeof(Py_UNICODE)) 90 alloc * sizeof(Py_UNICODE)); 95 self->buf_size = alloc; [all...] |
/external/python/cpython3/Modules/_io/ |
bytesio.c | 111 size_t alloc = PyBytes_GET_SIZE(self->buf); local 120 if (size < alloc / 2) { 122 alloc = size + 1; 124 else if (size < alloc) { 128 else if (size <= alloc * 1.125) { 130 alloc = size + (size >> 3) + (size < 9 ? 3 : 6); 134 alloc = size + 1; 137 if (alloc > ((size_t)-1) / sizeof(char)) 141 if (unshare_buffer(self, alloc) < 0) 145 if (_PyBytes_Resize(&self->buf, alloc) < 0 [all...] |
stringio.c | 78 size_t alloc = self->buf_size; local 90 if (size < alloc / 2) { 92 alloc = size + 1; 94 else if (size < alloc) { 98 else if (size <= alloc * 1.125) { 100 alloc = size + (size >> 3) + (size < 9 ? 3 : 6); 104 alloc = size + 1; 107 if (alloc > PY_SIZE_MAX / sizeof(Py_UCS4)) 109 new_buf = (Py_UCS4 *)PyMem_Realloc(self->buf, alloc * sizeof(Py_UCS4)); 114 self->buf_size = alloc; [all...] |
/external/skia/src/core/ |
SkBlitter.cpp | 85 SkArenaAlloc* alloc) { 89 int16_t* runs = alloc->makeArrayDefault<int16_t>(runSize); 90 SkAlpha* alphas = alloc->makeArrayDefault<SkAlpha>(runSize); 114 SkSTArenaAlloc<SkCoverageDeltaMask::MAX_SIZE> alloc; local 115 SkCoverageDeltaMask mask(&alloc, rowIR); 697 Context* onMakeContext(const ContextRec& rec, SkArenaAlloc* alloc) const override { 700 proxyContext = as_SB(fProxy)->makeContext(rec, alloc); 705 return alloc->make<Sk3DShaderContext>(*this, rec, proxyContext); 933 SkArenaAlloc* alloc, 935 SkASSERT(alloc != nullptr) [all...] |
SkScan_DAAPath.cpp | 337 constexpr int STACK_SIZE = 12 << 10; // 12K stack size alloc; Google3 has 16K limit. 343 // Set alloc to record's alloc if and only if we're in the init-once phase. We have to do that 346 SkArenaAlloc* alloc = isInitOnce ? record->fAlloc : &stackAlloc; local 349 record = alloc->make<SkDAARecord>(alloc); 359 SkCoverageDeltaMask deltaMask(alloc, clippedIR); 365 SkCoverageDeltaList* deltaList = alloc->make<SkCoverageDeltaList>( 366 alloc, clippedIR.fTop, clippedIR.fBottom, forceRLE); 378 clipBounds, isEvenOdd, isInverse, isConvex, alloc); [all...] |
/external/skqp/src/core/ |
SkBlitter.cpp | 107 SkSTArenaAlloc<SkCoverageDeltaMask::MAX_SIZE> alloc; local 108 SkCoverageDeltaMask mask(&alloc, rowIR); 694 Context* onMakeContext(const ContextRec& rec, SkArenaAlloc* alloc) const override { 697 proxyContext = as_SB(fProxy)->makeContext(rec, alloc); 702 return alloc->make<Sk3DShaderContext>(*this, rec, proxyContext); 930 SkArenaAlloc* alloc, 932 SkASSERT(alloc != nullptr); 938 return alloc->make<SkNullBlitter>(); 964 return alloc->make<SkNullBlitter>(); [all...] |
SkScan_DAAPath.cpp | 339 constexpr int STACK_SIZE = 12 << 10; // 12K stack size alloc; Google3 has 16K limit. 343 SkSTArenaAlloc<STACK_SIZE> alloc; // avoid heap allocation with SkSTArenaAlloc local 348 SkCoverageDeltaMask deltaMask(&alloc, clippedIR); 353 SkCoverageDeltaList deltaList(&alloc, clippedIR.fTop, clippedIR.fBottom, forceRLE);
|
/external/tensorflow/tensorflow/contrib/gdr/ |
gdr_memory_manager.cc | 190 void* Alloc(size_t alignment, size_t num_bytes) override { 433 Allocator* alloc = ProcessState::singleton()->GetCUDAHostAllocator(0); local 434 Tensor* host_copy = new Tensor(alloc, tensor.dtype(), tensor.shape()); 535 Allocator* alloc = ProcessState::singleton()->GetCUDAHostAllocator(0); local 536 host_copy = Tensor(alloc, tensor->dtype(), tensor->shape());
|
/frameworks/native/libs/binder/ |
MemoryDealer.cpp | 153 ssize_t alloc(size_t size, uint32_t flags); 312 ssize_t offset = alloc(size, flags); 326 ssize_t SimpleBestFitAllocator::alloc(size_t size, uint32_t flags) function in class:android::SimpleBestFitAllocator
|
/hardware/google/av/codec2/vndk/ |
C2AllocatorIon.cpp | 212 static Impl *Alloc(int ionFd, size_t size, size_t align, unsigned heapMask, unsigned flags, C2Allocator::id_t id) { 404 mImpl(Impl::Alloc(ionFd, size, align, heapMask, flags, id)) { } 522 std::shared_ptr<C2AllocationIon> alloc local 524 ret = alloc->status(); 526 *allocation = alloc; 544 std::shared_ptr<C2AllocationIon> alloc local 546 c2_status_t ret = alloc->status(); 548 *allocation = alloc;
|
/hardware/google/av/codec2/vndk/bufferpool/ |
AccessorImpl.cpp | 53 const std::shared_ptr<BufferPoolAllocation> &alloc, 57 mAllocation(alloc), mAllocSize(allocSize), mConfig(allocConfig) {} 182 std::shared_ptr<BufferPoolAllocation> alloc; local 184 status = mAllocator->allocate(params, &alloc, &allocSize); 187 status = mBufferPool.addNewBuffer(alloc, allocSize, params, bufferId, handle); 480 const std::shared_ptr<BufferPoolAllocation> &alloc, 489 bufferId, alloc, allocSize, params); 495 *handle = alloc->handle(); 509 "%zu(%zu size) used buffers - %zu/%zu (recycle/alloc) - "
|
/hardware/google/av/media/sfplugin/ |
Codec2Buffer.cpp | 500 std::function<sp<ABuffer>(size_t)> alloc) { 518 buffer = alloc(converter.backBufferSize()); 562 const std::shared_ptr<C2Allocator> &alloc) 564 mAlloc(alloc) { 585 std::shared_ptr<C2GraphicAllocation> alloc; local 586 c2_status_t err = mAlloc->priorGraphicAllocation(handle, &alloc); 591 std::shared_ptr<C2GraphicBlock> block = _C2BlockFactory::CreateGraphicBlock(alloc); 609 std::function<sp<ABuffer>(size_t)> alloc) { 631 aBuffer = alloc(converter.backBufferSize()); 653 std::function<sp<ABuffer>(size_t)> alloc) { [all...] |
/hardware/interfaces/bluetooth/1.0/default/ |
bt_vendor_lib.h | 258 * Vendor lib needs to request a buffer through the alloc callout function 311 * needs to be first allocated through the alloc callout function. 350 malloc_cb alloc; member in struct:__anon47999
|