/external/mesa3d/src/gallium/drivers/nouveau/ |
nouveau_mm.c | 53 mm_slab_alloc(struct mm_slab *slab) 57 if (slab->free == 0) 60 for (i = 0; i < (slab->count + 31) / 32; ++i) { 61 b = ffs(slab->bits[i]) - 1; 64 assert(n < slab->count); 65 slab->free--; 66 slab->bits[i] &= ~(1 << b); 74 mm_slab_free(struct mm_slab *slab, int i) 76 assert(i < slab->count); 77 slab->bits[i / 32] |= 1 << (i % 32) 123 struct mm_slab *slab; local 168 struct mm_slab *slab; local 219 struct mm_slab *slab = (struct mm_slab *)alloc->priv; local 269 struct mm_slab *slab, *next; local [all...] |
/external/swiftshader/third_party/LLVM/lib/Support/ |
Allocator.cpp | 44 /// StartNewSlab - Allocate a new slab and move the bump pointers over into 45 /// the new slab. Modifies CurPtr and End. 48 // to allocate more. Increase slab size to reduce mallocs and possibly memory 62 void BumpPtrAllocator::DeallocateSlabs(MemSlab *Slab) { 63 while (Slab) { 64 MemSlab *NextSlab = Slab->NextPtr; 68 sys::Memory::setRangeWritable(Slab + 1, Slab->Size - sizeof(MemSlab)); 69 memset(Slab + 1, 0xCD, Slab->Size - sizeof(MemSlab)) [all...] |
/external/mesa3d/src/gallium/auxiliary/pipebuffer/ |
pb_slab.c | 33 /* All slab allocations from the same heap and with the same size belong 43 * The intention is to keep serving allocations from the same slab as long 46 * Due to a race in new slab allocation, additional slabs in this list 56 struct pb_slab *slab = entry->slab; local 59 LIST_ADD(&entry->head, &slab->free); 60 slab->num_free++; 62 /* Add slab to the group's list if it isn't already linked. */ 63 if (!slab->head.next) { 65 LIST_ADDTAIL(&slab->head, &group->slabs) 103 struct pb_slab *slab; local [all...] |
pb_bufmgr_slab.c | 55 * Buffer in a slab. 63 struct pb_slab *slab; member in struct:pb_slab_buffer 69 /** Offset relative to the start of the slab buffer. */ 79 * Slab -- a contiguous piece of memory. 116 * Alignment, usage to be used to allocate the slab buffers. 139 * This buffer manager just dispatches buffer allocations to the appropriate slab 140 * manager, according to the requested buffer size, or by passes the slab 191 * Delete a buffer from the slab delayed list and put 192 * it on the slab FREE list. 198 struct pb_slab *slab = buf->slab local 302 struct pb_slab *slab; local 379 struct pb_slab *slab; local [all...] |
pb_slab.h | 31 * Helper library for carving out smaller allocations (called "(slab) entries") 55 /* Descriptor of a slab entry. 63 struct pb_slab *slab; /* the slab that contains this buffer */ member in struct:pb_slab_entry 67 /* Descriptor of a slab from which many entries are carved out. 81 /* Callback function that is called when a new slab needs to be allocated 85 * of entries. All entries that belong to the slab must be added to the free 96 /* Callback function that is called when all entries of a slab have been freed. 98 * The callback must free the slab and all its entries. It must not call any of 107 /* Manager of slab allocations. The user of this utility library should embe [all...] |
/external/llvm/unittests/Support/ |
AllocatorTest.cpp | 103 // Test allocating just over the slab size. This tests a bug where before the 108 // Fill the slab right up until the end pointer. 112 // If we don't allocate a new slab, then we will have overflowed. 117 // Test allocating with a size larger than the initial slab size. 125 // Test requesting alignment that goes past the end of the current slab. 130 // Aligning the current slab pointer is likely to move it past the end of the 131 // slab, which would confuse any unsigned comparisons with the difference of 138 // Mock slab allocator that returns slabs aligned on 4096 bytes. There is no 147 // Allocate space for the alignment, the slab, and a void* that goes right 148 // before the slab [all...] |
/external/swiftshader/third_party/LLVM/unittests/Support/ |
AllocatorTest.cpp | 82 // Test allocating just over the slab size. This tests a bug where before the 87 // Fill the slab right up until the end pointer. 91 // If we don't allocate a new slab, then we will have overflowed. 96 // Mock slab allocator that returns slabs aligned on 4096 bytes. There is no 105 // Allocate space for the alignment, the slab, and a void* that goes right 106 // before the slab. 110 // Make the slab. 111 MemSlab *Slab = (MemSlab*)(((uintptr_t)MemBase+sizeof(void*)+Alignment-1) & 113 Slab->Size = Size; 114 Slab->NextPtr = 0 [all...] |
/external/swiftshader/third_party/LLVM/include/llvm/Support/ |
Allocator.h | 51 /// MemSlab - This structure lives at the beginning of every slab allocated by 68 virtual void Deallocate(MemSlab *Slab) = 0; 71 /// MallocSlabAllocator - The default slab allocator for the bump allocator 83 virtual void Deallocate(MemSlab *Slab); 99 /// allocate a separate slab. 107 /// CurSlab - The slab that we are currently allocating into. 111 /// CurPtr - The current pointer into the current slab. This points to the 112 /// next free byte in the slab. 115 /// End - The end of the current slab. 128 /// StartNewSlab - Allocate a new slab and move the bump pointers over int [all...] |
/hardware/intel/common/libwsbm/src/ |
wsbm_slabpool.c | 277 * Add a _SlabKernelBO to the free slab manager. 307 * Get a _SlabKernelBO for us to use as storage for a slab. 449 struct _WsbmSlab *slab; local 455 slab = calloc(1, sizeof(*slab)); 456 if (!slab) 459 slab->kbo = wsbmAllocKernelBO(header); 460 if (!slab->kbo) { 465 numBuffers = slab->kbo->actualSize / header->bufSize; 471 slab->buffers = calloc(numBuffers, sizeof(*slab->buffers)) 529 struct _WsbmSlab *slab = buf->parent; local 580 struct _WsbmSlab *slab; local 655 struct _WsbmSlab *slab; local 801 struct _WsbmSlab *slab; local [all...] |
/external/mesa3d/src/util/ |
Makefile.sources | 34 slab.c \ 35 slab.h \
|
/external/ltp/testcases/kernel/fs/fs_bind/ |
TODO | 8 Check for leaks of vfsmount structs by diffing vfs slab cache obj numbers
|
/frameworks/base/cmds/incident_helper/testdata/ |
procrank_short.txt | 7 RAM: 3843972K total, 281424K free, 116764K buffers, 1777452K cached, 1136K shmem, 217916K slab
|
/external/mesa3d/src/gallium/winsys/amdgpu/drm/ |
amdgpu_bo.c | 161 assert(bo->bo && "must not be called for slab entries"); 193 assert(bo->bo); /* slab buffers have a separate vtbl */ 296 real = bo->u.slab.real; 326 real = bo->bo ? bo : bo->u.slab.real; 459 bo = container_of(entry, bo, u.slab.entry); 470 pb_slab_free(&bo->ws->bo_slabs, &bo->u.slab.entry); 483 struct amdgpu_slab *slab = CALLOC_STRUCT(amdgpu_slab); local 488 if (!slab) 509 slab->buffer = amdgpu_winsys_bo(amdgpu_bo_create(&ws->base, 512 if (!slab->buffer 556 struct amdgpu_slab *slab = amdgpu_slab(pslab); local [all...] |
amdgpu_bo.h | 55 } slab; member in union:amdgpu_winsys_bo::__anon30263 61 amdgpu_bo_handle bo; /* NULL for slab entries */ 98 void amdgpu_bo_slab_free(void *priv, struct pb_slab *slab); 107 struct amdgpu_slab *amdgpu_slab(struct pb_slab *slab) 109 return (struct amdgpu_slab *)slab;
|
/external/mesa3d/src/gallium/winsys/radeon/drm/ |
radeon_drm_bo.c | 81 for (num_idle = 0; num_idle < bo->u.slab.num_fences; ++num_idle) { 82 if (radeon_real_bo_is_busy(bo->u.slab.fences[num_idle])) { 86 radeon_bo_reference(&bo->u.slab.fences[num_idle], NULL); 88 memmove(&bo->u.slab.fences[0], &bo->u.slab.fences[num_idle], 89 (bo->u.slab.num_fences - num_idle) * sizeof(bo->u.slab.fences[0])); 90 bo->u.slab.num_fences -= num_idle; 111 while (bo->u.slab.num_fences) { 113 radeon_bo_reference(&fence, bo->u.slab.fences[0]) 742 struct radeon_slab *slab = CALLOC_STRUCT(radeon_slab); local 815 struct radeon_slab *slab = (struct radeon_slab *)pslab; local [all...] |
/external/llvm/include/llvm/Support/ |
Allocator.h | 125 /// the slab, or the next N bytes in the next slab. 128 /// size into their own slab. 141 "that objects larger than a slab go into their own memory " 187 /// \brief Deallocate all but the current slab and reset the current pointer 190 // Deallocate all but the first slab, and deallocate all custom-sized slabs. 224 // will point to the allocation of the entire slab. 231 // If Size is really big, allocate a separate slab for it. 235 // We own the new slab and don't want anyone reading anyting other than 236 // pieces returned from this method. So poison the whole slab [all...] |
/external/swiftshader/third_party/llvm-subzero/include/llvm/Support/ |
Allocator.h | 128 /// the slab, or the next N bytes in the next slab. 131 /// size into their own slab. 144 "that objects larger than a slab go into their own memory " 190 /// \brief Deallocate all but the current slab and reset the current pointer 193 // Deallocate all but the first slab, and deallocate all custom-sized slabs. 227 // will point to the allocation of the entire slab. 234 // If Size is really big, allocate a separate slab for it. 238 // We own the new slab and don't want anyone reading anyting other than 239 // pieces returned from this method. So poison the whole slab [all...] |
/prebuilts/clang/host/darwin-x86/clang-4053586/prebuilt_include/llvm/include/llvm/Support/ |
Allocator.h | 128 /// the slab, or the next N bytes in the next slab. 131 /// size into their own slab. 144 "that objects larger than a slab go into their own memory " 190 /// \brief Deallocate all but the current slab and reset the current pointer 193 // Deallocate all but the first slab, and deallocate all custom-sized slabs. 233 // will point to the allocation of the entire slab. 240 // If Size is really big, allocate a separate slab for it. 244 // We own the new slab and don't want anyone reading anyting other than 245 // pieces returned from this method. So poison the whole slab [all...] |
/prebuilts/clang/host/darwin-x86/clang-4393122/include/llvm/Support/ |
Allocator.h | 128 /// the slab, or the next N bytes in the next slab. 131 /// size into their own slab. 144 "that objects larger than a slab go into their own memory " 190 /// \brief Deallocate all but the current slab and reset the current pointer 193 // Deallocate all but the first slab, and deallocate all custom-sized slabs. 233 // will point to the allocation of the entire slab. 240 // If Size is really big, allocate a separate slab for it. 244 // We own the new slab and don't want anyone reading anyting other than 245 // pieces returned from this method. So poison the whole slab [all...] |
/prebuilts/clang/host/darwin-x86/clang-4479392/include/llvm/Support/ |
Allocator.h | 128 /// the slab, or the next N bytes in the next slab. 131 /// size into their own slab. 144 "that objects larger than a slab go into their own memory " 190 /// \brief Deallocate all but the current slab and reset the current pointer 193 // Deallocate all but the first slab, and deallocate all custom-sized slabs. 233 // will point to the allocation of the entire slab. 240 // If Size is really big, allocate a separate slab for it. 244 // We own the new slab and don't want anyone reading anyting other than 245 // pieces returned from this method. So poison the whole slab [all...] |
/prebuilts/clang/host/darwin-x86/clang-4579689/include/llvm/Support/ |
Allocator.h | 128 /// the slab, or the next N bytes in the next slab. 131 /// size into their own slab. 144 "that objects larger than a slab go into their own memory " 190 /// \brief Deallocate all but the current slab and reset the current pointer 193 // Deallocate all but the first slab, and deallocate all custom-sized slabs. 233 // will point to the allocation of the entire slab. 240 // If Size is really big, allocate a separate slab for it. 244 // We own the new slab and don't want anyone reading anyting other than 245 // pieces returned from this method. So poison the whole slab [all...] |
/prebuilts/clang/host/darwin-x86/clang-4630689/include/llvm/Support/ |
Allocator.h | 128 /// the slab, or the next N bytes in the next slab. 131 /// size into their own slab. 144 "that objects larger than a slab go into their own memory " 190 /// \brief Deallocate all but the current slab and reset the current pointer 193 // Deallocate all but the first slab, and deallocate all custom-sized slabs. 233 // will point to the allocation of the entire slab. 240 // If Size is really big, allocate a separate slab for it. 244 // We own the new slab and don't want anyone reading anyting other than 245 // pieces returned from this method. So poison the whole slab [all...] |
/prebuilts/clang/host/darwin-x86/clang-4639204/include/llvm/Support/ |
Allocator.h | 128 /// the slab, or the next N bytes in the next slab. 131 /// size into their own slab. 144 "that objects larger than a slab go into their own memory " 190 /// \brief Deallocate all but the current slab and reset the current pointer 193 // Deallocate all but the first slab, and deallocate all custom-sized slabs. 233 // will point to the allocation of the entire slab. 240 // If Size is really big, allocate a separate slab for it. 244 // We own the new slab and don't want anyone reading anyting other than 245 // pieces returned from this method. So poison the whole slab [all...] |
/prebuilts/clang/host/darwin-x86/clang-4691093/include/llvm/Support/ |
Allocator.h | 128 /// the slab, or the next N bytes in the next slab. 131 /// size into their own slab. 144 "that objects larger than a slab go into their own memory " 190 /// \brief Deallocate all but the current slab and reset the current pointer 193 // Deallocate all but the first slab, and deallocate all custom-sized slabs. 233 // will point to the allocation of the entire slab. 240 // If Size is really big, allocate a separate slab for it. 244 // We own the new slab and don't want anyone reading anyting other than 245 // pieces returned from this method. So poison the whole slab [all...] |
/prebuilts/clang/host/linux-x86/clang-4053586/prebuilt_include/llvm/include/llvm/Support/ |
Allocator.h | 128 /// the slab, or the next N bytes in the next slab. 131 /// size into their own slab. 144 "that objects larger than a slab go into their own memory " 190 /// \brief Deallocate all but the current slab and reset the current pointer 193 // Deallocate all but the first slab, and deallocate all custom-sized slabs. 233 // will point to the allocation of the entire slab. 240 // If Size is really big, allocate a separate slab for it. 244 // We own the new slab and don't want anyone reading anyting other than 245 // pieces returned from this method. So poison the whole slab [all...] |