HomeSort by relevance Sort by last modified time
    Searched full:slab (Results 1 - 25 of 238) sorted by null

1 2 3 4 5 6 7 8 910

  /external/mesa3d/src/gallium/drivers/nouveau/
nouveau_mm.c 53 mm_slab_alloc(struct mm_slab *slab)
57 if (slab->free == 0)
60 for (i = 0; i < (slab->count + 31) / 32; ++i) {
61 b = ffs(slab->bits[i]) - 1;
64 assert(n < slab->count);
65 slab->free--;
66 slab->bits[i] &= ~(1 << b);
74 mm_slab_free(struct mm_slab *slab, int i)
76 assert(i < slab->count);
77 slab->bits[i / 32] |= 1 << (i % 32)
123 struct mm_slab *slab; local
168 struct mm_slab *slab; local
219 struct mm_slab *slab = (struct mm_slab *)alloc->priv; local
269 struct mm_slab *slab, *next; local
    [all...]
  /external/swiftshader/third_party/LLVM/lib/Support/
Allocator.cpp 44 /// StartNewSlab - Allocate a new slab and move the bump pointers over into
45 /// the new slab. Modifies CurPtr and End.
48 // to allocate more. Increase slab size to reduce mallocs and possibly memory
62 void BumpPtrAllocator::DeallocateSlabs(MemSlab *Slab) {
63 while (Slab) {
64 MemSlab *NextSlab = Slab->NextPtr;
68 sys::Memory::setRangeWritable(Slab + 1, Slab->Size - sizeof(MemSlab));
69 memset(Slab + 1, 0xCD, Slab->Size - sizeof(MemSlab))
    [all...]
  /external/mesa3d/src/gallium/auxiliary/pipebuffer/
pb_slab.c 33 /* All slab allocations from the same heap and with the same size belong
43 * The intention is to keep serving allocations from the same slab as long
46 * Due to a race in new slab allocation, additional slabs in this list
56 struct pb_slab *slab = entry->slab; local
59 LIST_ADD(&entry->head, &slab->free);
60 slab->num_free++;
62 /* Add slab to the group's list if it isn't already linked. */
63 if (!slab->head.next) {
65 LIST_ADDTAIL(&slab->head, &group->slabs)
103 struct pb_slab *slab; local
    [all...]
pb_bufmgr_slab.c 55 * Buffer in a slab.
63 struct pb_slab *slab; member in struct:pb_slab_buffer
69 /** Offset relative to the start of the slab buffer. */
79 * Slab -- a contiguous piece of memory.
116 * Alignment, usage to be used to allocate the slab buffers.
139 * This buffer manager just dispatches buffer allocations to the appropriate slab
140 * manager, according to the requested buffer size, or by passes the slab
191 * Delete a buffer from the slab delayed list and put
192 * it on the slab FREE list.
198 struct pb_slab *slab = buf->slab local
302 struct pb_slab *slab; local
379 struct pb_slab *slab; local
    [all...]
pb_slab.h 31 * Helper library for carving out smaller allocations (called "(slab) entries")
55 /* Descriptor of a slab entry.
63 struct pb_slab *slab; /* the slab that contains this buffer */ member in struct:pb_slab_entry
67 /* Descriptor of a slab from which many entries are carved out.
81 /* Callback function that is called when a new slab needs to be allocated
85 * of entries. All entries that belong to the slab must be added to the free
96 /* Callback function that is called when all entries of a slab have been freed.
98 * The callback must free the slab and all its entries. It must not call any of
107 /* Manager of slab allocations. The user of this utility library should embe
    [all...]
  /external/llvm/unittests/Support/
AllocatorTest.cpp 103 // Test allocating just over the slab size. This tests a bug where before the
108 // Fill the slab right up until the end pointer.
112 // If we don't allocate a new slab, then we will have overflowed.
117 // Test allocating with a size larger than the initial slab size.
125 // Test requesting alignment that goes past the end of the current slab.
130 // Aligning the current slab pointer is likely to move it past the end of the
131 // slab, which would confuse any unsigned comparisons with the difference of
138 // Mock slab allocator that returns slabs aligned on 4096 bytes. There is no
147 // Allocate space for the alignment, the slab, and a void* that goes right
148 // before the slab
    [all...]
  /external/swiftshader/third_party/LLVM/unittests/Support/
AllocatorTest.cpp 82 // Test allocating just over the slab size. This tests a bug where before the
87 // Fill the slab right up until the end pointer.
91 // If we don't allocate a new slab, then we will have overflowed.
96 // Mock slab allocator that returns slabs aligned on 4096 bytes. There is no
105 // Allocate space for the alignment, the slab, and a void* that goes right
106 // before the slab.
110 // Make the slab.
111 MemSlab *Slab = (MemSlab*)(((uintptr_t)MemBase+sizeof(void*)+Alignment-1) &
113 Slab->Size = Size;
114 Slab->NextPtr = 0
    [all...]
  /external/swiftshader/third_party/LLVM/include/llvm/Support/
Allocator.h 51 /// MemSlab - This structure lives at the beginning of every slab allocated by
68 virtual void Deallocate(MemSlab *Slab) = 0;
71 /// MallocSlabAllocator - The default slab allocator for the bump allocator
83 virtual void Deallocate(MemSlab *Slab);
99 /// allocate a separate slab.
107 /// CurSlab - The slab that we are currently allocating into.
111 /// CurPtr - The current pointer into the current slab. This points to the
112 /// next free byte in the slab.
115 /// End - The end of the current slab.
128 /// StartNewSlab - Allocate a new slab and move the bump pointers over int
    [all...]
  /hardware/intel/common/libwsbm/src/
wsbm_slabpool.c 277 * Add a _SlabKernelBO to the free slab manager.
307 * Get a _SlabKernelBO for us to use as storage for a slab.
449 struct _WsbmSlab *slab; local
455 slab = calloc(1, sizeof(*slab));
456 if (!slab)
459 slab->kbo = wsbmAllocKernelBO(header);
460 if (!slab->kbo) {
465 numBuffers = slab->kbo->actualSize / header->bufSize;
471 slab->buffers = calloc(numBuffers, sizeof(*slab->buffers))
529 struct _WsbmSlab *slab = buf->parent; local
580 struct _WsbmSlab *slab; local
655 struct _WsbmSlab *slab; local
801 struct _WsbmSlab *slab; local
    [all...]
  /external/mesa3d/src/util/
Makefile.sources 34 slab.c \
35 slab.h \
  /external/ltp/testcases/kernel/fs/fs_bind/
TODO 8 Check for leaks of vfsmount structs by diffing vfs slab cache obj numbers
  /frameworks/base/cmds/incident_helper/testdata/
procrank_short.txt 7 RAM: 3843972K total, 281424K free, 116764K buffers, 1777452K cached, 1136K shmem, 217916K slab
  /external/mesa3d/src/gallium/winsys/amdgpu/drm/
amdgpu_bo.c 161 assert(bo->bo && "must not be called for slab entries");
193 assert(bo->bo); /* slab buffers have a separate vtbl */
296 real = bo->u.slab.real;
326 real = bo->bo ? bo : bo->u.slab.real;
459 bo = container_of(entry, bo, u.slab.entry);
470 pb_slab_free(&bo->ws->bo_slabs, &bo->u.slab.entry);
483 struct amdgpu_slab *slab = CALLOC_STRUCT(amdgpu_slab); local
488 if (!slab)
509 slab->buffer = amdgpu_winsys_bo(amdgpu_bo_create(&ws->base,
512 if (!slab->buffer
556 struct amdgpu_slab *slab = amdgpu_slab(pslab); local
    [all...]
amdgpu_bo.h 55 } slab; member in union:amdgpu_winsys_bo::__anon30263
61 amdgpu_bo_handle bo; /* NULL for slab entries */
98 void amdgpu_bo_slab_free(void *priv, struct pb_slab *slab);
107 struct amdgpu_slab *amdgpu_slab(struct pb_slab *slab)
109 return (struct amdgpu_slab *)slab;
  /external/mesa3d/src/gallium/winsys/radeon/drm/
radeon_drm_bo.c 81 for (num_idle = 0; num_idle < bo->u.slab.num_fences; ++num_idle) {
82 if (radeon_real_bo_is_busy(bo->u.slab.fences[num_idle])) {
86 radeon_bo_reference(&bo->u.slab.fences[num_idle], NULL);
88 memmove(&bo->u.slab.fences[0], &bo->u.slab.fences[num_idle],
89 (bo->u.slab.num_fences - num_idle) * sizeof(bo->u.slab.fences[0]));
90 bo->u.slab.num_fences -= num_idle;
111 while (bo->u.slab.num_fences) {
113 radeon_bo_reference(&fence, bo->u.slab.fences[0])
742 struct radeon_slab *slab = CALLOC_STRUCT(radeon_slab); local
815 struct radeon_slab *slab = (struct radeon_slab *)pslab; local
    [all...]
  /external/llvm/include/llvm/Support/
Allocator.h 125 /// the slab, or the next N bytes in the next slab.
128 /// size into their own slab.
141 "that objects larger than a slab go into their own memory "
187 /// \brief Deallocate all but the current slab and reset the current pointer
190 // Deallocate all but the first slab, and deallocate all custom-sized slabs.
224 // will point to the allocation of the entire slab.
231 // If Size is really big, allocate a separate slab for it.
235 // We own the new slab and don't want anyone reading anyting other than
236 // pieces returned from this method. So poison the whole slab
    [all...]
  /external/swiftshader/third_party/llvm-subzero/include/llvm/Support/
Allocator.h 128 /// the slab, or the next N bytes in the next slab.
131 /// size into their own slab.
144 "that objects larger than a slab go into their own memory "
190 /// \brief Deallocate all but the current slab and reset the current pointer
193 // Deallocate all but the first slab, and deallocate all custom-sized slabs.
227 // will point to the allocation of the entire slab.
234 // If Size is really big, allocate a separate slab for it.
238 // We own the new slab and don't want anyone reading anyting other than
239 // pieces returned from this method. So poison the whole slab
    [all...]
  /prebuilts/clang/host/darwin-x86/clang-4053586/prebuilt_include/llvm/include/llvm/Support/
Allocator.h 128 /// the slab, or the next N bytes in the next slab.
131 /// size into their own slab.
144 "that objects larger than a slab go into their own memory "
190 /// \brief Deallocate all but the current slab and reset the current pointer
193 // Deallocate all but the first slab, and deallocate all custom-sized slabs.
233 // will point to the allocation of the entire slab.
240 // If Size is really big, allocate a separate slab for it.
244 // We own the new slab and don't want anyone reading anyting other than
245 // pieces returned from this method. So poison the whole slab
    [all...]
  /prebuilts/clang/host/darwin-x86/clang-4393122/include/llvm/Support/
Allocator.h 128 /// the slab, or the next N bytes in the next slab.
131 /// size into their own slab.
144 "that objects larger than a slab go into their own memory "
190 /// \brief Deallocate all but the current slab and reset the current pointer
193 // Deallocate all but the first slab, and deallocate all custom-sized slabs.
233 // will point to the allocation of the entire slab.
240 // If Size is really big, allocate a separate slab for it.
244 // We own the new slab and don't want anyone reading anyting other than
245 // pieces returned from this method. So poison the whole slab
    [all...]
  /prebuilts/clang/host/darwin-x86/clang-4479392/include/llvm/Support/
Allocator.h 128 /// the slab, or the next N bytes in the next slab.
131 /// size into their own slab.
144 "that objects larger than a slab go into their own memory "
190 /// \brief Deallocate all but the current slab and reset the current pointer
193 // Deallocate all but the first slab, and deallocate all custom-sized slabs.
233 // will point to the allocation of the entire slab.
240 // If Size is really big, allocate a separate slab for it.
244 // We own the new slab and don't want anyone reading anyting other than
245 // pieces returned from this method. So poison the whole slab
    [all...]
  /prebuilts/clang/host/darwin-x86/clang-4579689/include/llvm/Support/
Allocator.h 128 /// the slab, or the next N bytes in the next slab.
131 /// size into their own slab.
144 "that objects larger than a slab go into their own memory "
190 /// \brief Deallocate all but the current slab and reset the current pointer
193 // Deallocate all but the first slab, and deallocate all custom-sized slabs.
233 // will point to the allocation of the entire slab.
240 // If Size is really big, allocate a separate slab for it.
244 // We own the new slab and don't want anyone reading anyting other than
245 // pieces returned from this method. So poison the whole slab
    [all...]
  /prebuilts/clang/host/darwin-x86/clang-4630689/include/llvm/Support/
Allocator.h 128 /// the slab, or the next N bytes in the next slab.
131 /// size into their own slab.
144 "that objects larger than a slab go into their own memory "
190 /// \brief Deallocate all but the current slab and reset the current pointer
193 // Deallocate all but the first slab, and deallocate all custom-sized slabs.
233 // will point to the allocation of the entire slab.
240 // If Size is really big, allocate a separate slab for it.
244 // We own the new slab and don't want anyone reading anyting other than
245 // pieces returned from this method. So poison the whole slab
    [all...]
  /prebuilts/clang/host/darwin-x86/clang-4639204/include/llvm/Support/
Allocator.h 128 /// the slab, or the next N bytes in the next slab.
131 /// size into their own slab.
144 "that objects larger than a slab go into their own memory "
190 /// \brief Deallocate all but the current slab and reset the current pointer
193 // Deallocate all but the first slab, and deallocate all custom-sized slabs.
233 // will point to the allocation of the entire slab.
240 // If Size is really big, allocate a separate slab for it.
244 // We own the new slab and don't want anyone reading anyting other than
245 // pieces returned from this method. So poison the whole slab
    [all...]
  /prebuilts/clang/host/darwin-x86/clang-4691093/include/llvm/Support/
Allocator.h 128 /// the slab, or the next N bytes in the next slab.
131 /// size into their own slab.
144 "that objects larger than a slab go into their own memory "
190 /// \brief Deallocate all but the current slab and reset the current pointer
193 // Deallocate all but the first slab, and deallocate all custom-sized slabs.
233 // will point to the allocation of the entire slab.
240 // If Size is really big, allocate a separate slab for it.
244 // We own the new slab and don't want anyone reading anyting other than
245 // pieces returned from this method. So poison the whole slab
    [all...]
  /prebuilts/clang/host/linux-x86/clang-4053586/prebuilt_include/llvm/include/llvm/Support/
Allocator.h 128 /// the slab, or the next N bytes in the next slab.
131 /// size into their own slab.
144 "that objects larger than a slab go into their own memory "
190 /// \brief Deallocate all but the current slab and reset the current pointer
193 // Deallocate all but the first slab, and deallocate all custom-sized slabs.
233 // will point to the allocation of the entire slab.
240 // If Size is really big, allocate a separate slab for it.
244 // We own the new slab and don't want anyone reading anyting other than
245 // pieces returned from this method. So poison the whole slab
    [all...]

Completed in 1953 milliseconds

1 2 3 4 5 6 7 8 910