/external/skia/tests/ |
GrAllocatorTest.cpp | 73 // Test combinations of allocators with and without stack storage and with different block 75 SkTArray<GrTAllocator<C>*> allocators; local 77 allocators.push_back(&a1); 79 allocators.push_back(&a2); 81 allocators.push_back(&a5); 84 allocators.push_back(&a1); 86 allocators.push_back(&sa3); 88 allocators.push_back(&sa4); 90 for (int i = 0; i < allocators.count(); ++i) { 91 check_allocator(allocators[i], 0, 0, reporter) [all...] |
/external/skqp/tests/ |
GrAllocatorTest.cpp | 73 // Test combinations of allocators with and without stack storage and with different block 75 SkTArray<GrTAllocator<C>*> allocators; local 77 allocators.push_back(&a1); 79 allocators.push_back(&a2); 81 allocators.push_back(&a5); 84 allocators.push_back(&a1); 86 allocators.push_back(&sa3); 88 allocators.push_back(&sa4); 90 for (int i = 0; i < allocators.count(); ++i) { 91 check_allocator(allocators[i], 0, 0, reporter) [all...] |
/device/google/cuttlefish_common/host/commands/launch/ |
vsoc_shared_memory.cc | 109 const std::vector<VSoCRegionAllocator>& allocators, 118 header->region_count = allocators.size(); 121 for (size_t idx = 0; idx < allocators.size(); ++idx) { 122 region_idx_by_name[allocators[idx].region_layout->region_name()] = idx; 128 for (size_t idx = 0; idx < allocators.size(); ++idx) { 130 const auto& region = *allocators[idx].region_layout; 131 WriteRegionDescription(shmem_region_desc, allocators[idx]); 166 std::vector<VSoCRegionAllocator> allocators; local 171 allocators.emplace_back(*layout, file_size /* offset */, 175 file_size += allocators.back().region_size [all...] |
/external/libcxx/test/support/ |
debug_mode_helper.h | 372 CHECKPOINT("testing swap with non-equal allocators");
|
/external/python/cpython3/Modules/ |
_tracemalloc.c | 30 } allocators; variable in typeref:struct:__anon37322 297 return allocators.raw.malloc(allocators.raw.ctx, size); 303 allocators.raw.free(allocators.raw.ctx, ptr); 968 PyMem_GetAllocator(PYMEM_DOMAIN_RAW, &allocators.raw); 1103 alloc.ctx = &allocators.raw; 1104 PyMem_GetAllocator(PYMEM_DOMAIN_RAW, &allocators.raw); 1113 alloc.ctx = &allocators.mem; 1114 PyMem_GetAllocator(PYMEM_DOMAIN_MEM, &allocators.mem) [all...] |
/bionic/libc/bionic/ |
bionic_allocator.cpp | 257 BionicSmallObjectAllocator* allocators = local 262 new (allocators + i) BionicSmallObjectAllocator(type, 1 << type); 265 allocators_ = allocators;
|
/external/libxcam/wrapper/gstreamer/ |
gstxcambufferpool.cpp | 39 #include <gst/allocators/gstdmabuf.h>
|
gstxcamfilter.cpp | 25 #include <gst/allocators/gstdmabuf.h> [all...] |
/external/tensorflow/tensorflow/contrib/verbs/ |
README.md | 30 3. Following HKUST research on the use of GPU direct, and their [GDR implementation](https://github.com/tensorflow/tensorflow/blob/master/tensorflow/contrib/gdr/README.md), there is a smart way to benefit from the TensorFlow allocation theme which is mostly pool based, i.e allocators pre-allocate a large memory block, and allocate the tensors from there. By attaching a custom Visitor to relevant allocators, we can do a single registration of the entire memory block, which zeros the registration overhead. Once the block is registered, each new tensor allocated will be at a registered address, which will allow us to do direct RDMA writes to it.
|
/cts/tests/tests/graphics/src/android/graphics/cts/ |
ImageDecoderTest.java | 335 private static final int[] ALLOCATORS = new int[] { 349 for (int allocator : ALLOCATORS) { 383 for (int allocator : ALLOCATORS) { 1083 int allocators[] = new int[] { ImageDecoder.ALLOCATOR_DEFAULT, local 1797 int[] allocators = { ImageDecoder.ALLOCATOR_HARDWARE, ImageDecoder.ALLOCATOR_DEFAULT }; local [all...] |