/art/runtime/gc/accounting/ |
space_bitmap.cc | 31 template<size_t kAlignment> 32 size_t SpaceBitmap<kAlignment>::ComputeBitmapSize(uint64_t capacity) { 33 const uint64_t kBytesCoveredPerWord = kAlignment * kBitsPerIntPtrT; 37 template<size_t kAlignment> 38 size_t SpaceBitmap<kAlignment>::ComputeHeapSize(uint64_t bitmap_bytes) { 39 return bitmap_bytes * kBitsPerByte * kAlignment; 42 template<size_t kAlignment> 43 SpaceBitmap<kAlignment>* SpaceBitmap<kAlignment>::CreateFromMemMap( 51 template<size_t kAlignment> [all...] |
bitmap.cc | 78 template<size_t kAlignment> 79 MemoryRangeBitmap<kAlignment>* MemoryRangeBitmap<kAlignment>::Create( 81 CHECK_ALIGNED(cover_begin, kAlignment); 82 CHECK_ALIGNED(cover_end, kAlignment); 83 const size_t num_bits = (cover_end - cover_begin) / kAlignment; 88 template<size_t kAlignment> 89 MemoryRangeBitmap<kAlignment>* MemoryRangeBitmap<kAlignment>::CreateFromMemMap(
|
space_bitmap-inl.h | 32 template<size_t kAlignment> 33 inline bool SpaceBitmap<kAlignment>::AtomicTestAndSet(const mirror::Object* obj) { 54 template<size_t kAlignment> 55 inline bool SpaceBitmap<kAlignment>::Test(const mirror::Object* obj) const { 64 template<size_t kAlignment> template<typename Visitor> 65 inline void SpaceBitmap<kAlignment>::VisitMarkedRange(uintptr_t visit_begin, uintptr_t visit_end, 69 for (uintptr_t i = visit_begin; i < visit_end; i += kAlignment) { 85 const size_t bit_start = (offset_start / kAlignment) % kBitsPerIntPtrT; 86 const size_t bit_end = (offset_end / kAlignment) % kBitsPerIntPtrT; 111 mirror::Object* obj = reinterpret_cast<mirror::Object*>(ptr_base + shift * kAlignment); [all...] |
space_bitmap.h | 41 template<size_t kAlignment> 48 // heap_begin of heap_capacity bytes, where objects are guaranteed to be kAlignment-aligned. 63 return offset / kAlignment / kBitsPerIntPtrT; 68 return static_cast<T>(index * kAlignment * kBitsPerIntPtrT); 73 return (static_cast<size_t>(1)) << ((offset / kAlignment) % kBitsPerIntPtrT); 119 for (; visit_begin < visit_end; visit_begin += kAlignment) { 210 static void WalkInstanceFields(SpaceBitmap<kAlignment>* visited, ObjectCallback* callback, 234 template<size_t kAlignment> 235 std::ostream& operator << (std::ostream& stream, const SpaceBitmap<kAlignment>& bitmap);
|
space_bitmap_test.cc | 113 template <size_t kAlignment> 127 size_t offset = RoundDown(r.next() % heap_capacity, kAlignment); 141 size_t offset = RoundDown(r.next() % heap_capacity, kAlignment); 143 size_t end = offset + RoundDown(r.next() % (remain + 1), kAlignment); 149 for (uintptr_t k = offset; k < end; k += kAlignment) {
|
bitmap.h | 127 // One bit per kAlignment in range (start, end] 128 template<size_t kAlignment> 148 const uintptr_t addr = CoverBegin() + bit_index * kAlignment; 156 return (addr - CoverBegin()) / kAlignment; 182 : Bitmap(mem_map, num_bits), cover_begin_(begin), cover_end_(begin + kAlignment * num_bits) {
|
card_table.h | 43 template<size_t kAlignment> class SpaceBitmap;
|
/external/skia/src/gpu/ |
GrMemoryPool.h | 86 kAlignment = 8, 87 kHeaderSize = GR_CT_ALIGN_UP(sizeof(BlockHeader), kAlignment), 88 kPerAllocPad = GR_CT_ALIGN_UP(sizeof(AllocHeader), kAlignment),
|
GrMemoryPool.cpp | 21 fMinAllocSize = GrSizeAlignUp(minAllocSize + kPerAllocPad, kAlignment), 22 fPreallocSize = GrSizeAlignUp(preallocSize + kPerAllocPad, kAlignment); 44 size = GrSizeAlignUp(size, kAlignment); 123 SkASSERT(!(reinterpret_cast<intptr_t>(block) % kAlignment)); 159 SkASSERT(!(b % kAlignment)); 160 SkASSERT(!(totalSize % kAlignment)); 161 SkASSERT(!(userSize % kAlignment)); 162 SkASSERT(!(block->fCurrPtr % kAlignment));
|
/external/webrtc/talk/session/media/ |
planarfunctions_unittest.cc | 64 static const int kAlignment = 16; 165 uint8_t* image_pointer = new uint8_t[y_size + u_size + v_size + kAlignment]; 166 y_pointer = ALIGNP(image_pointer, kAlignment); 167 u_pointer = ALIGNP(&image_pointer[y_size], kAlignment); 168 v_pointer = ALIGNP(&image_pointer[y_size + u_size], kAlignment); 207 uint8_t* image_pointer = new uint8_t[2 * height * awidth + kAlignment]; 208 yuv_pointer = ALIGNP(image_pointer, kAlignment); 287 ((height + 1) / 2) * ((width + 1) / 2) * 2 + kAlignment]; 288 y_pointer = ALIGNP(image_pointer, kAlignment); 326 ((height + 1) / 2) * ((width + 1) / 2) * 2 + kAlignment]; [all...] |
yuvscaler_unittest.cc | 53 static const int kAlignment = 16; 105 new uint8_t[isize + kAlignment + memoffset]()); 107 new uint8_t[osize + kAlignment + memoffset]()); 109 new uint8_t[osize + kAlignment + memoffset]()); 111 uint8_t* ibuf = ALIGNP(ibuffer.get(), kAlignment) + memoffset; 112 uint8_t* obuf = ALIGNP(obuffer.get(), kAlignment) + memoffset; 113 uint8_t* xbuf = ALIGNP(xbuffer.get(), kAlignment) + memoffset; 212 scoped_ptr<uint8_t[]> ibuffer(new uint8_t[I420_SIZE(iw, ih) + kAlignment]); 213 scoped_ptr<uint8_t[]> obuffer(new uint8_t[I420_SIZE(ow, oh) + kAlignment]); 215 uint8_t* ibuf = ALIGNP(ibuffer.get(), kAlignment); [all...] |
/art/runtime/gc/space/ |
bump_pointer_space-inl.h | 30 num_bytes = RoundUp(num_bytes, kAlignment); 47 num_bytes = RoundUp(num_bytes, kAlignment); 66 DCHECK_ALIGNED(num_bytes, kAlignment); 93 *usable_size = RoundUp(num_bytes, kAlignment);
|
region_space-inl.h | 29 num_bytes = RoundUp(num_bytes, kAlignment); 46 DCHECK_ALIGNED(num_bytes, kAlignment); 118 DCHECK_ALIGNED(num_bytes, kAlignment); 146 *usable_size = RoundUp(num_bytes, kAlignment); 261 return reinterpret_cast<mirror::Object*>(RoundUp(position, kAlignment)); 268 DCHECK_ALIGNED(num_bytes, kAlignment);
|
large_object_space.cc | 266 return AlignSize() * FreeListSpace::kAlignment; 271 DCHECK_ALIGNED(size, FreeListSpace::kAlignment); 272 alloc_size_ = (size / FreeListSpace::kAlignment) | (free ? kFlagFree : 0u); 305 // Return how many kAlignment units there are before the free block. 311 return GetPrevFree() * FreeListSpace::kAlignment; 315 DCHECK_ALIGNED(bytes, FreeListSpace::kAlignment); 316 prev_free_ = bytes / FreeListSpace::kAlignment; 323 // Contains the size of the previous free block with kAlignment as the unit. If 0 then the 327 // Allocation size of this object in kAlignment as the unit. 355 CHECK_EQ(size % kAlignment, 0U) [all...] |
bump_pointer_space.h | 160 static constexpr size_t kAlignment = 8; 188 size_t unused_; // Ensures alignment of kAlignment. 191 static_assert(sizeof(BlockHeader) % kAlignment == 0, 192 "continuous block must be kAlignment aligned");
|
large_object_space.h | 162 static constexpr size_t kAlignment = kPageSize; 179 return (address - reinterpret_cast<uintptr_t>(Begin())) / kAlignment; 185 return reinterpret_cast<uintptr_t>(Begin()) + slot * kAlignment;
|
/external/gemmlowp/test/ |
test_allocator.cc | 32 !(reinterpret_cast<std::uintptr_t>(int32_array) % Allocator::kAlignment)); 34 !(reinterpret_cast<std::uintptr_t>(int8_array) % Allocator::kAlignment));
|
/external/v8/src/ |
zone.cc | 94 size = RoundUp(size, kAlignment); 98 if (kPointerSize == 4 && kAlignment == 4) { 101 DCHECK(kAlignment >= kPointerSize); 119 DCHECK(IsAddressAligned(result, kAlignment, 0)); 160 position_ = RoundUp(start, kAlignment); 224 DCHECK_EQ(size, RoundDown(size, kAlignment)); 233 static const size_t kSegmentOverhead = sizeof(Segment) + kAlignment; 262 Address result = RoundUp(segment->start(), kAlignment);
|
zone.h | 72 static const size_t kAlignment = 8; 75 static const size_t kAlignment = kPointerSize; 113 // is guaranteed to be aligned as dictated by kAlignment.
|
/external/gemmlowp/internal/ |
allocator.h | 97 static const std::size_t kAlignment = kDefaultCacheLineSize; 110 storage_ = memalign(kAlignment, storage_size_); 112 if (posix_memalign(&storage_, kAlignment, storage_size_)) { 152 const std::size_t bytes = RoundUp<kAlignment>(n * sizeof(T));
|
/art/runtime/gc/ |
heap-inl.h | 74 byte_count = RoundUp(byte_count, space::BumpPointerSpace::kAlignment); 252 alloc_size = RoundUp(alloc_size, space::BumpPointerSpace::kAlignment); 314 DCHECK_ALIGNED(alloc_size, space::BumpPointerSpace::kAlignment); 338 alloc_size = RoundUp(alloc_size, space::RegionSpace::kAlignment); 345 DCHECK_ALIGNED(alloc_size, space::RegionSpace::kAlignment);
|
/external/webrtc/talk/media/base/ |
videoframe_unittest.h | 69 static const int kAlignment = 16; 576 rtc::scoped_ptr<uint8_t[]> buf(new uint8_t[buf_size + kAlignment]); 577 uint8_t* y = ALIGNP(buf.get(), kAlignment); 597 rtc::scoped_ptr<uint8_t[]> buf(new uint8_t[buf_size + kAlignment]); 598 uint8_t* yuy2 = ALIGNP(buf.get(), kAlignment); 614 rtc::scoped_ptr<uint8_t[]> buf(new uint8_t[buf_size + kAlignment + 1]); 615 uint8_t* yuy2 = ALIGNP(buf.get(), kAlignment) + 1; 771 rtc::scoped_ptr<uint8_t[]> outbuf(new uint8_t[out_size + kAlignment]); 772 uint8_t* out = ALIGNP(outbuf.get(), kAlignment); 787 rtc::scoped_ptr<uint8_t[]> outbuf(new uint8_t[out_size + kAlignment]); [all...] |
/external/boringssl/src/tool/ |
speed.cc | 177 static const unsigned kAlignment = 16; 188 std::unique_ptr<uint8_t[]> in_storage(new uint8_t[chunk_len + kAlignment]); 189 std::unique_ptr<uint8_t[]> out_storage(new uint8_t[chunk_len + overhead_len + kAlignment]); 193 uint8_t *const in = align(in_storage.get(), kAlignment); 195 uint8_t *const out = align(out_storage.get(), kAlignment);
|
/art/runtime/entrypoints/quick/ |
quick_alloc_entrypoints.cc | 39 byte_count = RoundUp(byte_count, gc::space::BumpPointerSpace::kAlignment); \ 65 byte_count = RoundUp(byte_count, gc::space::BumpPointerSpace::kAlignment); \ 90 byte_count = RoundUp(byte_count, gc::space::BumpPointerSpace::kAlignment); \
|
/art/runtime/base/ |
arena_allocator.h | 307 bytes = RoundUp(bytes, kAlignment); 369 static constexpr size_t kAlignment = 8;
|