HomeSort by relevance Sort by last modified time
    Searched refs:kAlignment (Results 1 - 25 of 35) sorted by null

1 2

  /art/runtime/gc/accounting/
space_bitmap.cc 31 template<size_t kAlignment>
32 size_t SpaceBitmap<kAlignment>::ComputeBitmapSize(uint64_t capacity) {
33 const uint64_t kBytesCoveredPerWord = kAlignment * kBitsPerIntPtrT;
37 template<size_t kAlignment>
38 size_t SpaceBitmap<kAlignment>::ComputeHeapSize(uint64_t bitmap_bytes) {
39 return bitmap_bytes * kBitsPerByte * kAlignment;
42 template<size_t kAlignment>
43 SpaceBitmap<kAlignment>* SpaceBitmap<kAlignment>::CreateFromMemMap(
51 template<size_t kAlignment>
    [all...]
bitmap.cc 78 template<size_t kAlignment>
79 MemoryRangeBitmap<kAlignment>* MemoryRangeBitmap<kAlignment>::Create(
81 CHECK_ALIGNED(cover_begin, kAlignment);
82 CHECK_ALIGNED(cover_end, kAlignment);
83 const size_t num_bits = (cover_end - cover_begin) / kAlignment;
88 template<size_t kAlignment>
89 MemoryRangeBitmap<kAlignment>* MemoryRangeBitmap<kAlignment>::CreateFromMemMap(
space_bitmap-inl.h 32 template<size_t kAlignment>
33 inline bool SpaceBitmap<kAlignment>::AtomicTestAndSet(const mirror::Object* obj) {
54 template<size_t kAlignment>
55 inline bool SpaceBitmap<kAlignment>::Test(const mirror::Object* obj) const {
64 template<size_t kAlignment> template<typename Visitor>
65 inline void SpaceBitmap<kAlignment>::VisitMarkedRange(uintptr_t visit_begin, uintptr_t visit_end,
69 for (uintptr_t i = visit_begin; i < visit_end; i += kAlignment) {
85 const size_t bit_start = (offset_start / kAlignment) % kBitsPerIntPtrT;
86 const size_t bit_end = (offset_end / kAlignment) % kBitsPerIntPtrT;
111 mirror::Object* obj = reinterpret_cast<mirror::Object*>(ptr_base + shift * kAlignment);
    [all...]
space_bitmap.h 41 template<size_t kAlignment>
48 // heap_begin of heap_capacity bytes, where objects are guaranteed to be kAlignment-aligned.
63 return offset / kAlignment / kBitsPerIntPtrT;
68 return static_cast<T>(index * kAlignment * kBitsPerIntPtrT);
73 return (static_cast<size_t>(1)) << ((offset / kAlignment) % kBitsPerIntPtrT);
119 for (; visit_begin < visit_end; visit_begin += kAlignment) {
210 static void WalkInstanceFields(SpaceBitmap<kAlignment>* visited, ObjectCallback* callback,
234 template<size_t kAlignment>
235 std::ostream& operator << (std::ostream& stream, const SpaceBitmap<kAlignment>& bitmap);
space_bitmap_test.cc 113 template <size_t kAlignment>
127 size_t offset = RoundDown(r.next() % heap_capacity, kAlignment);
141 size_t offset = RoundDown(r.next() % heap_capacity, kAlignment);
143 size_t end = offset + RoundDown(r.next() % (remain + 1), kAlignment);
149 for (uintptr_t k = offset; k < end; k += kAlignment) {
bitmap.h 127 // One bit per kAlignment in range (start, end]
128 template<size_t kAlignment>
148 const uintptr_t addr = CoverBegin() + bit_index * kAlignment;
156 return (addr - CoverBegin()) / kAlignment;
182 : Bitmap(mem_map, num_bits), cover_begin_(begin), cover_end_(begin + kAlignment * num_bits) {
card_table.h 43 template<size_t kAlignment> class SpaceBitmap;
  /external/skia/src/gpu/
GrMemoryPool.h 86 kAlignment = 8,
87 kHeaderSize = GR_CT_ALIGN_UP(sizeof(BlockHeader), kAlignment),
88 kPerAllocPad = GR_CT_ALIGN_UP(sizeof(AllocHeader), kAlignment),
GrMemoryPool.cpp 21 fMinAllocSize = GrSizeAlignUp(minAllocSize + kPerAllocPad, kAlignment),
22 fPreallocSize = GrSizeAlignUp(preallocSize + kPerAllocPad, kAlignment);
44 size = GrSizeAlignUp(size, kAlignment);
123 SkASSERT(!(reinterpret_cast<intptr_t>(block) % kAlignment));
159 SkASSERT(!(b % kAlignment));
160 SkASSERT(!(totalSize % kAlignment));
161 SkASSERT(!(userSize % kAlignment));
162 SkASSERT(!(block->fCurrPtr % kAlignment));
  /external/webrtc/talk/session/media/
planarfunctions_unittest.cc 64 static const int kAlignment = 16;
165 uint8_t* image_pointer = new uint8_t[y_size + u_size + v_size + kAlignment];
166 y_pointer = ALIGNP(image_pointer, kAlignment);
167 u_pointer = ALIGNP(&image_pointer[y_size], kAlignment);
168 v_pointer = ALIGNP(&image_pointer[y_size + u_size], kAlignment);
207 uint8_t* image_pointer = new uint8_t[2 * height * awidth + kAlignment];
208 yuv_pointer = ALIGNP(image_pointer, kAlignment);
287 ((height + 1) / 2) * ((width + 1) / 2) * 2 + kAlignment];
288 y_pointer = ALIGNP(image_pointer, kAlignment);
326 ((height + 1) / 2) * ((width + 1) / 2) * 2 + kAlignment];
    [all...]
yuvscaler_unittest.cc 53 static const int kAlignment = 16;
105 new uint8_t[isize + kAlignment + memoffset]());
107 new uint8_t[osize + kAlignment + memoffset]());
109 new uint8_t[osize + kAlignment + memoffset]());
111 uint8_t* ibuf = ALIGNP(ibuffer.get(), kAlignment) + memoffset;
112 uint8_t* obuf = ALIGNP(obuffer.get(), kAlignment) + memoffset;
113 uint8_t* xbuf = ALIGNP(xbuffer.get(), kAlignment) + memoffset;
212 scoped_ptr<uint8_t[]> ibuffer(new uint8_t[I420_SIZE(iw, ih) + kAlignment]);
213 scoped_ptr<uint8_t[]> obuffer(new uint8_t[I420_SIZE(ow, oh) + kAlignment]);
215 uint8_t* ibuf = ALIGNP(ibuffer.get(), kAlignment);
    [all...]
  /art/runtime/gc/space/
bump_pointer_space-inl.h 30 num_bytes = RoundUp(num_bytes, kAlignment);
47 num_bytes = RoundUp(num_bytes, kAlignment);
66 DCHECK_ALIGNED(num_bytes, kAlignment);
93 *usable_size = RoundUp(num_bytes, kAlignment);
region_space-inl.h 29 num_bytes = RoundUp(num_bytes, kAlignment);
46 DCHECK_ALIGNED(num_bytes, kAlignment);
118 DCHECK_ALIGNED(num_bytes, kAlignment);
146 *usable_size = RoundUp(num_bytes, kAlignment);
261 return reinterpret_cast<mirror::Object*>(RoundUp(position, kAlignment));
268 DCHECK_ALIGNED(num_bytes, kAlignment);
large_object_space.cc 266 return AlignSize() * FreeListSpace::kAlignment;
271 DCHECK_ALIGNED(size, FreeListSpace::kAlignment);
272 alloc_size_ = (size / FreeListSpace::kAlignment) | (free ? kFlagFree : 0u);
305 // Return how many kAlignment units there are before the free block.
311 return GetPrevFree() * FreeListSpace::kAlignment;
315 DCHECK_ALIGNED(bytes, FreeListSpace::kAlignment);
316 prev_free_ = bytes / FreeListSpace::kAlignment;
323 // Contains the size of the previous free block with kAlignment as the unit. If 0 then the
327 // Allocation size of this object in kAlignment as the unit.
355 CHECK_EQ(size % kAlignment, 0U)
    [all...]
bump_pointer_space.h 160 static constexpr size_t kAlignment = 8;
188 size_t unused_; // Ensures alignment of kAlignment.
191 static_assert(sizeof(BlockHeader) % kAlignment == 0,
192 "continuous block must be kAlignment aligned");
large_object_space.h 162 static constexpr size_t kAlignment = kPageSize;
179 return (address - reinterpret_cast<uintptr_t>(Begin())) / kAlignment;
185 return reinterpret_cast<uintptr_t>(Begin()) + slot * kAlignment;
  /external/gemmlowp/test/
test_allocator.cc 32 !(reinterpret_cast<std::uintptr_t>(int32_array) % Allocator::kAlignment));
34 !(reinterpret_cast<std::uintptr_t>(int8_array) % Allocator::kAlignment));
  /external/v8/src/
zone.cc 94 size = RoundUp(size, kAlignment);
98 if (kPointerSize == 4 && kAlignment == 4) {
101 DCHECK(kAlignment >= kPointerSize);
119 DCHECK(IsAddressAligned(result, kAlignment, 0));
160 position_ = RoundUp(start, kAlignment);
224 DCHECK_EQ(size, RoundDown(size, kAlignment));
233 static const size_t kSegmentOverhead = sizeof(Segment) + kAlignment;
262 Address result = RoundUp(segment->start(), kAlignment);
zone.h 72 static const size_t kAlignment = 8;
75 static const size_t kAlignment = kPointerSize;
113 // is guaranteed to be aligned as dictated by kAlignment.
  /external/gemmlowp/internal/
allocator.h 97 static const std::size_t kAlignment = kDefaultCacheLineSize;
110 storage_ = memalign(kAlignment, storage_size_);
112 if (posix_memalign(&storage_, kAlignment, storage_size_)) {
152 const std::size_t bytes = RoundUp<kAlignment>(n * sizeof(T));
  /art/runtime/gc/
heap-inl.h 74 byte_count = RoundUp(byte_count, space::BumpPointerSpace::kAlignment);
252 alloc_size = RoundUp(alloc_size, space::BumpPointerSpace::kAlignment);
314 DCHECK_ALIGNED(alloc_size, space::BumpPointerSpace::kAlignment);
338 alloc_size = RoundUp(alloc_size, space::RegionSpace::kAlignment);
345 DCHECK_ALIGNED(alloc_size, space::RegionSpace::kAlignment);
  /external/webrtc/talk/media/base/
videoframe_unittest.h 69 static const int kAlignment = 16;
576 rtc::scoped_ptr<uint8_t[]> buf(new uint8_t[buf_size + kAlignment]);
577 uint8_t* y = ALIGNP(buf.get(), kAlignment);
597 rtc::scoped_ptr<uint8_t[]> buf(new uint8_t[buf_size + kAlignment]);
598 uint8_t* yuy2 = ALIGNP(buf.get(), kAlignment);
614 rtc::scoped_ptr<uint8_t[]> buf(new uint8_t[buf_size + kAlignment + 1]);
615 uint8_t* yuy2 = ALIGNP(buf.get(), kAlignment) + 1;
771 rtc::scoped_ptr<uint8_t[]> outbuf(new uint8_t[out_size + kAlignment]);
772 uint8_t* out = ALIGNP(outbuf.get(), kAlignment);
787 rtc::scoped_ptr<uint8_t[]> outbuf(new uint8_t[out_size + kAlignment]);
    [all...]
  /external/boringssl/src/tool/
speed.cc 177 static const unsigned kAlignment = 16;
188 std::unique_ptr<uint8_t[]> in_storage(new uint8_t[chunk_len + kAlignment]);
189 std::unique_ptr<uint8_t[]> out_storage(new uint8_t[chunk_len + overhead_len + kAlignment]);
193 uint8_t *const in = align(in_storage.get(), kAlignment);
195 uint8_t *const out = align(out_storage.get(), kAlignment);
  /art/runtime/entrypoints/quick/
quick_alloc_entrypoints.cc 39 byte_count = RoundUp(byte_count, gc::space::BumpPointerSpace::kAlignment); \
65 byte_count = RoundUp(byte_count, gc::space::BumpPointerSpace::kAlignment); \
90 byte_count = RoundUp(byte_count, gc::space::BumpPointerSpace::kAlignment); \
  /art/runtime/base/
arena_allocator.h 307 bytes = RoundUp(bytes, kAlignment);
369 static constexpr size_t kAlignment = 8;

Completed in 526 milliseconds

1 2