Home | History | Annotate | Download | only in sanitizer_common

Lines Matching refs:kRegionSize

355     return (reinterpret_cast<uptr>(p) / kRegionSize) % kNumClassesRounded;
363 uptr reg_beg = (uptr)p & ~(kRegionSize - 1);
384 return reinterpret_cast<void*>(kSpaceBeg + (kRegionSize * (class_id + 1)) -
445 uptr region_beg = kSpaceBeg + class_id * kRegionSize;
460 static const uptr kRegionSize = kSpaceSize / kNumClassesRounded;
463 // kRegionSize must be >= 2^32.
464 COMPILER_CHECK((kRegionSize) >= (1ULL << (SANITIZER_WORDSIZE / 2)));
496 uptr offset = chunk % kRegionSize;
514 uptr region_beg = kSpaceBeg + kRegionSize * class_id;
534 MapWithCallback(region_beg + kRegionSize -
539 if (region->mapped_user + region->mapped_meta > kRegionSize) {
542 kRegionSize / 1024 / 1024, size);
597 // a result of a single call to MmapAlignedOrDie(kRegionSize, kRegionSize).
598 // Since the regions are aligned by kRegionSize, there are exactly
650 uptr meta = (beg + kRegionSize) - (n + 1) * kMetadataSize;
706 res += kRegionSize;
713 UnmapWithCallback((i * kRegionSize), kRegionSize);
736 uptr max_chunks_in_region = kRegionSize / (chunk_size + kMetadataSize);
737 uptr region_beg = region * kRegionSize;
754 static const uptr kRegionSize = 1 << kRegionSizeLog;
755 static const uptr kNumPossibleRegions = kSpaceSize / kRegionSize;
771 return mem & ~(kRegionSize - 1);
776 uptr res = reinterpret_cast<uptr>(MmapAlignedOrDie(kRegionSize, kRegionSize,
778 MapUnmapCallback().OnMap(res, kRegionSize);
779 stat->Add(AllocatorStatMmapped, kRegionSize);
780 CHECK_EQ(0U, (res & (kRegionSize - 1)));
794 uptr n_chunks = kRegionSize / (size + kMetadataSize);