Home | History | Annotate | Download | only in sanitizer_common

Lines Matching defs:kRegionSize

380       return ((reinterpret_cast<uptr>(p)) / kRegionSize) % kNumClassesRounded;
381 return ((reinterpret_cast<uptr>(p) - SpaceBeg()) / kRegionSize) %
390 uptr reg_beg = (uptr)p & ~(kRegionSize - 1);
412 (kRegionSize * (class_id + 1)) -
473 uptr region_beg = SpaceBeg() + class_id * kRegionSize;
493 static const uptr kRegionSize = kSpaceSize / kNumClassesRounded;
501 // kRegionSize must be >= 2^32.
502 COMPILER_CHECK((kRegionSize) >= (1ULL << (SANITIZER_WORDSIZE / 2)));
530 uptr offset = chunk % kRegionSize;
548 uptr region_beg = SpaceBeg() + kRegionSize * class_id;
568 MapWithCallback(region_beg + kRegionSize -
573 if (region->mapped_user + region->mapped_meta > kRegionSize) {
576 kRegionSize / 1024 / 1024, size);
694 // a result of a single call to MmapAlignedOrDie(kRegionSize, kRegionSize).
695 // Since the regions are aligned by kRegionSize, there are exactly
747 uptr meta = (beg + kRegionSize) - (n + 1) * kMetadataSize;
806 res += kRegionSize;
813 UnmapWithCallback((i * kRegionSize), kRegionSize);
836 uptr max_chunks_in_region = kRegionSize / (chunk_size + kMetadataSize);
837 uptr region_beg = region * kRegionSize;
858 static const uptr kRegionSize = 1 << kRegionSizeLog;
859 static const uptr kNumPossibleRegions = kSpaceSize / kRegionSize;
875 return mem & ~(kRegionSize - 1);
880 uptr res = reinterpret_cast<uptr>(MmapAlignedOrDie(kRegionSize, kRegionSize,
882 MapUnmapCallback().OnMap(res, kRegionSize);
883 stat->Add(AllocatorStatMapped, kRegionSize);
884 CHECK_EQ(0U, (res & (kRegionSize - 1)));
898 uptr n_chunks = kRegionSize / (size + kMetadataSize);