Lines Matching full:uptr
97 uptr buffer_size;
112 uptr size;
142 static inline bool CanBeAHeapPointer(uptr p) {
145 const uptr kMinAddress = 4 * 4096;
161 void ScanRangeForPointers(uptr begin, uptr end,
164 const uptr alignment = flags()->pointer_alignment();
166 uptr pp = begin;
171 if (!CanBeAHeapPointer(reinterpret_cast<uptr>(p))) continue;
172 uptr chunk = PointsIntoChunk(p);
198 void ForEachExtraStackRangeCb(uptr begin, uptr end, void* arg) {
206 InternalScopedBuffer<uptr> registers(SuspendedThreadsList::RegisterCount());
207 uptr registers_begin = reinterpret_cast<uptr>(registers.data());
208 uptr registers_end = registers_begin + registers.size();
209 for (uptr i = 0; i < suspended_threads.thread_count(); i++) {
210 uptr os_id = static_cast<uptr>(suspended_threads.GetThreadID(i));
212 uptr stack_begin, stack_end, tls_begin, tls_end, cache_begin, cache_end;
222 uptr sp;
270 static void ProcessRootRegion(Frontier *frontier, uptr root_begin,
271 uptr root_end) {
273 uptr begin, end, prot;
277 uptr intersection_begin = Max(root_begin, begin);
278 uptr intersection_end = Min(end, root_end);
294 for (uptr i = 0; i < root_regions->size(); i++) {
296 uptr begin_addr = reinterpret_cast<uptr>(region.begin);
303 uptr next_chunk = frontier->back();
313 static void MarkIndirectlyLeakedCb(uptr chunk, void *arg) {
324 static void CollectIgnoredCb(uptr chunk, void *arg) {
361 uptr size = 0;
362 const uptr *trace = StackDepotGet(stack_trace_id, &size);
368 static void CollectLeaksCb(uptr chunk, void *arg) {
375 uptr resolution = flags()->resolution;
378 uptr size = 0;
379 const uptr *trace = StackDepotGet(m.stack_trace_id(), &size);
399 for (uptr i = 0; i < matched.size(); i++)
400 Printf("%7zu %10zu %s\n", static_cast<uptr>(matched[i]->hit_count),
442 uptr unsuppressed_count = param.leak_report.UnsuppressedLeakCount();
462 static Suppression *GetSuppressionForAddr(uptr addr) {
467 uptr module_offset;
474 static const uptr kMaxAddrFrames = 16;
476 for (uptr i = 0; i < kMaxAddrFrames; i++) new (&addr_frames[i]) AddressInfo();
477 uptr addr_frames_num = Symbolizer::Get()->SymbolizePC(
479 for (uptr i = 0; i < addr_frames_num; i++) {
488 uptr size = 0;
489 const uptr *trace = StackDepotGet(stack_trace_id, &size);
490 for (uptr i = 0; i < size; i++) {
505 const uptr kMaxLeaksConsidered = 5000;
507 void LeakReport::AddLeakedChunk(uptr chunk, u32 stack_trace_id,
508 uptr leaked_size, ChunkTag tag) {
511 uptr i;
539 void LeakReport::ReportTopLeaks(uptr num_leaks_to_report) {
547 uptr unsuppressed_count = UnsuppressedLeakCount();
551 uptr leaks_reported = 0;
552 for (uptr i = 0; i < leaks_.size(); i++) {
559 uptr remaining = unsuppressed_count - leaks_reported;
564 void LeakReport::PrintReportForLeak(uptr index) {
581 void LeakReport::PrintLeakedObjectsForLeak(uptr index) {
583 for (uptr j = 0; j < leaked_objects_.size(); j++) {
592 uptr bytes = 0, allocations = 0;
593 for (uptr i = 0; i < leaks_.size(); i++) {
606 for (uptr i = 0; i < leaks_.size(); i++) {
616 uptr LeakReport::UnsuppressedLeakCount() {
617 uptr result = 0;
618 for (uptr i = 0; i < leaks_.size(); i++)
649 void __lsan_register_root_region(const void *begin, uptr size) {
660 void __lsan_unregister_root_region(const void *begin, uptr size) {
665 for (uptr i = 0; i < root_regions->size(); i++) {
669 uptr last_index = root_regions->size() - 1;