/external/valgrind/memcheck/tests/ |
calloc-overflow.c | 13 size_t szB = 0x1000000010000001ULL; 15 size_t szB = 0x10000001UL; 17 x = calloc(szB, 0x10);
|
/external/valgrind/tests/ |
malloc.h | 14 static void* memalign16(size_t szB) 19 posix_memalign((void **)&x, 16, szB); 21 x = memalign(16, szB); 30 static void* memalign32(size_t szB) 35 posix_memalign((void **)&x, 32, szB); 37 x = memalign(32, szB); 46 static void* memalign64(size_t szB) 51 posix_memalign((void **)&x, 64, szB); 53 x = memalign(64, szB);
|
/external/valgrind/include/ |
pub_tool_xtmemory.h | 46 szB: nr of bytes for the allocated/freed block(s) 54 extern void VG_(XTMemory_Full_alloc)(SizeT szB, 56 extern void VG_(XTMemory_Full_free)(SizeT szB,
|
pub_tool_gdbserver.h | 110 extern Bool VG_(is_watched)(PointKind kind, Addr addr, Int szB); 180 If no size in the string, keeps the current value of szB. 184 set *address and *szB to 0, 188 SizeT* szB,
|
pub_tool_mallocfree.h | 52 extern void VG_(out_of_memory_NORETURN) ( const HChar* who, SizeT szB );
|
pub_tool_debuginfo.h | 208 SizeT szB; /* size in bytes */ 229 SizeT szB;
|
/external/valgrind/helgrind/ |
hg_addrdescr.h | 58 /*OUT*/SizeT* szB,
|
libhb.h | 160 Thr* thr, Addr a, SizeT szB, Bool isW ); 165 SizeT SzB, 168 /* Call fn for each recorded access history that overlaps with range [a, a+szB[. 170 void libhb_event_map_access_history ( Addr a, SizeT szB, Access_t fn );
|
/external/valgrind/exp-sgcheck/ |
sg_main.c | 174 if (fb1->szB < fb2->szB) return -1; 175 if (fb1->szB > fb2->szB) return 1; 187 /* Returns True if all fields except .szB are the same. szBs may or 229 " StackBlock{ off %ld szB %lu spRel:%c isVec:%c \"%s\" }\n", 230 sb->base, sb->szB, sb->spRel ? 'Y' : 'N', 299 /* StackBlock{ off 16 szB 66 spRel:Y isVec:Y "sz" } 300 StackBlock{ off 16 szB 130 spRel:Y isVec:Y "sz" } 301 StackBlock{ off 208 szB 16 spRel:Y isVec:Y "ar" [all...] |
h_main.c | 95 SizeT szB; /* may be zero */ 110 } else if (a < seg->addr + seg->szB && seg->szB > 0) { 115 *n = a - (seg->addr + seg->szB); 136 return seg->szB; 291 seg->szB = size; 419 if (new_size <= seg->szB) { 436 VG_(memcpy)((void*)p_new, p_old, seg->szB); 458 return ( seg ? seg->szB : 0 );
|
/external/valgrind/memcheck/ |
mc_malloc_wrappers.c | 82 MC_Chunk* create_MC_Chunk ( ThreadId tid, Addr p, SizeT szB, 106 const int l = (mc->szB >= MC_(clo_freelist_big_blocks) ? 0 : 1); 117 if (mc->szB >= MC_(clo_freelist_vol)) { 126 VG_(free_queue_volume) += (Long)mc->szB; 153 VG_(free_queue_volume) -= (Long)mc1->szB; 182 if (VG_(addr_is_in_block)( a, mc->data, mc->szB, 194 MC_Chunk* create_MC_Chunk ( ThreadId tid, Addr p, SizeT szB, 199 mc->szB = szB; 245 if (found_mc->szB != mc->sz [all...] |
mc_leakcheck.c | 275 a_hi = ((Addr)chunks[i]->data) + chunks[i]->szB; 277 a_hi++; // Special case for szB 0. See find_chunk_for. 305 a_mid_hi = chunks[mid]->data + chunks[mid]->szB; 312 if (chunks[mid]->szB == 0) 384 if (mc->szB > 1) { 385 m = find_chunk_for(mc->data + (mc->szB - 1), mallocs, n_mallocs); 552 tl_assert(ptr < ch->data + ch->szB + (ch->szB==0 ? 1 : 0)); 571 VG_(printf)("pushing %#lx-%#lx\n", ch->data, ch->data + ch->szB); 784 if (3 * sizeof(SizeT) + capacity + 1 == ch->szB [all...] |
mc_errors.c | 95 SizeT szB; // size of value in bytes 117 SizeT szB; // not used for exec (jump) errors 168 SizeT szB; // Size in bytes; 0 if unused. 336 lr->szB + lr->indirect_szB, 340 lr->szB, 359 lr->szB + lr->indirect_szB, d_bytes, 360 lr->szB, d_direct_bytes, 367 lr->szB + lr->indirect_szB ); 374 lr->szB, d_direct_bytes, 378 emit( " <leakedbytes>%lu</leakedbytes>\n", lr->szB); [all...] |
mc_include.h | 68 SizeT szB : (sizeof(SizeT)*8)-2; // Size requested; 30 or 62 bits. 173 Int MC_(get_otrack_shadow_offset) ( Int offset, Int szB ); 446 SizeT szB; // Sum of all MC_Chunk.szB values. 489 // the given range [address, address+szB[ is found. 490 void MC_(who_points_at) ( Addr address, SizeT szB); 540 void MC_(record_address_error) ( ThreadId tid, Addr a, Int szB, 543 void MC_(record_value_error) ( ThreadId tid, Int szB, UInt otag ); 551 Addr src, Addr dst, SizeT szB ); 580 Bool MC_(in_ignored_range_below_sp) ( Addr sp, Addr a, UInt szB ); [all...] |
/external/valgrind/coregrind/m_debuginfo/ |
priv_tytypes.h | 105 Int szB; 109 Int szB; 118 UWord szB; 126 Int szB;
|
readmacho.c | 98 Bool ML_(is_macho_object_file)( const void* buf, SizeT szB ) 117 if (szB < sizeof(struct fat_header)) 122 if (szB < sizeof(struct MACH_HEADER)) 225 sli.szB = arch.size; 241 if (sli.szB < sizeof(struct MACH_HEADER)) { 246 if (sli.szB > ML_(img_size)(sli.img)) { 251 if (sli.ioff >= 0 && sli.ioff + sli.szB <= ML_(img_size)(sli.img)) { 268 if (sli.szB < sizeof(struct MACH_HEADER) + mh.sizeofcmds) { 278 vg_assert(sli.szB > 0); 279 vg_assert(sli.ioff + sli.szB <= ML_(img_size)(sli.img)) [all...] |
readdwarf3.c | 209 return c->sli_next >= c->sli.ioff + c->sli.szB; 227 return c->sli.ioff + c->sli.szB - c->sli_next; 244 if (c->sli_next + sizeof(UChar) > c->sli.ioff + c->sli.szB) { 256 if (c->sli_next + sizeof(UShort) > c->sli.ioff + c->sli.szB) { 268 if (c->sli_next + sizeof(UInt) > c->sli.ioff + c->sli.szB) { 280 if (c->sli_next + sizeof(ULong) > c->sli.ioff + c->sli.szB) { 528 /* The use of escn_debug_{info,types}.szB seems safe to me even if 530 sections were not found), because DiSlice_INVALID.szB is always 533 if (die >= cc->escn_debug_info.szB) { 534 if (die >= cc->escn_debug_info.szB + cc->escn_debug_types.szB) [all...] |
priv_misc.h | 43 void* ML_(dinfo_zalloc)( const HChar* cc, SizeT szB ); 48 void ML_(dinfo_shrink_block)( void* ptr, SizeT szB );
|
misc.c | 45 void* ML_(dinfo_zalloc) ( const HChar* cc, SizeT szB ) { 47 vg_assert(szB > 0); 48 v = VG_(arena_malloc)( VG_AR_DINFO, cc, szB ); 49 VG_(memset)(v, 0, szB); 53 void ML_(dinfo_shrink_block)( void* ptr, SizeT szB ) { 54 VG_(arena_realloc_shrink)( VG_AR_DINFO, ptr, szB );
|
tytypes.c | 127 te->Te.TyBase.szB, te->Te.TyBase.enc, 132 VG_(printf)("Te_TyPtr(%d,0x%05lx)", te->Te.TyPorR.szB, 136 VG_(printf)("Te_TyRef(%d,0x%05lx)", te->Te.TyPorR.szB, 140 VG_(printf)("Te_TyMbr(%d,0x%05lx)", te->Te.TyPorR.szB, 144 VG_(printf)("Te_TyRvalRef(%d,0x%05lx)", te->Te.TyPorR.szB, 156 te->Te.TyStOrUn.szB, 169 te->Te.TyEnum.szB, te->Te.TyEnum.atomRs, 525 r = Int__cmp(te1->Te.TyBase.szB, te2->Te.TyBase.szB); 533 r = Int__cmp(te1->Te.TyPorR.szB, te2->Te.TyPorR.szB) [all...] |
/external/valgrind/coregrind/ |
m_xtmemory.c | 169 void VG_(XTMemory_Full_alloc)(SizeT szB, 172 XT_Full xtf = {szB, 1, szB, 1, 0, 0}; 175 void VG_(XTMemory_Full_free)(SizeT szB, 180 XT_Full xtf_sub = {szB, 1, 0, 0, 0, 0}; 184 XT_Full xtf_add = {0, 0, 0, 0, szB, 1};
|
/external/valgrind/coregrind/m_initimg/ |
initimg-solaris.c | 912 SizeT szB = (SizeT)VG_(client_rlimit_stack).rlim_cur; 913 if (szB < m1) 914 szB = m1; 915 if (szB > m16) 916 szB = m16; 919 szB = VG_(clo_main_stacksize); 920 if (szB < m1) 921 szB = m1; 923 szB = VG_PGROUNDUP(szB); [all...] |
/external/valgrind/massif/ |
ms_print.in | 263 my ($szB, $szB_scaled) = @_; 265 # For the label, if $szB is 999B or below, we print it as an integer. 277 if ($szB < 1000) { return sprintf("%5d", $szB); } 307 my ($szB) = @_; 311 my $szB_scaled = $szB; 325 return (max_label_2($szB, $szB_scaled), $unit); 331 my ($szB) = @_; 336 my $szB_scaled = $szB; 341 return (max_label_2($szB, $szB_scaled), $unit) [all...] |
/external/valgrind/exp-dhat/ |
dh_main.c | 620 static void* dh_malloc ( ThreadId tid, SizeT szB ) 622 return new_block( tid, NULL, szB, VG_(clo_alignment), /*is_zeroed*/False ); 625 static void* dh___builtin_new ( ThreadId tid, SizeT szB ) 627 return new_block( tid, NULL, szB, VG_(clo_alignment), /*is_zeroed*/False ); 630 static void* dh___builtin_vec_new ( ThreadId tid, SizeT szB ) 632 return new_block( tid, NULL, szB, VG_(clo_alignment), /*is_zeroed*/False ); 635 static void* dh_calloc ( ThreadId tid, SizeT m, SizeT szB ) 637 return new_block( tid, NULL, m*szB, VG_(clo_alignment), /*is_zeroed*/True ); 640 static void *dh_memalign ( ThreadId tid, SizeT alignB, SizeT szB ) 642 return new_block( tid, NULL, szB, alignB, False ) [all...] |
/external/valgrind/coregrind/m_aspacemgr/ |
aspacemgr-common.c | 423 Int szB; 430 szB = VG_STACK_GUARD_SZB 433 sres = VG_(am_mmap_anon_float_valgrind)( szB ); 439 aspacem_assert(VG_IS_PAGE_ALIGNED(szB)); 476 (ULong)(Addr)stack, szB); 483 (void)ML_(am_do_munmap_NO_NOTIFY)( (Addr)stack, szB );
|