Home | History | Annotate | Download | only in exp-dhat
      1 
      2 //--------------------------------------------------------------------*/
      3 //--- DHAT: a Dynamic Heap Analysis Tool                 dh_main.c ---*/
      4 //--------------------------------------------------------------------*/
      5 
      6 /*
      7    This file is part of DHAT, a Valgrind tool for profiling the
      8    heap usage of programs.
      9 
     10    Copyright (C) 2010-2011 Mozilla Inc
     11 
     12    This program is free software; you can redistribute it and/or
     13    modify it under the terms of the GNU General Public License as
     14    published by the Free Software Foundation; either version 2 of the
     15    License, or (at your option) any later version.
     16 
     17    This program is distributed in the hope that it will be useful, but
     18    WITHOUT ANY WARRANTY; without even the implied warranty of
     19    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
     20    General Public License for more details.
     21 
     22    You should have received a copy of the GNU General Public License
     23    along with this program; if not, write to the Free Software
     24    Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA
     25    02111-1307, USA.
     26 
     27    The GNU General Public License is contained in the file COPYING.
     28 */
     29 
     30 /* Contributed by Julian Seward <jseward (at) acm.org> */
     31 
     32 
     33 #include "pub_tool_basics.h"
     34 #include "pub_tool_libcbase.h"
     35 #include "pub_tool_libcassert.h"
     36 #include "pub_tool_libcprint.h"
     37 #include "pub_tool_machine.h"      // VG_(fnptr_to_fnentry)
     38 #include "pub_tool_mallocfree.h"
     39 #include "pub_tool_options.h"
     40 #include "pub_tool_replacemalloc.h"
     41 #include "pub_tool_tooliface.h"
     42 #include "pub_tool_wordfm.h"
     43 
     44 #define HISTOGRAM_SIZE_LIMIT 1024
     45 
     46 
     47 //------------------------------------------------------------//
     48 //--- Globals                                              ---//
     49 //------------------------------------------------------------//
     50 
     51 // Number of guest instructions executed so far.  This is
     52 // incremented directly from the generated code.
     53 static ULong g_guest_instrs_executed = 0;
     54 
     55 // Summary statistics for the entire run.
     56 static ULong g_tot_blocks = 0;   // total blocks allocated
     57 static ULong g_tot_bytes  = 0;   // total bytes allocated
     58 
     59 static ULong g_cur_blocks_live = 0; // curr # blocks live
     60 static ULong g_cur_bytes_live  = 0; // curr # bytes live
     61 
     62 static ULong g_max_blocks_live = 0; // bytes and blocks at
     63 static ULong g_max_bytes_live  = 0; // the max residency point
     64 
     65 
     66 //------------------------------------------------------------//
     67 //--- an Interval Tree of live blocks                      ---//
     68 //------------------------------------------------------------//
     69 
     70 /* Tracks information about live blocks. */
     71 typedef
     72    struct {
     73       Addr        payload;
     74       SizeT       req_szB;
     75       ExeContext* ap;  /* allocation ec */
     76       ULong       allocd_at; /* instruction number */
     77       ULong       n_reads;
     78       ULong       n_writes;
     79       /* Approx histogram, one byte per payload byte.  Counts latch up
     80          therefore at 0xFFFF.  Can be NULL if the block is resized or if
     81          the block is larger than HISTOGRAM_SIZE_LIMIT. */
     82       UShort*     histoW; /* [0 .. req_szB-1] */
     83    }
     84    Block;
     85 
     86 /* May not contain zero-sized blocks.  May not contain
     87    overlapping blocks. */
     88 static WordFM* interval_tree = NULL;  /* WordFM* Block* void */
     89 
     90 /* Here's the comparison function.  Since the tree is required
     91 to contain non-zero sized, non-overlapping blocks, it's good
     92 enough to consider any overlap as a match. */
     93 static Word interval_tree_Cmp ( UWord k1, UWord k2 )
     94 {
     95    Block* b1 = (Block*)k1;
     96    Block* b2 = (Block*)k2;
     97    tl_assert(b1->req_szB > 0);
     98    tl_assert(b2->req_szB > 0);
     99    if (b1->payload + b1->req_szB <= b2->payload) return -1;
    100    if (b2->payload + b2->req_szB <= b1->payload) return  1;
    101    return 0;
    102 }
    103 
    104 // 2-entry cache for find_Block_containing
    105 static Block* fbc_cache0 = NULL;
    106 static Block* fbc_cache1 = NULL;
    107 
    108 static UWord stats__n_fBc_cached = 0;
    109 static UWord stats__n_fBc_uncached = 0;
    110 static UWord stats__n_fBc_notfound = 0;
    111 
    112 static Block* find_Block_containing ( Addr a )
    113 {
    114    if (LIKELY(fbc_cache0
    115               && fbc_cache0->payload <= a
    116               && a < fbc_cache0->payload + fbc_cache0->req_szB)) {
    117       // found at 0
    118       stats__n_fBc_cached++;
    119       return fbc_cache0;
    120    }
    121    if (LIKELY(fbc_cache1
    122               && fbc_cache1->payload <= a
    123               && a < fbc_cache1->payload + fbc_cache1->req_szB)) {
    124       // found at 1; swap 0 and 1
    125       Block* tmp = fbc_cache0;
    126       fbc_cache0 = fbc_cache1;
    127       fbc_cache1 = tmp;
    128       stats__n_fBc_cached++;
    129       return fbc_cache0;
    130    }
    131    Block fake;
    132    fake.payload = a;
    133    fake.req_szB = 1;
    134    UWord foundkey = 1;
    135    UWord foundval = 1;
    136    Bool found = VG_(lookupFM)( interval_tree,
    137                                &foundkey, &foundval, (UWord)&fake );
    138    if (!found) {
    139       stats__n_fBc_notfound++;
    140       return NULL;
    141    }
    142    tl_assert(foundval == 0); // we don't store vals in the interval tree
    143    tl_assert(foundkey != 1);
    144    Block* res = (Block*)foundkey;
    145    tl_assert(res != &fake);
    146    // put at the top position
    147    fbc_cache1 = fbc_cache0;
    148    fbc_cache0 = res;
    149    stats__n_fBc_uncached++;
    150    return res;
    151 }
    152 
    153 // delete a block; asserts if not found.  (viz, 'a' must be
    154 // known to be present.)
    155 static void delete_Block_starting_at ( Addr a )
    156 {
    157    Block fake;
    158    fake.payload = a;
    159    fake.req_szB = 1;
    160    Bool found = VG_(delFromFM)( interval_tree,
    161                                 NULL, NULL, (Addr)&fake );
    162    tl_assert(found);
    163    fbc_cache0 = fbc_cache1 = NULL;
    164 }
    165 
    166 
    167 //------------------------------------------------------------//
    168 //--- a FM of allocation points (APs)                      ---//
    169 //------------------------------------------------------------//
    170 
    171 typedef
    172    struct {
    173       // the allocation point that we're summarising stats for
    174       ExeContext* ap;
    175       // used when printing results
    176       Bool shown;
    177       // The current number of blocks and bytes live for this AP
    178       ULong cur_blocks_live;
    179       ULong cur_bytes_live;
    180       // The number of blocks and bytes live at the max-liveness
    181       // point.  Note this is a bit subtle.  max_blocks_live is not
    182       // the maximum number of live blocks, but rather the number of
    183       // blocks live at the point of maximum byte liveness.  These are
    184       // not necessarily the same thing.
    185       ULong max_blocks_live;
    186       ULong max_bytes_live;
    187       // Total number of blocks and bytes allocated by this AP.
    188       ULong tot_blocks;
    189       ULong tot_bytes;
    190       // Sum of death ages for all blocks allocated by this AP,
    191       // that have subsequently been freed.
    192       ULong death_ages_sum;
    193       ULong deaths;
    194       // Total number of reads and writes in all blocks allocated
    195       // by this AP.
    196       ULong n_reads;
    197       ULong n_writes;
    198       /* Histogram information.  We maintain a histogram aggregated for
    199          all retiring Blocks allocated by this AP, but only if:
    200          - this AP has only ever allocated objects of one size
    201          - that size is <= HISTOGRAM_SIZE_LIMIT
    202          What we need therefore is a mechanism to see if this AP
    203          has only ever allocated blocks of one size.
    204 
    205          3 states:
    206             Unknown          because no retirement yet
    207             Exactly xsize    all retiring blocks are of this size
    208             Mixed            multiple different sizes seen
    209       */
    210       enum { Unknown=999, Exactly, Mixed } xsize_tag;
    211       SizeT xsize;
    212       UInt* histo; /* [0 .. xsize-1] */
    213    }
    214    APInfo;
    215 
    216 /* maps ExeContext*'s to APInfo*'s.  Note that the keys must match the
    217    .ap field in the values. */
    218 static WordFM* apinfo = NULL;  /* WordFM* ExeContext* APInfo* */
    219 
    220 
    221 /* 'bk' is being introduced (has just been allocated).  Find the
    222    relevant APInfo entry for it, or create one, based on the block's
    223    allocation EC.  Then, update the APInfo to the extent that we
    224    actually can, to reflect the allocation. */
    225 static void intro_Block ( Block* bk )
    226 {
    227    tl_assert(bk);
    228    tl_assert(bk->ap);
    229 
    230    APInfo* api   = NULL;
    231    UWord   keyW  = 0;
    232    UWord   valW  = 0;
    233    Bool    found = VG_(lookupFM)( apinfo,
    234                                   &keyW, &valW, (UWord)bk->ap );
    235    if (found) {
    236       api = (APInfo*)valW;
    237       tl_assert(keyW == (UWord)bk->ap);
    238    } else {
    239       api = VG_(malloc)( "dh.main.intro_Block.1", sizeof(APInfo) );
    240       VG_(memset)(api, 0, sizeof(*api));
    241       api->ap = bk->ap;
    242       Bool present = VG_(addToFM)( apinfo,
    243                                    (UWord)bk->ap, (UWord)api );
    244       tl_assert(!present);
    245       // histo stuff
    246       tl_assert(api->deaths == 0);
    247       api->xsize_tag = Unknown;
    248       api->xsize = 0;
    249       if (0) VG_(printf)("api %p   -->  Unknown\n", api);
    250    }
    251 
    252    tl_assert(api->ap == bk->ap);
    253 
    254    /* So: update stats to reflect an allocation */
    255 
    256    // # live blocks
    257    api->cur_blocks_live++;
    258 
    259    // # live bytes
    260    api->cur_bytes_live += bk->req_szB;
    261    if (api->cur_bytes_live > api->max_bytes_live) {
    262       api->max_bytes_live  = api->cur_bytes_live;
    263       api->max_blocks_live = api->cur_blocks_live;
    264    }
    265 
    266    // total blocks and bytes allocated here
    267    api->tot_blocks++;
    268    api->tot_bytes += bk->req_szB;
    269 
    270    // update summary globals
    271    g_tot_blocks++;
    272    g_tot_bytes += bk->req_szB;
    273 
    274    g_cur_blocks_live++;
    275    g_cur_bytes_live += bk->req_szB;
    276    if (g_cur_bytes_live > g_max_bytes_live) {
    277       g_max_bytes_live = g_cur_bytes_live;
    278       g_max_blocks_live = g_cur_blocks_live;
    279    }
    280 }
    281 
    282 
    283 /* 'bk' is retiring (being freed).  Find the relevant APInfo entry for
    284    it, which must already exist.  Then, fold info from 'bk' into that
    285    entry.  'because_freed' is True if the block is retiring because
    286    the client has freed it.  If it is False then the block is retiring
    287    because the program has finished, in which case we want to skip the
    288    updates of the total blocks live etc for this AP, but still fold in
    289    the access counts and histo data that have so far accumulated for
    290    the block. */
    291 static void retire_Block ( Block* bk, Bool because_freed )
    292 {
    293    tl_assert(bk);
    294    tl_assert(bk->ap);
    295 
    296    APInfo* api   = NULL;
    297    UWord   keyW  = 0;
    298    UWord   valW  = 0;
    299    Bool    found = VG_(lookupFM)( apinfo,
    300                                   &keyW, &valW, (UWord)bk->ap );
    301 
    302    tl_assert(found);
    303    api = (APInfo*)valW;
    304    tl_assert(api->ap == bk->ap);
    305 
    306    // update stats following this free.
    307    if (0)
    308    VG_(printf)("ec %p  api->c_by_l %llu  bk->rszB %llu\n",
    309                bk->ap, api->cur_bytes_live, (ULong)bk->req_szB);
    310 
    311    // update total blocks live etc for this AP
    312    if (because_freed) {
    313       tl_assert(api->cur_blocks_live >= 1);
    314       tl_assert(api->cur_bytes_live >= bk->req_szB);
    315       api->cur_blocks_live--;
    316       api->cur_bytes_live -= bk->req_szB;
    317 
    318       api->deaths++;
    319 
    320       tl_assert(bk->allocd_at <= g_guest_instrs_executed);
    321       api->death_ages_sum += (g_guest_instrs_executed - bk->allocd_at);
    322 
    323       // update global summary stats
    324       tl_assert(g_cur_blocks_live > 0);
    325       g_cur_blocks_live--;
    326       tl_assert(g_cur_bytes_live >= bk->req_szB);
    327       g_cur_bytes_live -= bk->req_szB;
    328    }
    329 
    330    // access counts
    331    api->n_reads  += bk->n_reads;
    332    api->n_writes += bk->n_writes;
    333 
    334    // histo stuff.  First, do state transitions for xsize/xsize_tag.
    335    switch (api->xsize_tag) {
    336 
    337       case Unknown:
    338          tl_assert(api->xsize == 0);
    339          tl_assert(api->deaths == 1 || api->deaths == 0);
    340          tl_assert(!api->histo);
    341          api->xsize_tag = Exactly;
    342          api->xsize = bk->req_szB;
    343          if (0) VG_(printf)("api %p   -->  Exactly(%lu)\n", api, api->xsize);
    344          // and allocate the histo
    345          if (bk->histoW) {
    346             api->histo = VG_(malloc)("dh.main.retire_Block.1",
    347                                      api->xsize * sizeof(UInt));
    348             VG_(memset)(api->histo, 0, api->xsize * sizeof(UInt));
    349          }
    350          break;
    351 
    352       case Exactly:
    353          //tl_assert(api->deaths > 1);
    354          if (bk->req_szB != api->xsize) {
    355             if (0) VG_(printf)("api %p   -->  Mixed(%lu -> %lu)\n",
    356                                api, api->xsize, bk->req_szB);
    357             api->xsize_tag = Mixed;
    358             api->xsize = 0;
    359             // deallocate the histo, if any
    360             if (api->histo) {
    361                VG_(free)(api->histo);
    362                api->histo = NULL;
    363             }
    364          }
    365          break;
    366 
    367       case Mixed:
    368          //tl_assert(api->deaths > 1);
    369          break;
    370 
    371       default:
    372         tl_assert(0);
    373    }
    374 
    375    // See if we can fold the histo data from this block into
    376    // the data for the AP
    377    if (api->xsize_tag == Exactly && api->histo && bk->histoW) {
    378       tl_assert(api->xsize == bk->req_szB);
    379       UWord i;
    380       for (i = 0; i < api->xsize; i++) {
    381          // FIXME: do something better in case of overflow of api->histo[..]
    382          // Right now, at least don't let it overflow/wrap around
    383          if (api->histo[i] <= 0xFFFE0000)
    384             api->histo[i] += (UInt)bk->histoW[i];
    385       }
    386       if (0) VG_(printf)("fold in, AP = %p\n", api);
    387    }
    388 
    389 
    390 
    391 #if 0
    392    if (bk->histoB) {
    393       VG_(printf)("block retiring, histo %lu: ", bk->req_szB);
    394       UWord i;
    395       for (i = 0; i < bk->req_szB; i++)
    396         VG_(printf)("%u ", (UInt)bk->histoB[i]);
    397       VG_(printf)("\n");
    398    } else {
    399       VG_(printf)("block retiring, no histo %lu\n", bk->req_szB);
    400    }
    401 #endif
    402 }
    403 
    404 /* This handles block resizing.  When a block with AP 'ec' has a
    405    size change of 'delta', call here to update the APInfo. */
    406 static void apinfo_change_cur_bytes_live( ExeContext* ec, Long delta )
    407 {
    408    APInfo* api   = NULL;
    409    UWord   keyW  = 0;
    410    UWord   valW  = 0;
    411    Bool    found = VG_(lookupFM)( apinfo,
    412                                   &keyW, &valW, (UWord)ec );
    413 
    414    tl_assert(found);
    415    api = (APInfo*)valW;
    416    tl_assert(api->ap == ec);
    417 
    418    if (delta < 0) {
    419       tl_assert(api->cur_bytes_live >= -delta);
    420       tl_assert(g_cur_bytes_live >= -delta);
    421    }
    422 
    423    // adjust current live size
    424    api->cur_bytes_live += delta;
    425    g_cur_bytes_live += delta;
    426 
    427    if (delta > 0 && api->cur_bytes_live > api->max_bytes_live) {
    428       api->max_bytes_live  = api->cur_bytes_live;
    429       api->max_blocks_live = api->cur_blocks_live;
    430    }
    431 
    432    // update global summary stats
    433    if (delta > 0 && g_cur_bytes_live > g_max_bytes_live) {
    434       g_max_bytes_live = g_cur_bytes_live;
    435       g_max_blocks_live = g_cur_blocks_live;
    436    }
    437    if (delta > 0)
    438       g_tot_bytes += delta;
    439 
    440    // adjust total allocation size
    441    if (delta > 0)
    442       api->tot_bytes += delta;
    443 }
    444 
    445 
    446 //------------------------------------------------------------//
    447 //--- update both Block and APInfos after {m,re}alloc/free ---//
    448 //------------------------------------------------------------//
    449 
    450 static
    451 void* new_block ( ThreadId tid, void* p, SizeT req_szB, SizeT req_alignB,
    452                   Bool is_zeroed )
    453 {
    454    tl_assert(p == NULL); // don't handle custom allocators right now
    455    SizeT actual_szB /*, slop_szB*/;
    456 
    457    if ((SSizeT)req_szB < 0) return NULL;
    458 
    459    if (req_szB == 0)
    460       req_szB = 1;  /* can't allow zero-sized blocks in the interval tree */
    461 
    462    // Allocate and zero if necessary
    463    if (!p) {
    464       p = VG_(cli_malloc)( req_alignB, req_szB );
    465       if (!p) {
    466          return NULL;
    467       }
    468       if (is_zeroed) VG_(memset)(p, 0, req_szB);
    469       actual_szB = VG_(malloc_usable_size)(p);
    470       tl_assert(actual_szB >= req_szB);
    471       /* slop_szB = actual_szB - req_szB; */
    472    } else {
    473       /* slop_szB = 0; */
    474    }
    475 
    476    // Make new HP_Chunk node, add to malloc_list
    477    Block* bk = VG_(malloc)("dh.new_block.1", sizeof(Block));
    478    bk->payload   = (Addr)p;
    479    bk->req_szB   = req_szB;
    480    bk->ap        = VG_(record_ExeContext)(tid, 0/*first word delta*/);
    481    bk->allocd_at = g_guest_instrs_executed;
    482    bk->n_reads   = 0;
    483    bk->n_writes  = 0;
    484    // set up histogram array, if the block isn't too large
    485    bk->histoW = NULL;
    486    if (req_szB <= HISTOGRAM_SIZE_LIMIT) {
    487       bk->histoW = VG_(malloc)("dh.new_block.2", req_szB * sizeof(UShort));
    488       VG_(memset)(bk->histoW, 0, req_szB * sizeof(UShort));
    489    }
    490 
    491    Bool present = VG_(addToFM)( interval_tree, (UWord)bk, (UWord)0/*no val*/);
    492    tl_assert(!present);
    493    fbc_cache0 = fbc_cache1 = NULL;
    494 
    495    intro_Block(bk);
    496 
    497    if (0) VG_(printf)("ALLOC %ld -> %p\n", req_szB, p);
    498 
    499    return p;
    500 }
    501 
    502 static
    503 void die_block ( void* p, Bool custom_free )
    504 {
    505    tl_assert(!custom_free);  // at least for now
    506 
    507    Block* bk = find_Block_containing( (Addr)p );
    508 
    509    if (!bk) {
    510      return; // bogus free
    511    }
    512 
    513    tl_assert(bk->req_szB > 0);
    514    // assert the block finder is behaving sanely
    515    tl_assert(bk->payload <= (Addr)p);
    516    tl_assert( (Addr)p < bk->payload + bk->req_szB );
    517 
    518    if (bk->payload != (Addr)p) {
    519       return; // bogus free
    520    }
    521 
    522    if (0) VG_(printf)(" FREE %p %llu\n",
    523                       p, g_guest_instrs_executed - bk->allocd_at);
    524 
    525    retire_Block(bk, True/*because_freed*/);
    526 
    527    VG_(cli_free)( (void*)bk->payload );
    528    delete_Block_starting_at( bk->payload );
    529    if (bk->histoW) {
    530       VG_(free)( bk->histoW );
    531       bk->histoW = NULL;
    532    }
    533    VG_(free)( bk );
    534 }
    535 
    536 
    537 static
    538 void* renew_block ( ThreadId tid, void* p_old, SizeT new_req_szB )
    539 {
    540    if (0) VG_(printf)("REALL %p %ld\n", p_old, new_req_szB);
    541    void* p_new = NULL;
    542 
    543    tl_assert(new_req_szB > 0); // map 0 to 1
    544 
    545    // Find the old block.
    546    Block* bk = find_Block_containing( (Addr)p_old );
    547    if (!bk) {
    548       return NULL;   // bogus realloc
    549    }
    550 
    551    tl_assert(bk->req_szB > 0);
    552    // assert the block finder is behaving sanely
    553    tl_assert(bk->payload <= (Addr)p_old);
    554    tl_assert( (Addr)p_old < bk->payload + bk->req_szB );
    555 
    556    if (bk->payload != (Addr)p_old) {
    557       return NULL; // bogus realloc
    558    }
    559 
    560    // Keeping the histogram alive in any meaningful way across
    561    // block resizing is too darn complicated.  Just throw it away.
    562    if (bk->histoW) {
    563       VG_(free)(bk->histoW);
    564       bk->histoW = NULL;
    565    }
    566 
    567    // Actually do the allocation, if necessary.
    568    if (new_req_szB <= bk->req_szB) {
    569 
    570       // New size is smaller or same; block not moved.
    571       apinfo_change_cur_bytes_live(bk->ap,
    572                                    (Long)new_req_szB - (Long)bk->req_szB);
    573       bk->req_szB = new_req_szB;
    574       return p_old;
    575 
    576    } else {
    577 
    578       // New size is bigger;  make new block, copy shared contents, free old.
    579       p_new = VG_(cli_malloc)(VG_(clo_alignment), new_req_szB);
    580       if (!p_new) {
    581          // Nb: if realloc fails, NULL is returned but the old block is not
    582          // touched.  What an awful function.
    583          return NULL;
    584       }
    585       tl_assert(p_new != p_old);
    586 
    587       VG_(memcpy)(p_new, p_old, bk->req_szB);
    588       VG_(cli_free)(p_old);
    589 
    590       // Since the block has moved, we need to re-insert it into the
    591       // interval tree at the new place.  Do this by removing
    592       // and re-adding it.
    593       delete_Block_starting_at( (Addr)p_old );
    594       // now 'bk' is no longer in the tree, but the Block itself
    595       // is still alive
    596 
    597       // Update the metadata.
    598       apinfo_change_cur_bytes_live(bk->ap,
    599                                    (Long)new_req_szB - (Long)bk->req_szB);
    600       bk->payload = (Addr)p_new;
    601       bk->req_szB = new_req_szB;
    602 
    603       // and re-add
    604       Bool present
    605          = VG_(addToFM)( interval_tree, (UWord)bk, (UWord)0/*no val*/);
    606       tl_assert(!present);
    607       fbc_cache0 = fbc_cache1 = NULL;
    608 
    609       return p_new;
    610    }
    611    /*NOTREACHED*/
    612    tl_assert(0);
    613 }
    614 
    615 
    616 //------------------------------------------------------------//
    617 //--- malloc() et al replacement wrappers                  ---//
    618 //------------------------------------------------------------//
    619 
    620 static void* dh_malloc ( ThreadId tid, SizeT szB )
    621 {
    622    return new_block( tid, NULL, szB, VG_(clo_alignment), /*is_zeroed*/False );
    623 }
    624 
    625 static void* dh___builtin_new ( ThreadId tid, SizeT szB )
    626 {
    627    return new_block( tid, NULL, szB, VG_(clo_alignment), /*is_zeroed*/False );
    628 }
    629 
    630 static void* dh___builtin_vec_new ( ThreadId tid, SizeT szB )
    631 {
    632    return new_block( tid, NULL, szB, VG_(clo_alignment), /*is_zeroed*/False );
    633 }
    634 
    635 static void* dh_calloc ( ThreadId tid, SizeT m, SizeT szB )
    636 {
    637    return new_block( tid, NULL, m*szB, VG_(clo_alignment), /*is_zeroed*/True );
    638 }
    639 
    640 static void *dh_memalign ( ThreadId tid, SizeT alignB, SizeT szB )
    641 {
    642    return new_block( tid, NULL, szB, alignB, False );
    643 }
    644 
    645 static void dh_free ( ThreadId tid __attribute__((unused)), void* p )
    646 {
    647    die_block( p, /*custom_free*/False );
    648 }
    649 
    650 static void dh___builtin_delete ( ThreadId tid, void* p )
    651 {
    652    die_block( p, /*custom_free*/False);
    653 }
    654 
    655 static void dh___builtin_vec_delete ( ThreadId tid, void* p )
    656 {
    657    die_block( p, /*custom_free*/False );
    658 }
    659 
    660 static void* dh_realloc ( ThreadId tid, void* p_old, SizeT new_szB )
    661 {
    662    if (p_old == NULL) {
    663       return dh_malloc(tid, new_szB);
    664    }
    665    if (new_szB == 0) {
    666       dh_free(tid, p_old);
    667       return NULL;
    668    }
    669    return renew_block(tid, p_old, new_szB);
    670 }
    671 
    672 static SizeT dh_malloc_usable_size ( ThreadId tid, void* p )
    673 {
    674    tl_assert(0);
    675 //zz   HP_Chunk* hc = VG_(HT_lookup)( malloc_list, (UWord)p );
    676 //zz
    677 //zz   return ( hc ? hc->req_szB + hc->slop_szB : 0 );
    678 }
    679 
    680 //------------------------------------------------------------//
    681 //--- memory references                                    ---//
    682 //------------------------------------------------------------//
    683 
    684 static
    685 void inc_histo_for_block ( Block* bk, Addr addr, UWord szB )
    686 {
    687    UWord i, offMin, offMax1;
    688    offMin = addr - bk->payload;
    689    tl_assert(offMin < bk->req_szB);
    690    offMax1 = offMin + szB;
    691    if (offMax1 > bk->req_szB)
    692       offMax1 = bk->req_szB;
    693    //VG_(printf)("%lu %lu   (size of block %lu)\n", offMin, offMax1, bk->req_szB);
    694    for (i = offMin; i < offMax1; i++) {
    695       UShort n = bk->histoW[i];
    696       if (n < 0xFFFF) n++;
    697       bk->histoW[i] = n;
    698    }
    699 }
    700 
    701 static VG_REGPARM(2)
    702 void dh_handle_write ( Addr addr, UWord szB )
    703 {
    704    Block* bk = find_Block_containing(addr);
    705    if (bk) {
    706       bk->n_writes += szB;
    707       if (bk->histoW)
    708          inc_histo_for_block(bk, addr, szB);
    709    }
    710 }
    711 
    712 static VG_REGPARM(2)
    713 void dh_handle_read ( Addr addr, UWord szB )
    714 {
    715    Block* bk = find_Block_containing(addr);
    716    if (bk) {
    717       bk->n_reads += szB;
    718       if (bk->histoW)
    719          inc_histo_for_block(bk, addr, szB);
    720    }
    721 }
    722 
    723 
    724 // Handle reads and writes by syscalls (read == kernel
    725 // reads user space, write == kernel writes user space).
    726 // Assumes no such read or write spans a heap block
    727 // boundary and so we can treat it just as one giant
    728 // read or write.
    729 static
    730 void dh_handle_noninsn_read ( CorePart part, ThreadId tid, Char* s,
    731                               Addr base, SizeT size )
    732 {
    733    switch (part) {
    734       case Vg_CoreSysCall:
    735          dh_handle_read(base, size);
    736          break;
    737       case Vg_CoreSysCallArgInMem:
    738          break;
    739       case Vg_CoreTranslate:
    740          break;
    741       default:
    742          tl_assert(0);
    743    }
    744 }
    745 
    746 static
    747 void dh_handle_noninsn_write ( CorePart part, ThreadId tid,
    748                                Addr base, SizeT size )
    749 {
    750    switch (part) {
    751       case Vg_CoreSysCall:
    752          dh_handle_write(base, size);
    753          break;
    754       case Vg_CoreSignal:
    755          break;
    756       default:
    757          tl_assert(0);
    758    }
    759 }
    760 
    761 
    762 //------------------------------------------------------------//
    763 //--- Instrumentation                                      ---//
    764 //------------------------------------------------------------//
    765 
    766 #define binop(_op, _arg1, _arg2) IRExpr_Binop((_op),(_arg1),(_arg2))
    767 #define mkexpr(_tmp)             IRExpr_RdTmp((_tmp))
    768 #define mkU32(_n)                IRExpr_Const(IRConst_U32(_n))
    769 #define mkU64(_n)                IRExpr_Const(IRConst_U64(_n))
    770 #define assign(_t, _e)           IRStmt_WrTmp((_t), (_e))
    771 
    772 static
    773 void add_counter_update(IRSB* sbOut, Int n)
    774 {
    775    #if defined(VG_BIGENDIAN)
    776    # define END Iend_BE
    777    #elif defined(VG_LITTLEENDIAN)
    778    # define END Iend_LE
    779    #else
    780    # error "Unknown endianness"
    781    #endif
    782    // Add code to increment 'g_guest_instrs_executed' by 'n', like this:
    783    //   WrTmp(t1, Load64(&g_guest_instrs_executed))
    784    //   WrTmp(t2, Add64(RdTmp(t1), Const(n)))
    785    //   Store(&g_guest_instrs_executed, t2)
    786    IRTemp t1 = newIRTemp(sbOut->tyenv, Ity_I64);
    787    IRTemp t2 = newIRTemp(sbOut->tyenv, Ity_I64);
    788    IRExpr* counter_addr = mkIRExpr_HWord( (HWord)&g_guest_instrs_executed );
    789 
    790    IRStmt* st1 = assign(t1, IRExpr_Load(END, Ity_I64, counter_addr));
    791    IRStmt* st2 = assign(t2, binop(Iop_Add64, mkexpr(t1), mkU64(n)));
    792    IRStmt* st3 = IRStmt_Store(END, counter_addr, mkexpr(t2));
    793 
    794    addStmtToIRSB( sbOut, st1 );
    795    addStmtToIRSB( sbOut, st2 );
    796    addStmtToIRSB( sbOut, st3 );
    797 }
    798 
    799 static
    800 void addMemEvent(IRSB* sbOut, Bool isWrite, Int szB, IRExpr* addr,
    801                  Int goff_sp)
    802 {
    803    IRType   tyAddr   = Ity_INVALID;
    804    HChar*   hName    = NULL;
    805    void*    hAddr    = NULL;
    806    IRExpr** argv     = NULL;
    807    IRDirty* di       = NULL;
    808 
    809    const Int THRESH = 4096 * 4; // somewhat arbitrary
    810    const Int rz_szB = VG_STACK_REDZONE_SZB;
    811 
    812    tyAddr = typeOfIRExpr( sbOut->tyenv, addr );
    813    tl_assert(tyAddr == Ity_I32 || tyAddr == Ity_I64);
    814 
    815    if (isWrite) {
    816       hName = "dh_handle_write";
    817       hAddr = &dh_handle_write;
    818    } else {
    819       hName = "dh_handle_read";
    820       hAddr = &dh_handle_read;
    821    }
    822 
    823    argv = mkIRExprVec_2( addr, mkIRExpr_HWord(szB) );
    824 
    825    /* Add the helper. */
    826    tl_assert(hName);
    827    tl_assert(hAddr);
    828    tl_assert(argv);
    829    di = unsafeIRDirty_0_N( 2/*regparms*/,
    830                            hName, VG_(fnptr_to_fnentry)( hAddr ),
    831                            argv );
    832 
    833    /* Generate the guard condition: "(addr - (SP - RZ)) >u N", for
    834       some arbitrary N.  If that fails then addr is in the range (SP -
    835       RZ .. SP + N - RZ).  If N is smallish (a page?) then we can say
    836       addr is within a page of SP and so can't possibly be a heap
    837       access, and so can be skipped. */
    838    IRTemp sp = newIRTemp(sbOut->tyenv, tyAddr);
    839    addStmtToIRSB( sbOut, assign(sp, IRExpr_Get(goff_sp, tyAddr)));
    840 
    841    IRTemp sp_minus_rz = newIRTemp(sbOut->tyenv, tyAddr);
    842    addStmtToIRSB(
    843       sbOut,
    844       assign(sp_minus_rz,
    845              tyAddr == Ity_I32
    846                 ? binop(Iop_Sub32, mkexpr(sp), mkU32(rz_szB))
    847                 : binop(Iop_Sub64, mkexpr(sp), mkU64(rz_szB)))
    848    );
    849 
    850    IRTemp diff = newIRTemp(sbOut->tyenv, tyAddr);
    851    addStmtToIRSB(
    852       sbOut,
    853       assign(diff,
    854              tyAddr == Ity_I32
    855                 ? binop(Iop_Sub32, addr, mkexpr(sp_minus_rz))
    856                 : binop(Iop_Sub64, addr, mkexpr(sp_minus_rz)))
    857    );
    858 
    859    IRTemp guard = newIRTemp(sbOut->tyenv, Ity_I1);
    860    addStmtToIRSB(
    861       sbOut,
    862       assign(guard,
    863              tyAddr == Ity_I32
    864                 ? binop(Iop_CmpLT32U, mkU32(THRESH), mkexpr(diff))
    865                 : binop(Iop_CmpLT64U, mkU64(THRESH), mkexpr(diff)))
    866    );
    867    di->guard = mkexpr(guard);
    868 
    869    addStmtToIRSB( sbOut, IRStmt_Dirty(di) );
    870 }
    871 
    872 static
    873 IRSB* dh_instrument ( VgCallbackClosure* closure,
    874                       IRSB* sbIn,
    875                       VexGuestLayout* layout,
    876                       VexGuestExtents* vge,
    877                       IRType gWordTy, IRType hWordTy )
    878 {
    879    Int   i, n = 0;
    880    IRSB* sbOut;
    881    IRTypeEnv* tyenv = sbIn->tyenv;
    882 
    883    const Int goff_sp = layout->offset_SP;
    884 
    885    // We increment the instruction count in two places:
    886    // - just before any Ist_Exit statements;
    887    // - just before the IRSB's end.
    888    // In the former case, we zero 'n' and then continue instrumenting.
    889 
    890    sbOut = deepCopyIRSBExceptStmts(sbIn);
    891 
    892    // Copy verbatim any IR preamble preceding the first IMark
    893    i = 0;
    894    while (i < sbIn->stmts_used && sbIn->stmts[i]->tag != Ist_IMark) {
    895       addStmtToIRSB( sbOut, sbIn->stmts[i] );
    896       i++;
    897    }
    898 
    899    for (/*use current i*/; i < sbIn->stmts_used; i++) {
    900       IRStmt* st = sbIn->stmts[i];
    901 
    902       if (!st || st->tag == Ist_NoOp) continue;
    903 
    904       switch (st->tag) {
    905 
    906          case Ist_IMark: {
    907             n++;
    908             break;
    909          }
    910 
    911          case Ist_Exit: {
    912             if (n > 0) {
    913                // Add an increment before the Exit statement, then reset 'n'.
    914                add_counter_update(sbOut, n);
    915                n = 0;
    916             }
    917             break;
    918          }
    919 
    920          case Ist_WrTmp: {
    921             IRExpr* data = st->Ist.WrTmp.data;
    922             if (data->tag == Iex_Load) {
    923                IRExpr* aexpr = data->Iex.Load.addr;
    924                // Note also, endianness info is ignored.  I guess
    925                // that's not interesting.
    926                addMemEvent( sbOut, False/*!isWrite*/,
    927                             sizeofIRType(data->Iex.Load.ty),
    928                             aexpr, goff_sp );
    929             }
    930             break;
    931          }
    932 
    933          case Ist_Store: {
    934             IRExpr* data  = st->Ist.Store.data;
    935             IRExpr* aexpr = st->Ist.Store.addr;
    936             addMemEvent( sbOut, True/*isWrite*/,
    937                          sizeofIRType(typeOfIRExpr(tyenv, data)),
    938                          aexpr, goff_sp );
    939             break;
    940          }
    941 
    942          case Ist_Dirty: {
    943             Int      dataSize;
    944             IRDirty* d = st->Ist.Dirty.details;
    945             if (d->mFx != Ifx_None) {
    946                /* This dirty helper accesses memory.  Collect the details. */
    947                tl_assert(d->mAddr != NULL);
    948                tl_assert(d->mSize != 0);
    949                dataSize = d->mSize;
    950                // Large (eg. 28B, 108B, 512B on x86) data-sized
    951                // instructions will be done inaccurately, but they're
    952                // very rare and this avoids errors from hitting more
    953                // than two cache lines in the simulation.
    954                if (d->mFx == Ifx_Read || d->mFx == Ifx_Modify)
    955                   addMemEvent( sbOut, False/*!isWrite*/,
    956                                dataSize, d->mAddr, goff_sp );
    957                if (d->mFx == Ifx_Write || d->mFx == Ifx_Modify)
    958                   addMemEvent( sbOut, True/*isWrite*/,
    959                                dataSize, d->mAddr, goff_sp );
    960             } else {
    961                tl_assert(d->mAddr == NULL);
    962                tl_assert(d->mSize == 0);
    963             }
    964             break;
    965          }
    966 
    967          case Ist_CAS: {
    968             /* We treat it as a read and a write of the location.  I
    969                think that is the same behaviour as it was before IRCAS
    970                was introduced, since prior to that point, the Vex
    971                front ends would translate a lock-prefixed instruction
    972                into a (normal) read followed by a (normal) write. */
    973             Int    dataSize;
    974             IRCAS* cas = st->Ist.CAS.details;
    975             tl_assert(cas->addr != NULL);
    976             tl_assert(cas->dataLo != NULL);
    977             dataSize = sizeofIRType(typeOfIRExpr(tyenv, cas->dataLo));
    978             if (cas->dataHi != NULL)
    979                dataSize *= 2; /* since it's a doubleword-CAS */
    980             addMemEvent( sbOut, False/*!isWrite*/,
    981                          dataSize, cas->addr, goff_sp );
    982             addMemEvent( sbOut, True/*isWrite*/,
    983                          dataSize, cas->addr, goff_sp );
    984             break;
    985          }
    986 
    987          case Ist_LLSC: {
    988             IRType dataTy;
    989             if (st->Ist.LLSC.storedata == NULL) {
    990                /* LL */
    991                dataTy = typeOfIRTemp(tyenv, st->Ist.LLSC.result);
    992                addMemEvent( sbOut, False/*!isWrite*/,
    993                             sizeofIRType(dataTy),
    994                             st->Ist.LLSC.addr, goff_sp );
    995             } else {
    996                /* SC */
    997                dataTy = typeOfIRExpr(tyenv, st->Ist.LLSC.storedata);
    998                addMemEvent( sbOut, True/*isWrite*/,
    999                             sizeofIRType(dataTy),
   1000                             st->Ist.LLSC.addr, goff_sp );
   1001             }
   1002             break;
   1003          }
   1004 
   1005          default:
   1006             break;
   1007       }
   1008 
   1009       addStmtToIRSB( sbOut, st );
   1010    }
   1011 
   1012    if (n > 0) {
   1013       // Add an increment before the SB end.
   1014       add_counter_update(sbOut, n);
   1015    }
   1016    return sbOut;
   1017 }
   1018 
   1019 #undef binop
   1020 #undef mkexpr
   1021 #undef mkU32
   1022 #undef mkU64
   1023 #undef assign
   1024 
   1025 
   1026 //------------------------------------------------------------//
   1027 //--- Command line args                                    ---//
   1028 //------------------------------------------------------------//
   1029 
   1030 // FORWARDS
   1031 static Bool identify_metric ( /*OUT*/ULong(**get_metricP)(APInfo*),
   1032                               /*OUT*/Bool* increasingP,
   1033                               Char* metric_name );
   1034 
   1035 static Int    clo_show_top_n = 10;
   1036 static HChar* clo_sort_by    = "max-bytes-live";
   1037 
   1038 static Bool dh_process_cmd_line_option(Char* arg)
   1039 {
   1040    if VG_BINT_CLO(arg, "--show-top-n", clo_show_top_n, 1, 100000) {}
   1041 
   1042    else if VG_STR_CLO(arg, "--sort-by", clo_sort_by) {
   1043        ULong (*dummyFn)(APInfo*);
   1044        Bool dummyB;
   1045        Bool ok = identify_metric( &dummyFn, &dummyB, clo_sort_by);
   1046        if (!ok)
   1047           return False;
   1048        // otherwise it's OK, in which case leave it alone.
   1049        // show_top_n_apinfos will later convert the string by a
   1050        // second call to identify_metric.
   1051    }
   1052 
   1053    else
   1054       return VG_(replacement_malloc_process_cmd_line_option)(arg);
   1055 
   1056    return True;
   1057 }
   1058 
   1059 
   1060 static void dh_print_usage(void)
   1061 {
   1062    VG_(printf)(
   1063 "    --show-top-n=number       show the top <number> alloc points [10]\n"
   1064 "    --sort-by=string\n"
   1065 "            sort the allocation points by the metric\n"
   1066 "            defined by <string>, thusly:\n"
   1067 "                max-bytes-live    maximum live bytes [default]\n"
   1068 "                tot-bytes-allocd  total allocation (turnover)\n"
   1069 "                max-blocks-live   maximum live blocks\n"
   1070    );
   1071 }
   1072 
   1073 static void dh_print_debug_usage(void)
   1074 {
   1075    VG_(printf)(
   1076 "    (none)\n"
   1077    );
   1078 }
   1079 
   1080 
   1081 //------------------------------------------------------------//
   1082 //--- Finalisation                                         ---//
   1083 //------------------------------------------------------------//
   1084 
   1085 static void show_N_div_100( /*OUT*/HChar* buf, ULong n )
   1086 {
   1087    ULong nK = n / 100;
   1088    ULong nR = n % 100;
   1089    VG_(sprintf)(buf, "%llu.%s%llu", nK,
   1090                 nR < 10 ? "0" : "",
   1091                 nR);
   1092 }
   1093 
   1094 static void show_APInfo ( APInfo* api )
   1095 {
   1096    HChar bufA[80];
   1097    VG_(memset)(bufA, 0, sizeof(bufA));
   1098    if (api->tot_blocks > 0) {
   1099       show_N_div_100( bufA, ((ULong)api->tot_bytes * 100ULL)
   1100                               / (ULong)api->tot_blocks );
   1101    } else {
   1102       bufA[0] = 'N'; bufA[1] = 'a'; bufA[2] = 'N';
   1103    }
   1104 
   1105    VG_(umsg)("max-live:    %'llu in %'llu blocks\n",
   1106              api->max_bytes_live, api->max_blocks_live);
   1107    VG_(umsg)("tot-alloc:   %'llu in %'llu blocks (avg size %s)\n",
   1108              api->tot_bytes, api->tot_blocks, bufA);
   1109 
   1110    tl_assert(api->tot_blocks >= api->max_blocks_live);
   1111    tl_assert(api->tot_bytes >= api->max_bytes_live);
   1112 
   1113    if (api->deaths > 0) {
   1114       // Average Age at Death
   1115       ULong aad = api->deaths == 0
   1116                   ? 0 : (api->death_ages_sum / api->deaths);
   1117       // AAD as a fraction of the total program lifetime (so far)
   1118       // measured in ten-thousand-ths (aad_frac_10k == 10000 means the
   1119       // complete lifetime of the program.
   1120       ULong aad_frac_10k
   1121          = g_guest_instrs_executed == 0
   1122            ? 0 : (10000ULL * aad) / g_guest_instrs_executed;
   1123       HChar buf[16];
   1124       show_N_div_100(buf, aad_frac_10k);
   1125       VG_(umsg)("deaths:      %'llu, at avg age %'llu "
   1126                 "(%s%% of prog lifetime)\n",
   1127                 api->deaths, aad, buf );
   1128    } else {
   1129       VG_(umsg)("deaths:      none (none of these blocks were freed)\n");
   1130    }
   1131 
   1132    HChar bufR[80], bufW[80];
   1133    VG_(memset)(bufR, 0, sizeof(bufR));
   1134    VG_(memset)(bufW, 0, sizeof(bufW));
   1135    if (api->tot_bytes > 0) {
   1136       show_N_div_100(bufR, (100ULL * api->n_reads) / api->tot_bytes);
   1137       show_N_div_100(bufW, (100ULL * api->n_writes) / api->tot_bytes);
   1138    } else {
   1139       VG_(strcat)(bufR, "Inf");
   1140       VG_(strcat)(bufW, "Inf");
   1141    }
   1142 
   1143    VG_(umsg)("acc-ratios:  %s rd, %s wr "
   1144              " (%'llu b-read, %'llu b-written)\n",
   1145              bufR, bufW,
   1146              api->n_reads, api->n_writes);
   1147 
   1148    VG_(pp_ExeContext)(api->ap);
   1149 
   1150    if (api->histo && api->xsize_tag == Exactly) {
   1151       VG_(umsg)("\nAggregated access counts by offset:\n");
   1152       VG_(umsg)("\n");
   1153       UWord i;
   1154       if (api->xsize > 0)
   1155          VG_(umsg)("[   0]  ");
   1156       for (i = 0; i < api->xsize; i++) {
   1157          if (i > 0 && (i % 16) == 0 && i != api->xsize-1) {
   1158             VG_(umsg)("\n");
   1159             VG_(umsg)("[%4lu]  ", i);
   1160          }
   1161          VG_(umsg)("%u ", api->histo[i]);
   1162       }
   1163       VG_(umsg)("\n");
   1164    }
   1165 }
   1166 
   1167 
   1168 /* Metric-access functions for APInfos. */
   1169 static ULong get_metric__max_bytes_live ( APInfo* api ) {
   1170    return api->max_bytes_live;
   1171 }
   1172 static ULong get_metric__tot_bytes ( APInfo* api ) {
   1173    return api->tot_bytes;
   1174 }
   1175 static ULong get_metric__max_blocks_live ( APInfo* api ) {
   1176    return api->max_blocks_live;
   1177 }
   1178 
   1179 /* Given a string, return the metric-access function and also a Bool
   1180    indicating whether we want increasing or decreasing values of the
   1181    metric.  This is used twice, once in command line processing, and
   1182    then again in show_top_n_apinfos.  Returns False if the given
   1183    string could not be identified.*/
   1184 static Bool identify_metric ( /*OUT*/ULong(**get_metricP)(APInfo*),
   1185                               /*OUT*/Bool* increasingP,
   1186                               Char* metric_name )
   1187 {
   1188    if (0 == VG_(strcmp)(metric_name, "max-bytes-live")) {
   1189       *get_metricP = get_metric__max_bytes_live;
   1190       *increasingP = False;
   1191       return True;
   1192    }
   1193    if (0 == VG_(strcmp)(metric_name, "tot-bytes-allocd")) {
   1194       *get_metricP = get_metric__tot_bytes;
   1195       *increasingP = False;
   1196       return True;
   1197    }
   1198    if (0 == VG_(strcmp)(metric_name, "max-blocks-live")) {
   1199       *get_metricP = get_metric__max_blocks_live;
   1200       *increasingP = False;
   1201       return True;
   1202    }
   1203    return False;
   1204 }
   1205 
   1206 
   1207 static void show_top_n_apinfos ( void )
   1208 {
   1209    Int   i;
   1210    UWord keyW, valW;
   1211    ULong (*get_metric)(APInfo*);
   1212    Bool  increasing;
   1213 
   1214    HChar* metric_name = clo_sort_by;
   1215    tl_assert(metric_name); // ensured by clo processing
   1216 
   1217    Bool ok = identify_metric( &get_metric, &increasing, metric_name );
   1218    tl_assert(ok); // ensured by clo processing
   1219 
   1220    VG_(umsg)("\n");
   1221    VG_(umsg)("======== ORDERED BY %s \"%s\": "
   1222              "top %d allocators ========\n",
   1223              increasing ? "increasing" : "decreasing",
   1224              metric_name, clo_show_top_n );
   1225 
   1226    // Clear all .shown bits
   1227    VG_(initIterFM)( apinfo );
   1228    while (VG_(nextIterFM)( apinfo, &keyW, &valW )) {
   1229       APInfo* api = (APInfo*)valW;
   1230       tl_assert(api && api->ap == (ExeContext*)keyW);
   1231       api->shown = False;
   1232    }
   1233    VG_(doneIterFM)( apinfo );
   1234 
   1235    // Now print the top N entries.  Each one requires a
   1236    // complete scan of the set.  Duh.
   1237    for (i = 0; i < clo_show_top_n; i++) {
   1238       ULong   best_metric = increasing ? ~0ULL : 0ULL;
   1239       APInfo* best_api    = NULL;
   1240 
   1241       VG_(initIterFM)( apinfo );
   1242       while (VG_(nextIterFM)( apinfo, &keyW, &valW )) {
   1243          APInfo* api = (APInfo*)valW;
   1244          if (api->shown)
   1245             continue;
   1246          ULong metric = get_metric(api);
   1247          if (increasing ? (metric < best_metric) : (metric > best_metric)) {
   1248             best_metric = metric;
   1249             best_api = api;
   1250          }
   1251       }
   1252       VG_(doneIterFM)( apinfo );
   1253 
   1254       if (!best_api)
   1255          break; // all APIs have been shown.  Stop.
   1256 
   1257       VG_(umsg)("\n");
   1258       VG_(umsg)("-------------------- %d of %d --------------------\n",
   1259                 i+1, clo_show_top_n );
   1260       show_APInfo(best_api);
   1261       best_api->shown = True;
   1262    }
   1263 
   1264    VG_(umsg)("\n");
   1265 }
   1266 
   1267 
   1268 static void dh_fini(Int exit_status)
   1269 {
   1270    // Before printing statistics, we must harvest access counts for
   1271    // all the blocks that are still alive.  Not doing so gives
   1272    // access ratios which are too low (zero, in the worst case)
   1273    // for such blocks, since the accesses that do get made will
   1274    // (if we skip this step) not get folded into the AP summaries.
   1275    UWord keyW, valW;
   1276    VG_(initIterFM)( interval_tree );
   1277    while (VG_(nextIterFM)( interval_tree, &keyW, &valW )) {
   1278       Block* bk = (Block*)keyW;
   1279       tl_assert(valW == 0);
   1280       tl_assert(bk);
   1281       retire_Block(bk, False/*!because_freed*/);
   1282    }
   1283    VG_(doneIterFM)( interval_tree );
   1284 
   1285    // show results
   1286    VG_(umsg)("======== SUMMARY STATISTICS ========\n");
   1287    VG_(umsg)("\n");
   1288    VG_(umsg)("guest_insns:  %'llu\n", g_guest_instrs_executed);
   1289    VG_(umsg)("\n");
   1290    VG_(umsg)("max_live:     %'llu in %'llu blocks\n",
   1291              g_max_bytes_live, g_max_blocks_live);
   1292    VG_(umsg)("\n");
   1293    VG_(umsg)("tot_alloc:    %'llu in %'llu blocks\n",
   1294              g_tot_bytes, g_tot_blocks);
   1295    VG_(umsg)("\n");
   1296    if (g_tot_bytes > 0) {
   1297       VG_(umsg)("insns per allocated byte: %'llu\n",
   1298                 g_guest_instrs_executed / g_tot_bytes);
   1299       VG_(umsg)("\n");
   1300    }
   1301 
   1302    show_top_n_apinfos();
   1303 
   1304    VG_(umsg)("\n");
   1305    VG_(umsg)("\n");
   1306    VG_(umsg)("==============================================================\n");
   1307    VG_(umsg)("\n");
   1308    VG_(umsg)("Some hints: (see --help for command line option details):\n");
   1309    VG_(umsg)("\n");
   1310    VG_(umsg)("* summary stats for whole program are at the top of this output\n");
   1311    VG_(umsg)("\n");
   1312    VG_(umsg)("* --show-top-n=  controls how many alloc points are shown.\n");
   1313    VG_(umsg)("                 You probably want to set it much higher than\n");
   1314    VG_(umsg)("                 the default value (10)\n");
   1315    VG_(umsg)("\n");
   1316    VG_(umsg)("* --sort-by=     specifies the sort key for output.\n");
   1317    VG_(umsg)("                 See --help for details.\n");
   1318    VG_(umsg)("\n");
   1319    VG_(umsg)("* Each allocation stack, by default 12 frames, counts as\n");
   1320    VG_(umsg)("  a separate alloc point.  This causes the data to be spread out\n");
   1321    VG_(umsg)("  over far too many alloc points.  I strongly suggest using\n");
   1322    VG_(umsg)("  --num-callers=4 or some such, to reduce the spreading.\n");
   1323    VG_(umsg)("\n");
   1324 
   1325    if (VG_(clo_stats)) {
   1326       VG_(dmsg)(" dhat: find_Block_containing:\n");
   1327       VG_(dmsg)("             found: %'lu (%'lu cached + %'lu uncached)\n",
   1328                 stats__n_fBc_cached + stats__n_fBc_uncached,
   1329                 stats__n_fBc_cached,
   1330                 stats__n_fBc_uncached);
   1331       VG_(dmsg)("          notfound: %'lu\n", stats__n_fBc_notfound);
   1332       VG_(dmsg)("\n");
   1333    }
   1334 }
   1335 
   1336 
   1337 //------------------------------------------------------------//
   1338 //--- Initialisation                                       ---//
   1339 //------------------------------------------------------------//
   1340 
   1341 static void dh_post_clo_init(void)
   1342 {
   1343 }
   1344 
   1345 static void dh_pre_clo_init(void)
   1346 {
   1347    VG_(details_name)            ("DHAT");
   1348    VG_(details_version)         (NULL);
   1349    VG_(details_description)     ("a dynamic heap analysis tool");
   1350    VG_(details_copyright_author)(
   1351       "Copyright (C) 2010-2011, and GNU GPL'd, by Mozilla Inc");
   1352    VG_(details_bug_reports_to)  (VG_BUGS_TO);
   1353 
   1354    // Basic functions.
   1355    VG_(basic_tool_funcs)          (dh_post_clo_init,
   1356                                    dh_instrument,
   1357                                    dh_fini);
   1358 //zz
   1359    // Needs.
   1360    VG_(needs_libc_freeres)();
   1361    VG_(needs_command_line_options)(dh_process_cmd_line_option,
   1362                                    dh_print_usage,
   1363                                    dh_print_debug_usage);
   1364 //zz   VG_(needs_client_requests)     (dh_handle_client_request);
   1365 //zz   VG_(needs_sanity_checks)       (dh_cheap_sanity_check,
   1366 //zz                                   dh_expensive_sanity_check);
   1367    VG_(needs_malloc_replacement)  (dh_malloc,
   1368                                    dh___builtin_new,
   1369                                    dh___builtin_vec_new,
   1370                                    dh_memalign,
   1371                                    dh_calloc,
   1372                                    dh_free,
   1373                                    dh___builtin_delete,
   1374                                    dh___builtin_vec_delete,
   1375                                    dh_realloc,
   1376                                    dh_malloc_usable_size,
   1377                                    0 );
   1378 
   1379    VG_(track_pre_mem_read)        ( dh_handle_noninsn_read );
   1380    //VG_(track_pre_mem_read_asciiz) ( check_mem_is_defined_asciiz );
   1381    VG_(track_post_mem_write)      ( dh_handle_noninsn_write );
   1382 
   1383    tl_assert(!interval_tree);
   1384    tl_assert(!fbc_cache0);
   1385    tl_assert(!fbc_cache1);
   1386 
   1387    interval_tree = VG_(newFM)( VG_(malloc),
   1388                                "dh.main.interval_tree.1",
   1389                                VG_(free),
   1390                                interval_tree_Cmp );
   1391 
   1392    apinfo = VG_(newFM)( VG_(malloc),
   1393                         "dh.main.apinfo.1",
   1394                         VG_(free),
   1395                         NULL/*unboxedcmp*/ );
   1396 }
   1397 
   1398 VG_DETERMINE_INTERFACE_VERSION(dh_pre_clo_init)
   1399 
   1400 //--------------------------------------------------------------------//
   1401 //--- end                                                dh_main.c ---//
   1402 //--------------------------------------------------------------------//
   1403