summaryrefslogtreecommitdiff
path: root/exp-dhat
diff options
context:
space:
mode:
authorsewardj <sewardj@a5019735-40e9-0310-863c-91ae7b9d1cf9>2010-10-13 14:06:00 +0000
committersewardj <sewardj@a5019735-40e9-0310-863c-91ae7b9d1cf9>2010-10-13 14:06:00 +0000
commit5fce756babf4b7b1cc65540b7b72a3a8b25d7bb3 (patch)
tree1e617c9e897d8746e54ccb54eb2ebb43c08f8563 /exp-dhat
parent00d9d1e0cc810c0aa7974119b5c4750c7a71beec (diff)
Add documentation for exp-dhat.
git-svn-id: svn://svn.valgrind.org/valgrind/trunk@11438 a5019735-40e9-0310-863c-91ae7b9d1cf9
Diffstat (limited to 'exp-dhat')
-rw-r--r--exp-dhat/Makefile.am4
-rw-r--r--exp-dhat/docs/dh-manual.xml400
2 files changed, 401 insertions, 3 deletions
diff --git a/exp-dhat/Makefile.am b/exp-dhat/Makefile.am
index 8a88b69b..9ee6497d 100644
--- a/exp-dhat/Makefile.am
+++ b/exp-dhat/Makefile.am
@@ -2,9 +2,7 @@ include $(top_srcdir)/Makefile.tool.am
#SUBDIRS += perf
-#EXTRA_DIST = \
-# docs/dh-manual.xml \
-# docs/dh_print-manpage.xml
+EXTRA_DIST = docs/dh-manual.xml
#----------------------------------------------------------------------------
# Headers, etc
diff --git a/exp-dhat/docs/dh-manual.xml b/exp-dhat/docs/dh-manual.xml
new file mode 100644
index 00000000..49df7a6b
--- /dev/null
+++ b/exp-dhat/docs/dh-manual.xml
@@ -0,0 +1,400 @@
+<?xml version="1.0"?> <!-- -*- sgml -*- -->
+<!DOCTYPE chapter PUBLIC "-//OASIS//DTD DocBook XML V4.2//EN"
+ "http://www.oasis-open.org/docbook/xml/4.2/docbookx.dtd"
+[ <!ENTITY % vg-entities SYSTEM "../../docs/xml/vg-entities.xml"> %vg-entities; ]>
+
+
+<chapter id="dh-manual"
+ xreflabel="DHAT: a dynamic heap analysis tool">
+ <title>DHAT: a dynamic heap analysis tool</title>
+
+<para>To use this tool, you must specify
+<option>--tool=exp-dhat</option> on the Valgrind
+command line.</para>
+
+
+
+<sect1 id="dh-manual.overview" xreflabel="Overview">
+<title>Overview</title>
+
+<para>DHAT is a tool for examining how programs use their heap
+allocations.</para>
+
+<para>It tracks the allocated blocks, and inspects every memory access
+to find which block, if any, it is to. The following data is
+collected and presented per allocation point (allocation
+stack):</para>
+
+<itemizedlist>
+ <listitem><para>Total allocation (number of bytes and
+ blocks)</para></listitem>
+
+ <listitem><para>maximum live volume (number of bytes and
+ blocks)</para></listitem>
+
+ <listitem><para>average block lifetime (number of instructions
+ between allocation and freeing)</para></listitem>
+
+ <listitem><para>average number of reads and writes to each byte in
+ the block ("access ratios")</para></listitem>
+
+ <listitem><para>for allocation points which always allocate blocks
+ only of one size, and that size is 4096 bytes or less: counts
+ showing how often each byte offset inside the block is
+ accessed.</para></listitem>
+</itemizedlist>
+
+<para>Using these statistics it is possible to identify allocation
+points with the following characteristics:</para>
+
+<itemizedlist>
+
+ <listitem><para>potential process-lifetime leaks: blocks allocated
+ by the point just accumulate, and are freed only at the end of the
+ run.</para></listitem>
+
+ <listitem><para>excessive turnover: points which chew through a lot
+ of heap, even if it is not held onto for very long</para></listitem>
+
+ <listitem><para>excessively transient: points which allocate very
+ short lived blocks</para></listitem>
+
+ <listitem><para>useless or underused allocations: blocks which are
+ allocated but not completely filled in, or are filled in but not
+ subsequently read.</para></listitem>
+
+ <listitem><para>blocks with inefficient layout -- areas never
+ accessed, or with hot fields scattered throughout the
+ block.</para></listitem>
+</itemizedlist>
+
+<para>As with the Massif heap profiler, DHAT measures program progress
+by counting instructions, and so presents all age/time related figures
+as instruction counts. This sounds a little odd at first, but it
+makes runs repeatable in a way which is not possible if CPU time is
+used.</para>
+
+</sect1>
+
+
+
+
+<sect1 id="dh-manual.understanding" xreflabel="Understanding DHAT's output">
+<title>Understanding DHAT's output</title>
+
+
+<para>DHAT provides a lot of useful information on dynamic heap usage.
+Most of the art of using it is in interpretation of the resulting
+numbers. That is best illustrated via a set of examples.</para>
+
+
+<sect2>
+<title>Interpreting the max-live, tot-alloc and deaths fields</title>
+
+<sect3><title>A simple example</title></sect3>
+
+<screen><![CDATA[
+ ======== SUMMARY STATISTICS ========
+
+ guest_insns: 1,045,339,534
+ [...]
+ max-live: 63,490 in 984 blocks
+ tot-alloc: 1,904,700 in 29,520 blocks (avg size 64.52)
+ deaths: 29,520, at avg age 22,227,424
+ acc-ratios: 6.37 rd, 1.14 wr (12,141,526 b-read, 2,174,460 b-written)
+ at 0x4C275B8: malloc (vg_replace_malloc.c:236)
+ by 0x40350E: tcc_malloc (tinycc.c:6712)
+ by 0x404580: tok_alloc_new (tinycc.c:7151)
+ by 0x40870A: next_nomacro1 (tinycc.c:9305)
+]]></screen>
+
+<para>Over the entire run of the program, this stack (allocation
+point) allocated 29,520 blocks in total, containing 1,904,700 bytes in
+total. By looking at the max-live data, we see that not many blocks
+were simultaneously live, though: at the peak, there were 63,490
+allocated bytes in 984 blocks. This tells us that the program is
+steadily freeing such blocks as it runs, rather than hanging on to all
+of them until the end and freeing them all.</para>
+
+<para>The deaths entry tells us that 29,520 blocks allocated by this stack
+died (were freed) during the run of the program. Since 29,520 is
+also the number of blocks allocated in total, that tells us that
+all allocated blocks were freed by the end of the program.</para>
+
+<para>It also tells us that the average age at death was 22,227,424
+instructions. From the summary statistics we see that the program ran
+for 1,045,339,534 instructions, and so the average age at death is
+about 2% of the program's total run time.</para>
+
+<sect3><title>Example of a potential process-lifetime leak</title></sect3>
+
+<para>This next example (from a different program than the above)
+shows a potential process lifetime leak. A process lifetime leak
+occurs when a program keeps allocating data, but only frees the
+data just before it exits. Hence the program's heap grows constantly
+in size, yet Memcheck reports no leak, because the program has
+freed up everything at exit. This is particularly a hazard for
+long running programs.</para>
+
+<screen><![CDATA[
+ ======== SUMMARY STATISTICS ========
+
+ guest_insns: 418,901,537
+ [...]
+ max-live: 32,512 in 254 blocks
+ tot-alloc: 32,512 in 254 blocks (avg size 128.00)
+ deaths: 254, at avg age 300,467,389
+ acc-ratios: 0.26 rd, 0.20 wr (8,756 b-read, 6,604 b-written)
+ at 0x4C275B8: malloc (vg_replace_malloc.c:236)
+ by 0x4C27632: realloc (vg_replace_malloc.c:525)
+ by 0x56FF41D: QtFontStyle::pixelSize(unsigned short, bool) (qfontdatabase.cpp:269)
+ by 0x5700D69: loadFontConfig() (qfontdatabase_x11.cpp:1146)
+]]></screen>
+
+<para>There are two tell-tale signs that this might be a
+process-lifetime leak. Firstly, the max-live and tot-alloc numbers
+are identical. The only way that can happen is if these blocks are
+all allocated and then all deallocated.</para>
+
+<para>Secondly, the average age at death (300 million insns) is 71% of
+the total program lifetime (419 million insns), hence this is not a
+transient allocation-free spike -- rather, it is spread out over a
+large part of the entire run. One interpretation is, roughly, that
+all 254 blocks were allocated in the first half of the run, held onto
+for the second half, and then freed just before exit.</para>
+
+</sect2>
+
+
+<sect2>
+<title>Interpreting the acc-ratios fields</title>
+
+
+<sect3><title>A fairly harmless allocation point record</title></sect3>
+
+<screen><![CDATA[
+ max-live: 49,398 in 808 blocks
+ tot-alloc: 1,481,940 in 24,240 blocks (avg size 61.13)
+ deaths: 24,240, at avg age 34,611,026
+ acc-ratios: 2.13 rd, 0.91 wr (3,166,650 b-read, 1,358,820 b-written)
+ at 0x4C275B8: malloc (vg_replace_malloc.c:236)
+ by 0x40350E: tcc_malloc (tinycc.c:6712)
+ by 0x404580: tok_alloc_new (tinycc.c:7151)
+ by 0x4046C4: tok_alloc (tinycc.c:7190)
+]]></screen>
+
+<para>The acc-ratios field tells us that each byte in the blocks
+allocated here is read an average of 2.13 times before the block is
+deallocated. Given that the blocks have an average age at death of
+34,611,026, that's one read per block per approximately every 15
+million instructions. So from that standpoint the blocks aren't
+"working" very hard.</para>
+
+<para>More interesting is the write ratio: each byte is written an
+average of 0.91 times. This tells us that some parts of the allocated
+blocks are never written, at least 9% on average. To completely
+initialise the block would require writing each byte at least once,
+and that would give a write ratio of 1.0. The fact that some block
+areas are evidently unused might point to data alignment holes or
+other layout inefficiencies.</para>
+
+<para>Well, at least all the blocks are freed (24,240 allocations,
+24,240 deaths).</para>
+
+<para>If all the blocks had been the same size, DHAT would also show
+the access counts by block offset, so we could see where exactly these
+unused areas are. However, that isn't the case: the blocks have
+varying sizes, so DHAT can't perform such an analysis. We can see
+that they must have varying sizes since the average block size, 61.13,
+isn't a whole number.</para>
+
+
+<sect3><title>A more suspicious looking example</title></sect3>
+
+<screen><![CDATA[
+ max-live: 180,224 in 22 blocks
+ tot-alloc: 180,224 in 22 blocks (avg size 8192.00)
+ deaths: none (none of these blocks were freed)
+ acc-ratios: 0.00 rd, 0.00 wr (0 b-read, 0 b-written)
+ at 0x4C275B8: malloc (vg_replace_malloc.c:236)
+ by 0x40350E: tcc_malloc (tinycc.c:6712)
+ by 0x40369C: __sym_malloc (tinycc.c:6787)
+ by 0x403711: sym_malloc (tinycc.c:6805)
+]]></screen>
+
+<para>Here, both the read and write access ratios are zero. Hence
+this point is allocating blocks which are never used, neither read nor
+written. Indeed, they are also not freed ("deaths: none") and are
+simply leaked. So, here is 180k of completely useless allocation that
+could be removed.</para>
+
+<para>Re-running with Memcheck does indeed report the same leak. What
+DHAT can tell us, that Memcheck can't, is that not only are the blocks
+leaked, they are also never used.</para>
+
+<sect3><title>Another suspicious example</title></sect3>
+
+<para>Here's one where blocks are allocated, written to,
+but never read from. We see this immediately from the zero read
+access ratio. They do get freed, though:</para>
+
+<screen><![CDATA[
+ max-live: 54 in 3 blocks
+ tot-alloc: 1,620 in 90 blocks (avg size 18.00)
+ deaths: 90, at avg age 34,558,236
+ acc-ratios: 0.00 rd, 1.11 wr (0 b-read, 1,800 b-written)
+ at 0x4C275B8: malloc (vg_replace_malloc.c:236)
+ by 0x40350E: tcc_malloc (tinycc.c:6712)
+ by 0x4035BD: tcc_strdup (tinycc.c:6750)
+ by 0x41FEBB: tcc_add_sysinclude_path (tinycc.c:20931)
+]]></screen>
+
+<para>In the previous two examples, it is easy to see blocks that are
+never written to, or never read from, or some combination of both.
+Unfortunately, in C++ code, the situation is less clear. That's
+because an object's constructor will write to the underlying block,
+and its destructor will read from it. So the block's read and write
+ratios will be non-zero even if the object, once constructed, is never
+used, but only eventually destructed.</para>
+
+<para>Really, what we want is to measure only memory accesses in
+between the end of an object's construction and the start of its
+destruction. Unfortunately I do not know of a reliable way to
+determine when those transitions are made.</para>
+
+
+</sect2>
+
+<sect2>
+<title>Interpreting "Aggregated access counts by offset" data</title>
+
+<para>For allocation points that always allocate blocks of the same
+size, and which are 4096 bytes or smaller, DHAT counts accesses
+per offset, for example:</para>
+
+<screen><![CDATA[
+ max-live: 317,408 in 5,668 blocks
+ tot-alloc: 317,408 in 5,668 blocks (avg size 56.00)
+ deaths: 5,668, at avg age 622,890,597
+ acc-ratios: 1.03 rd, 1.28 wr (327,642 b-read, 408,172 b-written)
+ at 0x4C275B8: malloc (vg_replace_malloc.c:236)
+ by 0x5440C16: QDesignerPropertySheetPrivate::ensureInfo (qhash.h:515)
+ by 0x544350B: QDesignerPropertySheet::setVisible (qdesigner_propertysh...)
+ by 0x5446232: QDesignerPropertySheet::QDesignerPropertySheet (qdesigne...)
+
+ Aggregated access counts by offset:
+
+ [ 0] 28782 28782 28782 28782 28782 28782 28782 28782
+ [ 8] 20638 20638 20638 20638 0 0 0 0
+ [ 16] 22738 22738 22738 22738 22738 22738 22738 22738
+ [ 24] 6013 6013 6013 6013 6013 6013 6013 6013
+ [ 32] 18883 18883 18883 37422 0 0 0 0
+ [ 36] 5668 11915 5668 5668 11336 11336 11336 11336
+ [ 48] 6166 6166 6166 6166 0 0 0 0
+]]></screen>
+
+<para>This is fairly typical, for C++ code running on a 64-bit
+platform. Here, we have aggregated access statistics for 5668 blocks,
+all of size 56 bytes. Each byte has been accessed at least 5668
+times, except for offsets 12--15, 36--39 and 52--55. These are likely
+to be alignment holes.</para>
+
+<para>Careful interpretation of the numbers reveals useful information.
+Groups of N consecutive identical numbers that begin at an N-aligned
+offset, for N being 2, 4 or 8, are likely to indicate an N-byte object
+in the structure at that point. For example, the first 32 bytes of
+this object are likely to have the layout</para>
+
+<screen><![CDATA[
+ [0 ] 64-bit type
+ [8 ] 32-bit type
+ [12] 32-bit alignment hole
+ [16] 64-bit type
+ [24] 64-bit type
+]]></screen>
+
+<para>As a counterexample, it's also clear that, whatever is at offset 32,
+it is not a 32-bit value. That's because the last number of the group
+(37422) is not the same as the first three (18883 18883 18883).</para>
+
+<para>This example leads one to enquire (by reading the source code)
+whether the zeroes at 12--15 and 52--55 are alignment holes, and
+whether 48--51 is indeed a 32-bit type. If so, it might be possible
+to place what's at 48--51 at 12--15 instead, which would reduce
+the object size from 56 to 48 bytes.</para>
+
+<para>Bear in mind that the above inferences are all only "maybes". That's
+because they are based on dynamic data, not static analysis of the
+object layout. For example, the zeroes might not be alignment
+holes, but rather just parts of the structure which were not used
+at all for this particular run. Experience shows that's unlikely
+to be the case, but it could happen.</para>
+
+</sect2>
+
+</sect1>
+
+
+
+
+
+
+
+<sect1 id="dh-manual.options" xreflabel="DHAT Command-line Options">
+<title>DHAT Command-line Options</title>
+
+<para>DHAT-specific command-line options are:</para>
+
+<!-- start of xi:include in the manpage -->
+<variablelist id="dh.opts.list">
+
+ <varlistentry id="opt.show-top-n" xreflabel="--show-top-n">
+ <term>
+ <option><![CDATA[--show-top-n=<number>
+ [default: 10] ]]></option>
+ </term>
+ <listitem>
+ <para>At the end of the run, DHAT sorts the accumulated
+ allocation points according to some metric, and shows the
+ highest scoring entries. <varname>--show-top-n</varname>
+ controls how many entries are shown. The default of 10 is
+ quite small. For realistic applications you will probably need
+ to set it much higher, at least several hundred.</para>
+ </listitem>
+ </varlistentry>
+
+ <varlistentry id="opt.sort-by" xreflabel="--sort-by=string">
+ <term>
+ <option><![CDATA[--sort-by=<string> [default: max-bytes-live] ]]></option>
+ </term>
+ <listitem>
+ <para>At the end of the run, DHAT sorts the accumulated
+ allocation points according to some metric, and shows the
+ highest scoring entries. <varname>--sort-by</varname>
+ selects the metric used for sorting:</para>
+ <para><varname>max-bytes-live </varname> maximum live bytes [default]</para>
+ <para><varname>tot-bytes-allocd </varname> total allocation (turnover)</para>
+ <para><varname>max-blocks-live </varname> maximum live blocks</para>
+ <para>This controls the order in which allocation points are
+ displayed. You can choose to look at allocation points with
+ the highest maximum liveness, or the highest total turnover, or
+ by the highest number of live blocks. These give usefully
+ different pictures of program behaviour. For example, sorting
+ by maximum live blocks tends to show up allocation points
+ creating large numbers of small objects.</para>
+ </listitem>
+ </varlistentry>
+
+</variablelist>
+
+<para>One important point to note is that each allocation stack counts
+as a seperate allocation point. Because stacks by default have 12
+frames, this tends to spread data out over multiple allocation points.
+You may want to use the flag --num-callers=4 or some such small
+number, to reduce the spreading.</para>
+
+<!-- end of xi:include in the manpage -->
+
+</sect1>
+
+</chapter>