[llvm-dev] RFC: EfficiencySanitizer
Filipe Cabecinhas via llvm-dev
llvm-dev at lists.llvm.org
Tue Apr 19 10:18:14 PDT 2016
Thanks for proposing this. It seems like it might be an interesting
tool for us too. But this proposal seems a bit hand-wavy, and I think
it's missing some crucial info before we start heading this way.
On Sun, Apr 17, 2016 at 10:46 PM, Derek Bruening via llvm-dev
<llvm-dev at lists.llvm.org> wrote:
> TL;DR: We plan to build a suite of compiler-based dynamic instrumentation
> tools for analyzing targeted performance problems. These tools will all
> live under a new "EfficiencySanitizer" (or "esan") sanitizer umbrella, as
> they will share significant portions of their implementations.
I will bike-shed the name as much as anyone else, but I'll stay away for now :-)
> Our goal is to build a suite of dynamic instrumentation tools for analyzing
> particular performance problems that are difficult to evaluate using other
> profiling methods. Modern hardware performance counters provide insight
> into where time is spent and when micro-architectural events such as cache
> misses are occurring, but they are of limited effectiveness for contextual
> analysis: it is not easy to answer *why* a cache miss occurred.
> Examples of tools that we have planned include: identifying wasted or
> redundant computation, identifying cache fragmentation, and measuring
> working sets. See more details on these below.
> We believe that tools with overhead beyond about 5x are simply too
> heavyweight to easily apply to large, industrial-sized applications running
> real-world workloads. Our goal is for our tools to gather useful
> information with overhead less than 5x, and ideally closer to 3x, to
> facilitate deployment. We would prefer to trade off accuracy and build a
> less-accurate tool below our overhead ceiling than to build a high-accuracy
> but slow tool. We hope to hit a sweet spot of tools that gather trace-based
> contextual information not feasible with pure sampling yet are still
> practical to deploy.
This is also very important for us, and there are sanitizers which are
"harder to sell" because of the overhead.
> In a similar vein, we would prefer a targeted tool that analyzes one
> particular aspect of performance with low overhead than a more general tool
> that can answer more questions but has high overhead.
> Dynamic binary instrumentation is one option for these types of tools, but
> typically compiler-based instrumentation provides better performance, and we
> intend to focus only on analyzing applications for which source code is
> available. Studying instruction cache behavior with compiler
> instrumentation can be challenging, however, so we plan to at least
> initially focus on data performance.
> Many of our planned tools target specific performance issues with data
> accesses. They employ the technique of *shadow memory* to store metadata
> about application data references, using the compiler to instrument loads
> and stores with code to update the shadow memory. A companion runtime
> library intercepts libc calls if necessary to update shadow memory on
> non-application data references. The runtime library also intercepts heap
> allocations and other key events in order to perform its analyses. This is
> all very similar to how existing sanitizers such as AddressSanitizer,
> ThreadSanitizer, MemorySanitizer, etc. operate today.
> Example Tools
> We have several initial tools that we plan to build. These are not
> necessarily novel ideas on their own: some of these have already been
> explored in academia. The idea is to create practical, low-overhead,
> robust, and publicly available versions of these tools.
> *Cache fragmentation*: this tool gather data structure field hotness
> information, looking for data layout optimization opportunities by grouping
> hot fields together to avoid data cache fragmentation. Future enhancements
> may add field affinity information if it can be computed with low enough
> *Working set measurement*: this tool measures the data working set size of
> an application at each snapshot during execution. It can help to understand
> phased behavior as well as providing basic direction for further effort by
> the developer: e.g., knowing whether the working set is close to fitting in
> current L3 caches or is many times larger can help determine where to spend
> *Dead store detection*: this tool identifies dead stores (write-after-write
> patterns with no intervening read) as well as redundant stores (writes of
> the same value already in memory). Xref the Deadspy paper from CGO 2012.
> *Single-reference*: this tool identifies data cache lines brought in but
> only read once. These could be candidates for non-temporal loads.
Do you have any estimates on memory overhead (both memory usage
(+shadow), and code size) you expect? As well as estimated about the
At least for the tools you are currently starting to implement, it
would be nice to have some estimates and plans on what is going to
I would actually like to see a small RFC about each tool and what the
plan (overhead/slowdown, pseudo-code for instrumentation, UX, ...) is
before starting to commit.
I don't expect the plan to be very detailed, nor for everything to be
pinned down, of course. This seems to be a bit at a research stage,
and I'm totally ok with it. But would rather it not be as opaque as it
is now. The way the tools will report problems/data/etc, is important,
> We are proposing the name EfficiencySanitizer, or "esan" for short, to refer
> to this suite of dynamic instrumentation tools for improving program
> efficiency. As we have a number of different tools that share quite a bit
> of their implementation we plan to consider them sub-tools under the
> EfficiencySanitizer umbrella, rather than adding a whole bunch of separate
> instrumentation and runtime library components.
How much code is expected to be shared? Most? Similar to what the
sanitizers already share?
Do we expect the shadow memory mapping to be (mostly) the same among
all esan's tools?
Do you already have an idea of a few tools + the type of code that
would be shared (for example: "read/write instrumentation is mostly
the same among these tools", or "generating reports for these tools is
mostly the same", or something similar)?
> While these tools are not addressing correctness issues like other
> sanitizers, they will be sharing a lot of the existing sanitizer runtime
> library support code. Furthermore, users are already familiar with the
> sanitizer brand, and it seems better to extend that concept rather than add
> some new term.
Not the email to bike-shed the name, but I don't like "Efficiency"
that much here :-)
Thank you for working on this,
More information about the llvm-dev