[PATCH] D13259: LLE 6/6: Add LoopLoadElimination pass

Adam Nemet via llvm-commits llvm-commits at lists.llvm.org
Tue Sep 29 10:28:58 PDT 2015


anemet created this revision.
anemet added reviewers: hfinkel, dberlin.
anemet added a subscriber: llvm-commits.
Herald added a subscriber: sanjoy.

The goal of this pass is to perform store-to-load forwarding across the
backedge of a loop.  E.g.:

for (i)
   A[i + 1] = A[i] + B[i]

=>

T = A[0]
for (i)
   T = T + B[i]
   A[i + 1] = T

The pass relies on loop dependence analysis via LoopAccessAnalisys to
find opportunities of loop-carried dependences with a distance of one
between a store and a load.  Since it's using LoopAccessAnalysis, it was
easy to also add support for versioning away may-aliasing intervening
stores that would otherwise prevent this transformation.

This optimization is also performed by Load-PRE in GVN without the
option of multi-versioning.  As was discussed with Daniel Berlin in
http://reviews.llvm.org/D9548, this is inferior to a more loop-aware
solution applied here.  Hopefully, we will be able to remove some
complexity from GVN/MemorySSA as a consequence.

In the long run, we may want to extend this pass (or create a new one if
there is little overlap) to also eliminate loop-indepedent redundant
loads and store that *require* versioning due to may-aliasing
intervening stores/loads.  I have some motivating cases for store
elimination. My plan right now is to wait for MemorySSA to come online
first rather than using memdep for this.

The main motiviation for this pass is the 456.hmmer loop in SPECint2006
where after distributing the original loop and vectorizing the top part,
we are left with the critical path exposed in the bottom loop.  Being
able to promote the memory dependence into a register depedence (even
though the HW does perform store-to-load fowarding as well) results in a
major gain (~20%).  This gain also transfers over to x86: it's
around 8-10%.

Right now the pass is off by default and can be enabled
with -enable-loop-load-elim.  On the LNT testsuite, there are two
performance changes (negative number -> improvement):

  1. -28% in Polybench/linear-algebra/solvers/dynprog: the length of the
     critical paths is reduced
  2. +2% in Polybench/stencils/adi: Unfortunately, I couldn't reproduce this
     outside of LNT

The pass is scheduled after the loop vectorizer (which is after loop
distribution).  The rational is to try to reuse LAA state, rather than
recomputing it.  The order between LV and LLE is not critical because
normally LV does not touch scalar st->ld forwarding cases where
vectorizing would inhibit the CPU's st->ld forwarding to kick in.

LoopLoadElimination requires LAA to provide the full set of dependences
(including forward dependences).  LAA is known to omit loop-independent
dependences in certain situations.  The big comment before
removeDependencesFromMultipleStores explains why this should not occur
for the cases that we're interested in.

http://reviews.llvm.org/D13259

Files:
  include/llvm/Analysis/LoopAccessAnalysis.h
  include/llvm/InitializePasses.h
  include/llvm/Transforms/Scalar.h
  lib/Transforms/IPO/PassManagerBuilder.cpp
  lib/Transforms/Scalar/CMakeLists.txt
  lib/Transforms/Scalar/LoopLoadElimination.cpp
  lib/Transforms/Scalar/Scalar.cpp
  test/Transforms/LoopLoadElim/backward.ll
  test/Transforms/LoopLoadElim/def-store-before-load.ll
  test/Transforms/LoopLoadElim/forward.ll
  test/Transforms/LoopLoadElim/memcheck.ll
  test/Transforms/LoopLoadElim/multiple-stores-same-block.ll
  test/Transforms/LoopLoadElim/unknown-dep.ll

-------------- next part --------------
A non-text attachment was scrubbed...
Name: D13259.36003.patch
Type: text/x-patch
Size: 36996 bytes
Desc: not available
URL: <http://lists.llvm.org/pipermail/llvm-commits/attachments/20150929/01ff41f8/attachment.bin>


More information about the llvm-commits mailing list