[llvm-commits] [llvm] r127868 - /llvm/trunk/lib/CodeGen/InlineSpiller.cpp

Jakob Stoklund Olesen stoklund at 2pi.dk
Thu Mar 17 21:23:07 PDT 2011


Author: stoklund
Date: Thu Mar 17 23:23:06 2011
New Revision: 127868

URL: http://llvm.org/viewvc/llvm-project?rev=127868&view=rev
Log:
Hoist spills when the same value is known to be in less loopy sibling registers.

Stack slot real estate is virtually free compared to registers, so it is
advantageous to spill earlier even though the same value is now kept in both a
register and a stack slot.

Also eliminate redundant spills by extending the stack slot live range
underneath reloaded registers.

This can trigger a dead code elimination, removing copies and even reloads that
were only feeding spills.

Modified:
    llvm/trunk/lib/CodeGen/InlineSpiller.cpp

Modified: llvm/trunk/lib/CodeGen/InlineSpiller.cpp
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/CodeGen/InlineSpiller.cpp?rev=127868&r1=127867&r2=127868&view=diff
==============================================================================
--- llvm/trunk/lib/CodeGen/InlineSpiller.cpp (original)
+++ llvm/trunk/lib/CodeGen/InlineSpiller.cpp Thu Mar 17 23:23:06 2011
@@ -83,7 +83,11 @@
   };
 
   // Values in RegsToSpill defined by sibling copies.
-  DenseMap<VNInfo*, SibValueInfo> SibValues;
+  typedef DenseMap<VNInfo*, SibValueInfo> SibValueMap;
+  SibValueMap SibValues;
+
+  // Dead defs generated during spilling.
+  SmallVector<MachineInstr*, 8> DeadDefs;
 
   ~InlineSpiller() {}
 
@@ -110,9 +114,18 @@
   bool isSnippet(const LiveInterval &SnipLI);
   void collectRegsToSpill();
 
+  bool isRegToSpill(unsigned Reg) {
+    return std::find(RegsToSpill.begin(),
+                     RegsToSpill.end(), Reg) != RegsToSpill.end();
+  }
+
+  bool isSibling(unsigned Reg);
   void traceSiblingValue(unsigned, VNInfo*, VNInfo*);
   void analyzeSiblingValues();
 
+  bool hoistSpill(LiveInterval &SpillLI, MachineInstr *CopyMI);
+  void eliminateRedundantSpills(unsigned Reg, VNInfo *VNI);
+
   bool reMaterializeFor(MachineBasicBlock::iterator MI);
   void reMaterializeAll();
 
@@ -225,18 +238,13 @@
   for (MachineRegisterInfo::reg_iterator RI = MRI.reg_begin(Reg);
        MachineInstr *MI = RI.skipInstruction();) {
     unsigned SnipReg = isFullCopyOf(MI, Reg);
-    if (!SnipReg)
-      continue;
-    if (!TargetRegisterInfo::isVirtualRegister(SnipReg))
-      continue;
-    if (VRM.getOriginal(SnipReg) != Original)
+    if (!isSibling(SnipReg))
       continue;
     LiveInterval &SnipLI = LIS.getInterval(SnipReg);
     if (!isSnippet(SnipLI))
       continue;
     SnippetCopies.insert(MI);
-    if (std::find(RegsToSpill.begin(), RegsToSpill.end(),
-                  SnipReg) == RegsToSpill.end())
+    if (!isRegToSpill(SnipReg))
       RegsToSpill.push_back(SnipReg);
 
     DEBUG(dbgs() << "\talso spill snippet " << SnipLI << '\n');
@@ -256,6 +264,11 @@
 // the value has already been spilled, or we may want to hoist the spill from a
 // loop.
 
+bool InlineSpiller::isSibling(unsigned Reg) {
+  return TargetRegisterInfo::isVirtualRegister(Reg) &&
+           VRM.getOriginal(Reg) == Original;
+}
+
 /// traceSiblingValue - Trace a value that is about to be spilled back to the
 /// real defining instructions by looking through sibling copies. Always stay
 /// within the range of OrigVNI so the registers are known to carry the same
@@ -271,14 +284,13 @@
   DEBUG(dbgs() << "Tracing value " << PrintReg(UseReg) << ':'
                << UseVNI->id << '@' << UseVNI->def << '\n');
   SmallPtrSet<VNInfo*, 8> Visited;
-  SmallVector<std::pair<unsigned,VNInfo*>, 8> WorkList;
+  SmallVector<std::pair<unsigned, VNInfo*>, 8> WorkList;
   WorkList.push_back(std::make_pair(UseReg, UseVNI));
 
   // Best spill candidate seen so far. This must dominate UseVNI.
   SibValueInfo SVI(UseReg, UseVNI);
   MachineBasicBlock *UseMBB = LIS.getMBBFromIndex(UseVNI->def);
-  MachineBasicBlock *SpillMBB = UseMBB;
-  unsigned SpillDepth = Loops.getLoopDepth(SpillMBB);
+  unsigned SpillDepth = Loops.getLoopDepth(UseMBB);
   bool SeenOrigPHI = false; // Original PHI met.
 
   do {
@@ -289,13 +301,9 @@
       continue;
 
     // Is this value a better spill candidate?
-    if (VNI != SVI.SpillVNI) {
+    if (!isRegToSpill(Reg)) {
       MachineBasicBlock *MBB = LIS.getMBBFromIndex(VNI->def);
-      if (MBB == SpillMBB) {
-        // Prefer to spill a previous value in the same block.
-        if (VNI->def < SVI.SpillVNI->def)
-          SVI.SpillReg = Reg, SVI.SpillVNI = VNI;
-      } else if (MDT.dominates(MBB, UseMBB)) {
+      if (MBB != UseMBB && MDT.dominates(MBB, UseMBB)) {
         // This is a valid spill location dominating UseVNI.
         // Prefer to spill at a smaller loop depth.
         unsigned Depth = Loops.getLoopDepth(MBB);
@@ -304,7 +312,6 @@
                        << ':' << VNI->id << '@' << VNI->def << '\n');
           SVI.SpillReg = Reg;
           SVI.SpillVNI = VNI;
-          SpillMBB = MBB;
           SpillDepth = Depth;
         }
       }
@@ -335,8 +342,7 @@
 
     // Trace through sibling copies.
     if (unsigned SrcReg = isFullCopyOf(MI, Reg)) {
-      if (TargetRegisterInfo::isVirtualRegister(SrcReg) &&
-          VRM.getOriginal(SrcReg) == Original) {
+      if (isSibling(SrcReg)) {
         LiveInterval &SrcLI = LIS.getInterval(SrcReg);
         VNInfo *SrcVNI = SrcLI.getVNInfoAt(VNI->def.getUseIndex());
         assert(SrcVNI && "Copy from non-existing value");
@@ -406,6 +412,108 @@
   }
 }
 
+/// hoistSpill - Given a sibling copy that defines a value to be spilled, insert
+/// a spill at a better location.
+bool InlineSpiller::hoistSpill(LiveInterval &SpillLI, MachineInstr *CopyMI) {
+  SlotIndex Idx = LIS.getInstructionIndex(CopyMI);
+  VNInfo *VNI = SpillLI.getVNInfoAt(Idx.getDefIndex());
+  assert(VNI && VNI->def == Idx.getDefIndex() && "Not defined by copy");
+  SibValueMap::const_iterator I = SibValues.find(VNI);
+  if (I == SibValues.end())
+    return false;
+
+  const SibValueInfo &SVI = I->second;
+
+  // Let the normal folding code deal with the boring case.
+  if (!SVI.AllDefsAreReloads && SVI.SpillVNI == VNI)
+    return false;
+
+  // Conservatively extend the stack slot range to the range of the original
+  // value. We may be able to do better with stack slot coloring by being more
+  // careful here.
+  LiveInterval &StackInt = LSS.getInterval(StackSlot);
+  LiveInterval &OrigLI = LIS.getInterval(Original);
+  VNInfo *OrigVNI = OrigLI.getVNInfoAt(Idx);
+  StackInt.MergeValueInAsValue(OrigLI, OrigVNI, StackInt.getValNumInfo(0));
+
+  // Already spilled everywhere.
+  if (SVI.AllDefsAreReloads)
+    return true;
+
+  // We are going to spill SVI.SpillVNI immediately after its def, so clear out
+  // any later spills of the same value.
+  eliminateRedundantSpills(SVI.SpillReg, SVI.SpillVNI);
+
+  MachineBasicBlock *MBB = LIS.getMBBFromIndex(SVI.SpillVNI->def);
+  MachineBasicBlock::iterator MII;
+  if (SVI.SpillVNI->isPHIDef())
+    MII = MBB->SkipPHIsAndLabels(MBB->begin());
+  else {
+    MII = LIS.getInstructionFromIndex(SVI.SpillVNI->def);
+    ++MII;
+  }
+  // Insert spill without kill flag immediately after def.
+  TII.storeRegToStackSlot(*MBB, MII, SVI.SpillReg, false, StackSlot, RC, &TRI);
+  --MII; // Point to store instruction.
+  LIS.InsertMachineInstrInMaps(MII);
+  VRM.addSpillSlotUse(StackSlot, MII);
+  DEBUG(dbgs() << "\thoisted: " << SVI.SpillVNI->def << '\t' << *MII);
+  return true;
+}
+
+/// eliminateRedundantSpills - Reg:VNI is known to be on the stack. Remove any
+/// redundant spills of this value in Reg and sibling copies.
+void InlineSpiller::eliminateRedundantSpills(unsigned Reg, VNInfo *VNI) {
+  SmallVector<std::pair<unsigned, VNInfo*>, 8> WorkList;
+  WorkList.push_back(std::make_pair(Reg, VNI));
+  LiveInterval &StackInt = LSS.getInterval(StackSlot);
+
+  do {
+    tie(Reg, VNI) = WorkList.pop_back_val();
+    DEBUG(dbgs() << "Checking redundant spills for " << PrintReg(Reg) << ':'
+                 << VNI->id << '@' << VNI->def << '\n');
+
+    // Regs to spill are taken care of.
+    if (isRegToSpill(Reg))
+      continue;
+
+    // Add all of VNI's live range to StackInt.
+    LiveInterval &LI = LIS.getInterval(Reg);
+    StackInt.MergeValueInAsValue(LI, VNI, StackInt.getValNumInfo(0));
+
+    // Find all spills and copies of VNI.
+    for (MachineRegisterInfo::use_nodbg_iterator UI = MRI.use_nodbg_begin(Reg);
+         MachineInstr *MI = UI.skipInstruction();) {
+      if (!MI->isCopy() && !MI->getDesc().mayStore())
+        continue;
+      SlotIndex Idx = LIS.getInstructionIndex(MI);
+      if (LI.getVNInfoAt(Idx) != VNI)
+        continue;
+
+      // Follow sibling copies down the dominator tree.
+      if (unsigned DstReg = isFullCopyOf(MI, Reg)) {
+        if (isSibling(DstReg)) {
+           LiveInterval &DstLI = LIS.getInterval(DstReg);
+           VNInfo *DstVNI = DstLI.getVNInfoAt(Idx.getDefIndex());
+           assert(DstVNI && "Missing defined value");
+           assert(DstVNI->def == Idx.getDefIndex() && "Wrong copy def slot");
+           WorkList.push_back(std::make_pair(DstReg, DstVNI));
+        }
+        continue;
+      }
+
+      // Erase spills.
+      int FI;
+      if (Reg == TII.isStoreToStackSlot(MI, FI) && FI == StackSlot) {
+        DEBUG(dbgs() << "Redundant spill " << Idx << '\t' << *MI);
+        // eliminateDeadDefs won't normally remove stores, so switch opcode.
+        MI->setDesc(TII.get(TargetOpcode::KILL));
+        DeadDefs.push_back(MI);
+      }
+    }
+  } while (!WorkList.empty());
+}
+
 /// reMaterializeFor - Attempt to rematerialize before MI instead of reloading.
 bool InlineSpiller::reMaterializeFor(MachineBasicBlock::iterator MI) {
   SlotIndex UseIdx = LIS.getInstructionIndex(MI).getUseIndex();
@@ -679,6 +787,19 @@
     SmallVector<unsigned, 8> Ops;
     tie(Reads, Writes) = MI->readsWritesVirtualRegister(Reg, &Ops);
 
+    // Check for a sibling copy.
+    unsigned SibReg = isFullCopyOf(MI, Reg);
+    if (!isSibling(SibReg))
+      SibReg = 0;
+
+    // Hoist the spill of a sib-reg copy.
+    if (SibReg && Writes && !Reads && hoistSpill(OldLI, MI)) {
+      // This COPY is now dead, the value is already in the stack slot.
+      MI->getOperand(0).setIsDead();
+      DeadDefs.push_back(MI);
+      continue;
+    }
+
     // Attempt to fold memory ops.
     if (foldMemoryOperand(MI, Ops))
       continue;
@@ -728,11 +849,10 @@
                << LIS.getInterval(Original) << '\n');
   assert(edit.getParent().isSpillable() &&
          "Attempting to spill already spilled value.");
+  assert(DeadDefs.empty() && "Previous spill didn't remove dead defs");
 
   collectRegsToSpill();
-
   analyzeSiblingValues();
-
   reMaterializeAll();
 
   // Remat may handle everything.
@@ -759,6 +879,12 @@
   for (unsigned i = 0, e = RegsToSpill.size(); i != e; ++i)
     spillAroundUses(RegsToSpill[i]);
 
+  // Hoisted spills may cause dead code.
+  if (!DeadDefs.empty()) {
+    DEBUG(dbgs() << "Eliminating " << DeadDefs.size() << " dead defs\n");
+    Edit->eliminateDeadDefs(DeadDefs, LIS, VRM, TII);
+  }
+
   // Finally delete the SnippetCopies.
   for (MachineRegisterInfo::reg_iterator RI = MRI.reg_begin(edit.getReg());
        MachineInstr *MI = RI.skipInstruction();) {





More information about the llvm-commits mailing list