[llvm] r293737 - [CodeGen] Move MacroFusion to the target
Davide Italiano via llvm-commits
llvm-commits at lists.llvm.org
Fri Feb 17 18:47:15 PST 2017
GCC6 is complaining, can you please take a look?
[267/704] Building CXX object
lib/Target/AArch64/CMakeFiles/LLVMAArch64CodeGen.dir/AArch64MacroFusion.cpp.o
../lib/Target/AArch64/AArch64MacroFusion.cpp: In function ‘bool
{anonymous}::shouldScheduleAdjacent(const llvm::AArch64In
strInfo&, const llvm::AArch64Subtarget&, const llvm::MachineInstr*,
const llvm::MachineInstr*)’:
../lib/Target/AArch64/AArch64MacroFusion.cpp:37:32: warning: enumeral
and non-enumeral type in conditional expression [-W
extra]
unsigned FirstOpcode = First ?
~~~~~~^
First->getOpcode() : AArch64::INSTRUCTION_LIST_END;
~~~~~~~~~~~~~~~~~~~~~~~~~~~~
../lib/Target/AArch64/AArch64MacroFusion.cpp:39:34: warning: enumeral
and non-enumeral type in conditional expression [-$
extra]
unsigned SecondOpcode = Second ?
~~~~~~~^
Second->getOpcode() : AArch64::INSTRUCTION_LIST_END;
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
On Tue, Jan 31, 2017 at 6:54 PM, Evandro Menezes via llvm-commits
<llvm-commits at lists.llvm.org> wrote:
> Author: evandro
> Date: Tue Jan 31 20:54:34 2017
> New Revision: 293737
>
> URL: http://llvm.org/viewvc/llvm-project?rev=293737&view=rev
> Log:
> [CodeGen] Move MacroFusion to the target
>
> This patch moves the class for scheduling adjacent instructions,
> MacroFusion, to the target.
>
> In AArch64, it also expands the fusion to all instructions pairs in a
> scheduling block, beyond just among the predecessors of the branch at the
> end.
>
> Differential revision: https://reviews.llvm.org/D28489
>
> Added:
> llvm/trunk/lib/Target/AArch64/AArch64MacroFusion.cpp
> llvm/trunk/lib/Target/AArch64/AArch64MacroFusion.h
> llvm/trunk/lib/Target/X86/X86MacroFusion.cpp
> llvm/trunk/lib/Target/X86/X86MacroFusion.h
> Modified:
> llvm/trunk/include/llvm/CodeGen/MachineScheduler.h
> llvm/trunk/include/llvm/Target/TargetInstrInfo.h
> llvm/trunk/lib/CodeGen/MachineScheduler.cpp
> llvm/trunk/lib/Target/AArch64/AArch64InstrInfo.cpp
> llvm/trunk/lib/Target/AArch64/AArch64InstrInfo.h
> llvm/trunk/lib/Target/AArch64/AArch64TargetMachine.cpp
> llvm/trunk/lib/Target/AArch64/CMakeLists.txt
> llvm/trunk/lib/Target/X86/CMakeLists.txt
> llvm/trunk/lib/Target/X86/X86InstrInfo.cpp
> llvm/trunk/lib/Target/X86/X86InstrInfo.h
> llvm/trunk/lib/Target/X86/X86TargetMachine.cpp
> llvm/trunk/test/CodeGen/AArch64/misched-fusion.ll
>
> Modified: llvm/trunk/include/llvm/CodeGen/MachineScheduler.h
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/include/llvm/CodeGen/MachineScheduler.h?rev=293737&r1=293736&r2=293737&view=diff
> ==============================================================================
> --- llvm/trunk/include/llvm/CodeGen/MachineScheduler.h (original)
> +++ llvm/trunk/include/llvm/CodeGen/MachineScheduler.h Tue Jan 31 20:54:34 2017
> @@ -1033,9 +1033,6 @@ createStoreClusterDAGMutation(const Targ
> const TargetRegisterInfo *TRI);
>
> std::unique_ptr<ScheduleDAGMutation>
> -createMacroFusionDAGMutation(const TargetInstrInfo *TII);
> -
> -std::unique_ptr<ScheduleDAGMutation>
> createCopyConstrainDAGMutation(const TargetInstrInfo *TII,
> const TargetRegisterInfo *TRI);
>
>
> Modified: llvm/trunk/include/llvm/Target/TargetInstrInfo.h
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/include/llvm/Target/TargetInstrInfo.h?rev=293737&r1=293736&r2=293737&view=diff
> ==============================================================================
> --- llvm/trunk/include/llvm/Target/TargetInstrInfo.h (original)
> +++ llvm/trunk/include/llvm/Target/TargetInstrInfo.h Tue Jan 31 20:54:34 2017
> @@ -1070,15 +1070,6 @@ public:
> llvm_unreachable("target did not implement shouldClusterMemOps()");
> }
>
> - /// Can this target fuse the given instructions if they are scheduled
> - /// adjacent. Note that you have to add:
> - /// DAG.addMutation(createMacroFusionDAGMutation());
> - /// to TargetPassConfig::createMachineScheduler() to have an effect.
> - virtual bool shouldScheduleAdjacent(const MachineInstr &First,
> - const MachineInstr &Second) const {
> - llvm_unreachable("target did not implement shouldScheduleAdjacent()");
> - }
> -
> /// Reverses the branch condition of the specified condition list,
> /// returning false on success and true if it cannot be reversed.
> virtual
>
> Modified: llvm/trunk/lib/CodeGen/MachineScheduler.cpp
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/CodeGen/MachineScheduler.cpp?rev=293737&r1=293736&r2=293737&view=diff
> ==============================================================================
> --- llvm/trunk/lib/CodeGen/MachineScheduler.cpp (original)
> +++ llvm/trunk/lib/CodeGen/MachineScheduler.cpp Tue Jan 31 20:54:34 2017
> @@ -80,10 +80,6 @@ static cl::opt<bool> EnableMemOpCluster(
> cl::desc("Enable memop clustering."),
> cl::init(true));
>
> -// Experimental heuristics
> -static cl::opt<bool> EnableMacroFusion("misched-fusion", cl::Hidden,
> - cl::desc("Enable scheduling for macro fusion."), cl::init(true));
> -
> static cl::opt<bool> VerifyScheduling("verify-misched", cl::Hidden,
> cl::desc("Verify machine instrs before and after machine scheduling"));
>
> @@ -1544,76 +1540,6 @@ void BaseMemOpClusterMutation::apply(Sch
> }
>
> //===----------------------------------------------------------------------===//
> -// MacroFusion - DAG post-processing to encourage fusion of macro ops.
> -//===----------------------------------------------------------------------===//
> -
> -namespace {
> -/// \brief Post-process the DAG to create cluster edges between instructions
> -/// that may be fused by the processor into a single operation.
> -class MacroFusion : public ScheduleDAGMutation {
> - const TargetInstrInfo &TII;
> -public:
> - MacroFusion(const TargetInstrInfo &TII)
> - : TII(TII) {}
> -
> - void apply(ScheduleDAGInstrs *DAGInstrs) override;
> -};
> -} // anonymous
> -
> -namespace llvm {
> -
> -std::unique_ptr<ScheduleDAGMutation>
> -createMacroFusionDAGMutation(const TargetInstrInfo *TII) {
> - return EnableMacroFusion ? make_unique<MacroFusion>(*TII) : nullptr;
> -}
> -
> -} // namespace llvm
> -
> -/// \brief Callback from DAG postProcessing to create cluster edges to encourage
> -/// fused operations.
> -void MacroFusion::apply(ScheduleDAGInstrs *DAGInstrs) {
> - ScheduleDAGMI *DAG = static_cast<ScheduleDAGMI*>(DAGInstrs);
> -
> - // For now, assume targets can only fuse with the branch.
> - SUnit &ExitSU = DAG->ExitSU;
> - MachineInstr *Branch = ExitSU.getInstr();
> - if (!Branch)
> - return;
> -
> - for (SDep &PredDep : ExitSU.Preds) {
> - if (PredDep.isWeak())
> - continue;
> - SUnit &SU = *PredDep.getSUnit();
> - MachineInstr &Pred = *SU.getInstr();
> - if (!TII.shouldScheduleAdjacent(Pred, *Branch))
> - continue;
> -
> - // Create a single weak edge from SU to ExitSU. The only effect is to cause
> - // bottom-up scheduling to heavily prioritize the clustered SU. There is no
> - // need to copy predecessor edges from ExitSU to SU, since top-down
> - // scheduling cannot prioritize ExitSU anyway. To defer top-down scheduling
> - // of SU, we could create an artificial edge from the deepest root, but it
> - // hasn't been needed yet.
> - bool Success = DAG->addEdge(&ExitSU, SDep(&SU, SDep::Cluster));
> - (void)Success;
> - assert(Success && "No DAG nodes should be reachable from ExitSU");
> -
> - // Adjust latency of data deps between the nodes.
> - for (SDep &PredDep : ExitSU.Preds) {
> - if (PredDep.getSUnit() == &SU)
> - PredDep.setLatency(0);
> - }
> - for (SDep &SuccDep : SU.Succs) {
> - if (SuccDep.getSUnit() == &ExitSU)
> - SuccDep.setLatency(0);
> - }
> -
> - DEBUG(dbgs() << "Macro Fuse SU(" << SU.NodeNum << ")\n");
> - break;
> - }
> -}
> -
> -//===----------------------------------------------------------------------===//
> // CopyConstrain - DAG post-processing to encourage copy elimination.
> //===----------------------------------------------------------------------===//
>
>
> Modified: llvm/trunk/lib/Target/AArch64/AArch64InstrInfo.cpp
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/Target/AArch64/AArch64InstrInfo.cpp?rev=293737&r1=293736&r2=293737&view=diff
> ==============================================================================
> --- llvm/trunk/lib/Target/AArch64/AArch64InstrInfo.cpp (original)
> +++ llvm/trunk/lib/Target/AArch64/AArch64InstrInfo.cpp Tue Jan 31 20:54:34 2017
> @@ -1914,88 +1914,6 @@ bool AArch64InstrInfo::shouldClusterMemO
> return Offset1 + 1 == Offset2;
> }
>
> -bool AArch64InstrInfo::shouldScheduleAdjacent(
> - const MachineInstr &First, const MachineInstr &Second) const {
> - if (Subtarget.hasArithmeticBccFusion()) {
> - // Fuse CMN, CMP, TST followed by Bcc.
> - unsigned SecondOpcode = Second.getOpcode();
> - if (SecondOpcode == AArch64::Bcc) {
> - switch (First.getOpcode()) {
> - default:
> - return false;
> - case AArch64::ADDSWri:
> - case AArch64::ADDSWrr:
> - case AArch64::ADDSXri:
> - case AArch64::ADDSXrr:
> - case AArch64::ANDSWri:
> - case AArch64::ANDSWrr:
> - case AArch64::ANDSXri:
> - case AArch64::ANDSXrr:
> - case AArch64::SUBSWri:
> - case AArch64::SUBSWrr:
> - case AArch64::SUBSXri:
> - case AArch64::SUBSXrr:
> - case AArch64::BICSWrr:
> - case AArch64::BICSXrr:
> - return true;
> - case AArch64::ADDSWrs:
> - case AArch64::ADDSXrs:
> - case AArch64::ANDSWrs:
> - case AArch64::ANDSXrs:
> - case AArch64::SUBSWrs:
> - case AArch64::SUBSXrs:
> - case AArch64::BICSWrs:
> - case AArch64::BICSXrs:
> - // Shift value can be 0 making these behave like the "rr" variant...
> - return !hasShiftedReg(Second);
> - }
> - }
> - }
> - if (Subtarget.hasArithmeticCbzFusion()) {
> - // Fuse ALU operations followed by CBZ/CBNZ.
> - unsigned SecondOpcode = Second.getOpcode();
> - if (SecondOpcode == AArch64::CBNZW || SecondOpcode == AArch64::CBNZX ||
> - SecondOpcode == AArch64::CBZW || SecondOpcode == AArch64::CBZX) {
> - switch (First.getOpcode()) {
> - default:
> - return false;
> - case AArch64::ADDWri:
> - case AArch64::ADDWrr:
> - case AArch64::ADDXri:
> - case AArch64::ADDXrr:
> - case AArch64::ANDWri:
> - case AArch64::ANDWrr:
> - case AArch64::ANDXri:
> - case AArch64::ANDXrr:
> - case AArch64::EORWri:
> - case AArch64::EORWrr:
> - case AArch64::EORXri:
> - case AArch64::EORXrr:
> - case AArch64::ORRWri:
> - case AArch64::ORRWrr:
> - case AArch64::ORRXri:
> - case AArch64::ORRXrr:
> - case AArch64::SUBWri:
> - case AArch64::SUBWrr:
> - case AArch64::SUBXri:
> - case AArch64::SUBXrr:
> - return true;
> - case AArch64::ADDWrs:
> - case AArch64::ADDXrs:
> - case AArch64::ANDWrs:
> - case AArch64::ANDXrs:
> - case AArch64::SUBWrs:
> - case AArch64::SUBXrs:
> - case AArch64::BICWrs:
> - case AArch64::BICXrs:
> - // Shift value can be 0 making these behave like the "rr" variant...
> - return !hasShiftedReg(Second);
> - }
> - }
> - }
> - return false;
> -}
> -
> MachineInstr *AArch64InstrInfo::emitFrameIndexDebugValue(
> MachineFunction &MF, int FrameIx, uint64_t Offset, const MDNode *Var,
> const MDNode *Expr, const DebugLoc &DL) const {
>
> Modified: llvm/trunk/lib/Target/AArch64/AArch64InstrInfo.h
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/Target/AArch64/AArch64InstrInfo.h?rev=293737&r1=293736&r2=293737&view=diff
> ==============================================================================
> --- llvm/trunk/lib/Target/AArch64/AArch64InstrInfo.h (original)
> +++ llvm/trunk/lib/Target/AArch64/AArch64InstrInfo.h Tue Jan 31 20:54:34 2017
> @@ -138,9 +138,6 @@ public:
> bool shouldClusterMemOps(MachineInstr &FirstLdSt, MachineInstr &SecondLdSt,
> unsigned NumLoads) const override;
>
> - bool shouldScheduleAdjacent(const MachineInstr &First,
> - const MachineInstr &Second) const override;
> -
> MachineInstr *emitFrameIndexDebugValue(MachineFunction &MF, int FrameIx,
> uint64_t Offset, const MDNode *Var,
> const MDNode *Expr,
>
> Added: llvm/trunk/lib/Target/AArch64/AArch64MacroFusion.cpp
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/Target/AArch64/AArch64MacroFusion.cpp?rev=293737&view=auto
> ==============================================================================
> --- llvm/trunk/lib/Target/AArch64/AArch64MacroFusion.cpp (added)
> +++ llvm/trunk/lib/Target/AArch64/AArch64MacroFusion.cpp Tue Jan 31 20:54:34 2017
> @@ -0,0 +1,209 @@
> +//===- AArch64MacroFusion.cpp - AArch64 Macro Fusion ----------------------===//
> +//
> +// The LLVM Compiler Infrastructure
> +//
> +// This file is distributed under the University of Illinois Open Source
> +// License. See LICENSE.TXT for details.
> +//
> +//===----------------------------------------------------------------------===//
> +//
> +// \file This file contains the AArch64 implementation of the DAG scheduling mutation
> +// to pair instructions back to back.
> +//
> +//===----------------------------------------------------------------------===//
> +
> +#include "AArch64MacroFusion.h"
> +#include "AArch64Subtarget.h"
> +#include "llvm/Support/CommandLine.h"
> +#include "llvm/Target/TargetInstrInfo.h"
> +
> +#define DEBUG_TYPE "misched"
> +
> +using namespace llvm;
> +
> +static cl::opt<bool> EnableMacroFusion("aarch64-misched-fusion", cl::Hidden,
> + cl::desc("Enable scheduling for macro fusion."), cl::init(true));
> +
> +namespace {
> +
> +/// \brief Verify that the instruction pair, \param First and \param Second,
> +/// should be scheduled back to back. Given an anchor instruction, if the other
> +/// instruction is unspecified, then verify that the anchor instruction may be
> +/// part of a pair at all.
> +static bool shouldScheduleAdjacent(const AArch64InstrInfo &TII,
> + const AArch64Subtarget &ST,
> + const MachineInstr *First,
> + const MachineInstr *Second) {
> + unsigned FirstOpcode = First ?
> + First->getOpcode() : AArch64::INSTRUCTION_LIST_END;
> + unsigned SecondOpcode = Second ?
> + Second->getOpcode() : AArch64::INSTRUCTION_LIST_END;
> +
> + if (ST.hasArithmeticBccFusion())
> + // Fuse CMN, CMP, TST followed by Bcc.
> + if (SecondOpcode == AArch64::Bcc)
> + switch (FirstOpcode) {
> + default:
> + return false;
> + case AArch64::ADDSWri:
> + case AArch64::ADDSWrr:
> + case AArch64::ADDSXri:
> + case AArch64::ADDSXrr:
> + case AArch64::ANDSWri:
> + case AArch64::ANDSWrr:
> + case AArch64::ANDSXri:
> + case AArch64::ANDSXrr:
> + case AArch64::SUBSWri:
> + case AArch64::SUBSWrr:
> + case AArch64::SUBSXri:
> + case AArch64::SUBSXrr:
> + case AArch64::BICSWrr:
> + case AArch64::BICSXrr:
> + return true;
> + case AArch64::ADDSWrs:
> + case AArch64::ADDSXrs:
> + case AArch64::ANDSWrs:
> + case AArch64::ANDSXrs:
> + case AArch64::SUBSWrs:
> + case AArch64::SUBSXrs:
> + case AArch64::BICSWrs:
> + case AArch64::BICSXrs:
> + // Shift value can be 0 making these behave like the "rr" variant...
> + return !TII.hasShiftedReg(*First);
> + case AArch64::INSTRUCTION_LIST_END:
> + return true;
> + }
> +
> + if (ST.hasArithmeticCbzFusion())
> + // Fuse ALU operations followed by CBZ/CBNZ.
> + if (SecondOpcode == AArch64::CBNZW || SecondOpcode == AArch64::CBNZX ||
> + SecondOpcode == AArch64::CBZW || SecondOpcode == AArch64::CBZX)
> + switch (FirstOpcode) {
> + default:
> + return false;
> + case AArch64::ADDWri:
> + case AArch64::ADDWrr:
> + case AArch64::ADDXri:
> + case AArch64::ADDXrr:
> + case AArch64::ANDWri:
> + case AArch64::ANDWrr:
> + case AArch64::ANDXri:
> + case AArch64::ANDXrr:
> + case AArch64::EORWri:
> + case AArch64::EORWrr:
> + case AArch64::EORXri:
> + case AArch64::EORXrr:
> + case AArch64::ORRWri:
> + case AArch64::ORRWrr:
> + case AArch64::ORRXri:
> + case AArch64::ORRXrr:
> + case AArch64::SUBWri:
> + case AArch64::SUBWrr:
> + case AArch64::SUBXri:
> + case AArch64::SUBXrr:
> + return true;
> + case AArch64::ADDWrs:
> + case AArch64::ADDXrs:
> + case AArch64::ANDWrs:
> + case AArch64::ANDXrs:
> + case AArch64::SUBWrs:
> + case AArch64::SUBXrs:
> + case AArch64::BICWrs:
> + case AArch64::BICXrs:
> + // Shift value can be 0 making these behave like the "rr" variant...
> + return !TII.hasShiftedReg(*First);
> + case AArch64::INSTRUCTION_LIST_END:
> + return true;
> + }
> +
> + return false;
> +}
> +
> +/// \brief Implement the fusion of instruction pairs in the scheduling
> +/// \param DAG, anchored at the instruction in \param ASU. \param Preds
> +/// indicates if its dependencies in \param APreds are predecessors instead of
> +/// successors.
> +static bool scheduleAdjacentImpl(ScheduleDAGMI *DAG, SUnit *ASU,
> + SmallVectorImpl<SDep> &APreds, bool Preds) {
> + const AArch64InstrInfo *TII = static_cast<const AArch64InstrInfo *>(DAG->TII);
> + const AArch64Subtarget &ST = DAG->MF.getSubtarget<AArch64Subtarget>();
> +
> + const MachineInstr *AMI = ASU->getInstr();
> + if (!AMI || AMI->isPseudo() || AMI->isTransient() ||
> + (Preds && !shouldScheduleAdjacent(*TII, ST, nullptr, AMI)) ||
> + (!Preds && !shouldScheduleAdjacent(*TII, ST, AMI, nullptr)))
> + return false;
> +
> + for (SDep &BDep : APreds) {
> + if (BDep.isWeak())
> + continue;
> +
> + SUnit *BSU = BDep.getSUnit();
> + const MachineInstr *BMI = BSU->getInstr();
> + if (!BMI || BMI->isPseudo() || BMI->isTransient() ||
> + (Preds && !shouldScheduleAdjacent(*TII, ST, BMI, AMI)) ||
> + (!Preds && !shouldScheduleAdjacent(*TII, ST, AMI, BMI)))
> + continue;
> +
> + // Create a single weak edge between the adjacent instrs. The only
> + // effect is to cause bottom-up scheduling to heavily prioritize the
> + // clustered instrs.
> + if (Preds)
> + DAG->addEdge(ASU, SDep(BSU, SDep::Cluster));
> + else
> + DAG->addEdge(BSU, SDep(ASU, SDep::Cluster));
> +
> + // Adjust the latency between the 1st instr and its predecessors/successors.
> + for (SDep &Dep : APreds)
> + if (Dep.getSUnit() == BSU)
> + Dep.setLatency(0);
> +
> + // Adjust the latency between the 2nd instr and its successors/predecessors.
> + auto &BSuccs = Preds ? BSU->Succs : BSU->Preds;
> + for (SDep &Dep : BSuccs)
> + if (Dep.getSUnit() == ASU)
> + Dep.setLatency(0);
> +
> + DEBUG(dbgs() << "Macro fuse ";
> + Preds ? BSU->print(dbgs(), DAG) : ASU->print(dbgs(), DAG);
> + dbgs() << " - ";
> + Preds ? ASU->print(dbgs(), DAG) : BSU->print(dbgs(), DAG);
> + dbgs() << '\n');
> +
> + return true;
> + }
> +
> + return false;
> +}
> +
> +/// \brief Post-process the DAG to create cluster edges between instructions
> +/// that may be fused by the processor into a single operation.
> +class AArch64MacroFusion : public ScheduleDAGMutation {
> +public:
> + AArch64MacroFusion() {}
> +
> + void apply(ScheduleDAGInstrs *DAGInstrs) override;
> +};
> +
> +void AArch64MacroFusion::apply(ScheduleDAGInstrs *DAGInstrs) {
> + ScheduleDAGMI *DAG = static_cast<ScheduleDAGMI*>(DAGInstrs);
> +
> + // For each of the SUnits in the scheduling block, try to fuse the instruction
> + // in it with one in its successors.
> + for (SUnit &ASU : DAG->SUnits)
> + scheduleAdjacentImpl(DAG, &ASU, ASU.Succs, false);
> +
> + // Try to fuse the instruction in the ExitSU with one in its predecessors.
> + scheduleAdjacentImpl(DAG, &DAG->ExitSU, DAG->ExitSU.Preds, true);
> +}
> +
> +} // end namespace
> +
> +
> +namespace llvm {
> +
> +std::unique_ptr<ScheduleDAGMutation> createAArch64MacroFusionDAGMutation () {
> + return EnableMacroFusion ? make_unique<AArch64MacroFusion>() : nullptr;
> +}
> +
> +} // end namespace llvm
>
> Added: llvm/trunk/lib/Target/AArch64/AArch64MacroFusion.h
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/Target/AArch64/AArch64MacroFusion.h?rev=293737&view=auto
> ==============================================================================
> --- llvm/trunk/lib/Target/AArch64/AArch64MacroFusion.h (added)
> +++ llvm/trunk/lib/Target/AArch64/AArch64MacroFusion.h Tue Jan 31 20:54:34 2017
> @@ -0,0 +1,29 @@
> +//===- AArch64MacroFusion.h - AArch64 Macro Fusion ------------------------===//
> +//
> +// The LLVM Compiler Infrastructure
> +//
> +// \fileThis file is distributed under the University of Illinois Open Source
> +// License. See LICENSE.TXT for details.
> +//
> +//===----------------------------------------------------------------------===//
> +//
> +// This file contains the AArch64 definition of the DAG scheduling mutation
> +// to pair instructions back to back.
> +//
> +//===----------------------------------------------------------------------===//
> +
> +#include "AArch64InstrInfo.h"
> +#include "llvm/CodeGen/MachineScheduler.h"
> +
> +//===----------------------------------------------------------------------===//
> +// AArch64MacroFusion - DAG post-processing to encourage fusion of macro ops.
> +//===----------------------------------------------------------------------===//
> +
> +namespace llvm {
> +
> +/// Note that you have to add:
> +/// DAG.addMutation(createAArch64MacroFusionDAGMutation());
> +/// to AArch64PassConfig::createMachineScheduler() to have an effect.
> +std::unique_ptr<ScheduleDAGMutation> createAArch64MacroFusionDAGMutation();
> +
> +} // llvm
>
> Modified: llvm/trunk/lib/Target/AArch64/AArch64TargetMachine.cpp
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/Target/AArch64/AArch64TargetMachine.cpp?rev=293737&r1=293736&r2=293737&view=diff
> ==============================================================================
> --- llvm/trunk/lib/Target/AArch64/AArch64TargetMachine.cpp (original)
> +++ llvm/trunk/lib/Target/AArch64/AArch64TargetMachine.cpp Tue Jan 31 20:54:34 2017
> @@ -14,6 +14,7 @@
> #include "AArch64CallLowering.h"
> #include "AArch64InstructionSelector.h"
> #include "AArch64LegalizerInfo.h"
> +#include "AArch64MacroFusion.h"
> #ifdef LLVM_BUILD_GLOBAL_ISEL
> #include "AArch64RegisterBankInfo.h"
> #endif
> @@ -325,7 +326,7 @@ public:
> ScheduleDAGMILive *DAG = createGenericSchedLive(C);
> DAG->addMutation(createLoadClusterDAGMutation(DAG->TII, DAG->TRI));
> DAG->addMutation(createStoreClusterDAGMutation(DAG->TII, DAG->TRI));
> - DAG->addMutation(createMacroFusionDAGMutation(DAG->TII));
> + DAG->addMutation(createAArch64MacroFusionDAGMutation());
> return DAG;
> }
>
>
> Modified: llvm/trunk/lib/Target/AArch64/CMakeLists.txt
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/Target/AArch64/CMakeLists.txt?rev=293737&r1=293736&r2=293737&view=diff
> ==============================================================================
> --- llvm/trunk/lib/Target/AArch64/CMakeLists.txt (original)
> +++ llvm/trunk/lib/Target/AArch64/CMakeLists.txt Tue Jan 31 20:54:34 2017
> @@ -56,6 +56,7 @@ add_llvm_target(AArch64CodeGen
> AArch64ISelLowering.cpp
> AArch64InstrInfo.cpp
> AArch64LoadStoreOptimizer.cpp
> + AArch64MacroFusion.cpp
> AArch64MCInstLower.cpp
> AArch64PromoteConstant.cpp
> AArch64PBQPRegAlloc.cpp
>
> Modified: llvm/trunk/lib/Target/X86/CMakeLists.txt
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/Target/X86/CMakeLists.txt?rev=293737&r1=293736&r2=293737&view=diff
> ==============================================================================
> --- llvm/trunk/lib/Target/X86/CMakeLists.txt (original)
> +++ llvm/trunk/lib/Target/X86/CMakeLists.txt Tue Jan 31 20:54:34 2017
> @@ -43,6 +43,7 @@ set(sources
> X86EvexToVex.cpp
> X86MCInstLower.cpp
> X86MachineFunctionInfo.cpp
> + X86MacroFusion.cpp
> X86OptimizeLEAs.cpp
> X86PadShortFunction.cpp
> X86RegisterInfo.cpp
>
> Modified: llvm/trunk/lib/Target/X86/X86InstrInfo.cpp
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/Target/X86/X86InstrInfo.cpp?rev=293737&r1=293736&r2=293737&view=diff
> ==============================================================================
> --- llvm/trunk/lib/Target/X86/X86InstrInfo.cpp (original)
> +++ llvm/trunk/lib/Target/X86/X86InstrInfo.cpp Tue Jan 31 20:54:34 2017
> @@ -8419,165 +8419,6 @@ bool X86InstrInfo::shouldScheduleLoadsNe
> return true;
> }
>
> -bool X86InstrInfo::shouldScheduleAdjacent(const MachineInstr &First,
> - const MachineInstr &Second) const {
> - // Check if this processor supports macro-fusion. Since this is a minor
> - // heuristic, we haven't specifically reserved a feature. hasAVX is a decent
> - // proxy for SandyBridge+.
> - if (!Subtarget.hasAVX())
> - return false;
> -
> - enum {
> - FuseTest,
> - FuseCmp,
> - FuseInc
> - } FuseKind;
> -
> - switch (Second.getOpcode()) {
> - default:
> - return false;
> - case X86::JE_1:
> - case X86::JNE_1:
> - case X86::JL_1:
> - case X86::JLE_1:
> - case X86::JG_1:
> - case X86::JGE_1:
> - FuseKind = FuseInc;
> - break;
> - case X86::JB_1:
> - case X86::JBE_1:
> - case X86::JA_1:
> - case X86::JAE_1:
> - FuseKind = FuseCmp;
> - break;
> - case X86::JS_1:
> - case X86::JNS_1:
> - case X86::JP_1:
> - case X86::JNP_1:
> - case X86::JO_1:
> - case X86::JNO_1:
> - FuseKind = FuseTest;
> - break;
> - }
> - switch (First.getOpcode()) {
> - default:
> - return false;
> - case X86::TEST8rr:
> - case X86::TEST16rr:
> - case X86::TEST32rr:
> - case X86::TEST64rr:
> - case X86::TEST8ri:
> - case X86::TEST16ri:
> - case X86::TEST32ri:
> - case X86::TEST32i32:
> - case X86::TEST64i32:
> - case X86::TEST64ri32:
> - case X86::TEST8rm:
> - case X86::TEST16rm:
> - case X86::TEST32rm:
> - case X86::TEST64rm:
> - case X86::TEST8ri_NOREX:
> - case X86::AND16i16:
> - case X86::AND16ri:
> - case X86::AND16ri8:
> - case X86::AND16rm:
> - case X86::AND16rr:
> - case X86::AND32i32:
> - case X86::AND32ri:
> - case X86::AND32ri8:
> - case X86::AND32rm:
> - case X86::AND32rr:
> - case X86::AND64i32:
> - case X86::AND64ri32:
> - case X86::AND64ri8:
> - case X86::AND64rm:
> - case X86::AND64rr:
> - case X86::AND8i8:
> - case X86::AND8ri:
> - case X86::AND8rm:
> - case X86::AND8rr:
> - return true;
> - case X86::CMP16i16:
> - case X86::CMP16ri:
> - case X86::CMP16ri8:
> - case X86::CMP16rm:
> - case X86::CMP16rr:
> - case X86::CMP32i32:
> - case X86::CMP32ri:
> - case X86::CMP32ri8:
> - case X86::CMP32rm:
> - case X86::CMP32rr:
> - case X86::CMP64i32:
> - case X86::CMP64ri32:
> - case X86::CMP64ri8:
> - case X86::CMP64rm:
> - case X86::CMP64rr:
> - case X86::CMP8i8:
> - case X86::CMP8ri:
> - case X86::CMP8rm:
> - case X86::CMP8rr:
> - case X86::ADD16i16:
> - case X86::ADD16ri:
> - case X86::ADD16ri8:
> - case X86::ADD16ri8_DB:
> - case X86::ADD16ri_DB:
> - case X86::ADD16rm:
> - case X86::ADD16rr:
> - case X86::ADD16rr_DB:
> - case X86::ADD32i32:
> - case X86::ADD32ri:
> - case X86::ADD32ri8:
> - case X86::ADD32ri8_DB:
> - case X86::ADD32ri_DB:
> - case X86::ADD32rm:
> - case X86::ADD32rr:
> - case X86::ADD32rr_DB:
> - case X86::ADD64i32:
> - case X86::ADD64ri32:
> - case X86::ADD64ri32_DB:
> - case X86::ADD64ri8:
> - case X86::ADD64ri8_DB:
> - case X86::ADD64rm:
> - case X86::ADD64rr:
> - case X86::ADD64rr_DB:
> - case X86::ADD8i8:
> - case X86::ADD8mi:
> - case X86::ADD8mr:
> - case X86::ADD8ri:
> - case X86::ADD8rm:
> - case X86::ADD8rr:
> - case X86::SUB16i16:
> - case X86::SUB16ri:
> - case X86::SUB16ri8:
> - case X86::SUB16rm:
> - case X86::SUB16rr:
> - case X86::SUB32i32:
> - case X86::SUB32ri:
> - case X86::SUB32ri8:
> - case X86::SUB32rm:
> - case X86::SUB32rr:
> - case X86::SUB64i32:
> - case X86::SUB64ri32:
> - case X86::SUB64ri8:
> - case X86::SUB64rm:
> - case X86::SUB64rr:
> - case X86::SUB8i8:
> - case X86::SUB8ri:
> - case X86::SUB8rm:
> - case X86::SUB8rr:
> - return FuseKind == FuseCmp || FuseKind == FuseInc;
> - case X86::INC16r:
> - case X86::INC32r:
> - case X86::INC64r:
> - case X86::INC8r:
> - case X86::DEC16r:
> - case X86::DEC32r:
> - case X86::DEC64r:
> - case X86::DEC8r:
> - return FuseKind == FuseInc;
> - }
> -}
> -
> bool X86InstrInfo::
> reverseBranchCondition(SmallVectorImpl<MachineOperand> &Cond) const {
> assert(Cond.size() == 1 && "Invalid X86 branch condition!");
>
> Modified: llvm/trunk/lib/Target/X86/X86InstrInfo.h
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/Target/X86/X86InstrInfo.h?rev=293737&r1=293736&r2=293737&view=diff
> ==============================================================================
> --- llvm/trunk/lib/Target/X86/X86InstrInfo.h (original)
> +++ llvm/trunk/lib/Target/X86/X86InstrInfo.h Tue Jan 31 20:54:34 2017
> @@ -443,9 +443,6 @@ public:
> int64_t Offset1, int64_t Offset2,
> unsigned NumLoads) const override;
>
> - bool shouldScheduleAdjacent(const MachineInstr &First,
> - const MachineInstr &Second) const override;
> -
> void getNoopForMachoTarget(MCInst &NopInst) const override;
>
> bool
>
> Added: llvm/trunk/lib/Target/X86/X86MacroFusion.cpp
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/Target/X86/X86MacroFusion.cpp?rev=293737&view=auto
> ==============================================================================
> --- llvm/trunk/lib/Target/X86/X86MacroFusion.cpp (added)
> +++ llvm/trunk/lib/Target/X86/X86MacroFusion.cpp Tue Jan 31 20:54:34 2017
> @@ -0,0 +1,262 @@
> +//===- X86MacroFusion.cpp - X86 Macro Fusion ------------------------------===//
> +//
> +// The LLVM Compiler Infrastructure
> +//
> +// \file This file is distributed under the University of Illinois Open Source
> +// License. See LICENSE.TXT for details.
> +//
> +//===----------------------------------------------------------------------===//
> +//
> +// This file contains the X86 implementation of the DAG scheduling mutation to
> +// pair instructions back to back.
> +//
> +//===----------------------------------------------------------------------===//
> +
> +#include "X86MacroFusion.h"
> +#include "X86Subtarget.h"
> +#include "llvm/Support/CommandLine.h"
> +#include "llvm/Target/TargetInstrInfo.h"
> +
> +#define DEBUG_TYPE "misched"
> +
> +using namespace llvm;
> +
> +static cl::opt<bool> EnableMacroFusion("x86-misched-fusion", cl::Hidden,
> + cl::desc("Enable scheduling for macro fusion."), cl::init(true));
> +
> +namespace {
> +
> +/// \brief Verify that the instruction pair, \param First and \param Second,
> +/// should be scheduled back to back. If either instruction is unspecified,
> +/// then verify that the other instruction may be part of a pair at all.
> +static bool shouldScheduleAdjacent(const X86Subtarget &ST,
> + const MachineInstr *First,
> + const MachineInstr *Second) {
> + // Check if this processor supports macro-fusion. Since this is a minor
> + // heuristic, we haven't specifically reserved a feature. hasAVX is a decent
> + // proxy for SandyBridge+.
> + if (!ST.hasAVX())
> + return false;
> +
> + enum {
> + FuseTest,
> + FuseCmp,
> + FuseInc
> + } FuseKind;
> +
> + unsigned FirstOpcode = First ?
> + First->getOpcode() : X86::INSTRUCTION_LIST_END;
> + unsigned SecondOpcode = Second ?
> + Second->getOpcode() : X86::INSTRUCTION_LIST_END;
> +
> + switch (SecondOpcode) {
> + default:
> + return false;
> + case X86::JE_1:
> + case X86::JNE_1:
> + case X86::JL_1:
> + case X86::JLE_1:
> + case X86::JG_1:
> + case X86::JGE_1:
> + FuseKind = FuseInc;
> + break;
> + case X86::JB_1:
> + case X86::JBE_1:
> + case X86::JA_1:
> + case X86::JAE_1:
> + FuseKind = FuseCmp;
> + break;
> + case X86::JS_1:
> + case X86::JNS_1:
> + case X86::JP_1:
> + case X86::JNP_1:
> + case X86::JO_1:
> + case X86::JNO_1:
> + FuseKind = FuseTest;
> + break;
> + }
> +
> + switch (FirstOpcode) {
> + default:
> + return false;
> + case X86::TEST8rr:
> + case X86::TEST16rr:
> + case X86::TEST32rr:
> + case X86::TEST64rr:
> + case X86::TEST8ri:
> + case X86::TEST16ri:
> + case X86::TEST32ri:
> + case X86::TEST32i32:
> + case X86::TEST64i32:
> + case X86::TEST64ri32:
> + case X86::TEST8rm:
> + case X86::TEST16rm:
> + case X86::TEST32rm:
> + case X86::TEST64rm:
> + case X86::TEST8ri_NOREX:
> + case X86::AND16i16:
> + case X86::AND16ri:
> + case X86::AND16ri8:
> + case X86::AND16rm:
> + case X86::AND16rr:
> + case X86::AND32i32:
> + case X86::AND32ri:
> + case X86::AND32ri8:
> + case X86::AND32rm:
> + case X86::AND32rr:
> + case X86::AND64i32:
> + case X86::AND64ri32:
> + case X86::AND64ri8:
> + case X86::AND64rm:
> + case X86::AND64rr:
> + case X86::AND8i8:
> + case X86::AND8ri:
> + case X86::AND8rm:
> + case X86::AND8rr:
> + return true;
> + case X86::CMP16i16:
> + case X86::CMP16ri:
> + case X86::CMP16ri8:
> + case X86::CMP16rm:
> + case X86::CMP16rr:
> + case X86::CMP32i32:
> + case X86::CMP32ri:
> + case X86::CMP32ri8:
> + case X86::CMP32rm:
> + case X86::CMP32rr:
> + case X86::CMP64i32:
> + case X86::CMP64ri32:
> + case X86::CMP64ri8:
> + case X86::CMP64rm:
> + case X86::CMP64rr:
> + case X86::CMP8i8:
> + case X86::CMP8ri:
> + case X86::CMP8rm:
> + case X86::CMP8rr:
> + case X86::ADD16i16:
> + case X86::ADD16ri:
> + case X86::ADD16ri8:
> + case X86::ADD16ri8_DB:
> + case X86::ADD16ri_DB:
> + case X86::ADD16rm:
> + case X86::ADD16rr:
> + case X86::ADD16rr_DB:
> + case X86::ADD32i32:
> + case X86::ADD32ri:
> + case X86::ADD32ri8:
> + case X86::ADD32ri8_DB:
> + case X86::ADD32ri_DB:
> + case X86::ADD32rm:
> + case X86::ADD32rr:
> + case X86::ADD32rr_DB:
> + case X86::ADD64i32:
> + case X86::ADD64ri32:
> + case X86::ADD64ri32_DB:
> + case X86::ADD64ri8:
> + case X86::ADD64ri8_DB:
> + case X86::ADD64rm:
> + case X86::ADD64rr:
> + case X86::ADD64rr_DB:
> + case X86::ADD8i8:
> + case X86::ADD8mi:
> + case X86::ADD8mr:
> + case X86::ADD8ri:
> + case X86::ADD8rm:
> + case X86::ADD8rr:
> + case X86::SUB16i16:
> + case X86::SUB16ri:
> + case X86::SUB16ri8:
> + case X86::SUB16rm:
> + case X86::SUB16rr:
> + case X86::SUB32i32:
> + case X86::SUB32ri:
> + case X86::SUB32ri8:
> + case X86::SUB32rm:
> + case X86::SUB32rr:
> + case X86::SUB64i32:
> + case X86::SUB64ri32:
> + case X86::SUB64ri8:
> + case X86::SUB64rm:
> + case X86::SUB64rr:
> + case X86::SUB8i8:
> + case X86::SUB8ri:
> + case X86::SUB8rm:
> + case X86::SUB8rr:
> + return FuseKind == FuseCmp || FuseKind == FuseInc;
> + case X86::INC16r:
> + case X86::INC32r:
> + case X86::INC64r:
> + case X86::INC8r:
> + case X86::DEC16r:
> + case X86::DEC32r:
> + case X86::DEC64r:
> + case X86::DEC8r:
> + return FuseKind == FuseInc;
> + case X86::INSTRUCTION_LIST_END:
> + return true;
> + }
> +}
> +
> +/// \brief Post-process the DAG to create cluster edges between instructions
> +/// that may be fused by the processor into a single operation.
> +class X86MacroFusion : public ScheduleDAGMutation {
> +public:
> + X86MacroFusion() {}
> +
> + void apply(ScheduleDAGInstrs *DAGInstrs) override;
> +};
> +
> +void X86MacroFusion::apply(ScheduleDAGInstrs *DAGInstrs) {
> + ScheduleDAGMI *DAG = static_cast<ScheduleDAGMI*>(DAGInstrs);
> + const X86Subtarget &ST = DAG->MF.getSubtarget<X86Subtarget>();
> +
> + // For now, assume targets can only fuse with the branch.
> + SUnit &ExitSU = DAG->ExitSU;
> + MachineInstr *Branch = ExitSU.getInstr();
> + if (!shouldScheduleAdjacent(ST, nullptr, Branch))
> + return;
> +
> + for (SDep &PredDep : ExitSU.Preds) {
> + if (PredDep.isWeak())
> + continue;
> + SUnit &SU = *PredDep.getSUnit();
> + MachineInstr &Pred = *SU.getInstr();
> + if (!shouldScheduleAdjacent(ST, &Pred, Branch))
> + continue;
> +
> + // Create a single weak edge from SU to ExitSU. The only effect is to cause
> + // bottom-up scheduling to heavily prioritize the clustered SU. There is no
> + // need to copy predecessor edges from ExitSU to SU, since top-down
> + // scheduling cannot prioritize ExitSU anyway. To defer top-down scheduling
> + // of SU, we could create an artificial edge from the deepest root, but it
> + // hasn't been needed yet.
> + bool Success = DAG->addEdge(&ExitSU, SDep(&SU, SDep::Cluster));
> + (void)Success;
> + assert(Success && "No DAG nodes should be reachable from ExitSU");
> +
> + // Adjust latency of data deps between the nodes.
> + for (SDep &PredDep : ExitSU.Preds)
> + if (PredDep.getSUnit() == &SU)
> + PredDep.setLatency(0);
> + for (SDep &SuccDep : SU.Succs)
> + if (SuccDep.getSUnit() == &ExitSU)
> + SuccDep.setLatency(0);
> +
> + DEBUG(dbgs() << "Macro fuse ";
> + SU.print(dbgs(), DAG);
> + dbgs() << " - ExitSU" << '\n');
> +
> + break;
> + }
> +}
> +
> +} // end namespace
> +
> +namespace llvm {
> +
> +std::unique_ptr<ScheduleDAGMutation>
> +createX86MacroFusionDAGMutation () {
> + return EnableMacroFusion ? make_unique<X86MacroFusion>() : nullptr;
> +}
> +
> +} // end namespace llvm
>
> Added: llvm/trunk/lib/Target/X86/X86MacroFusion.h
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/Target/X86/X86MacroFusion.h?rev=293737&view=auto
> ==============================================================================
> --- llvm/trunk/lib/Target/X86/X86MacroFusion.h (added)
> +++ llvm/trunk/lib/Target/X86/X86MacroFusion.h Tue Jan 31 20:54:34 2017
> @@ -0,0 +1,30 @@
> +//===- X86MacroFusion.h - X86 Macro Fusion --------------------------------===//
> +//
> +// The LLVM Compiler Infrastructure
> +//
> +// \file This file is distributed under the University of Illinois Open Source
> +// License. See LICENSE.TXT for details.
> +//
> +//===----------------------------------------------------------------------===//
> +//
> +// This file contains the X86 definition of the DAG scheduling mutation to pair
> +// instructions back to back.
> +//
> +//===----------------------------------------------------------------------===//
> +
> +#include "X86InstrInfo.h"
> +#include "llvm/CodeGen/MachineScheduler.h"
> +
> +//===----------------------------------------------------------------------===//
> +// X86MacroFusion - DAG post-processing to encourage fusion of macro ops.
> +//===----------------------------------------------------------------------===//
> +
> +namespace llvm {
> +
> +/// Note that you have to add:
> +/// DAG.addMutation(createX86MacroFusionDAGMutation());
> +/// to X86PassConfig::createMachineScheduler() to have an effect.
> +std::unique_ptr<ScheduleDAGMutation>
> +createX86MacroFusionDAGMutation();
> +
> +} // end namespace llvm
>
> Modified: llvm/trunk/lib/Target/X86/X86TargetMachine.cpp
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/Target/X86/X86TargetMachine.cpp?rev=293737&r1=293736&r2=293737&view=diff
> ==============================================================================
> --- llvm/trunk/lib/Target/X86/X86TargetMachine.cpp (original)
> +++ llvm/trunk/lib/Target/X86/X86TargetMachine.cpp Tue Jan 31 20:54:34 2017
> @@ -14,6 +14,7 @@
> #include "X86TargetMachine.h"
> #include "X86.h"
> #include "X86CallLowering.h"
> +#include "X86MacroFusion.h"
> #include "X86TargetObjectFile.h"
> #include "X86TargetTransformInfo.h"
> #include "llvm/CodeGen/GlobalISel/GISelAccessor.h"
> @@ -289,7 +290,7 @@ public:
> ScheduleDAGInstrs *
> createMachineScheduler(MachineSchedContext *C) const override {
> ScheduleDAGMILive *DAG = createGenericSchedLive(C);
> - DAG->addMutation(createMacroFusionDAGMutation(DAG->TII));
> + DAG->addMutation(createX86MacroFusionDAGMutation());
> return DAG;
> }
>
>
> Modified: llvm/trunk/test/CodeGen/AArch64/misched-fusion.ll
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/AArch64/misched-fusion.ll?rev=293737&r1=293736&r2=293737&view=diff
> ==============================================================================
> --- llvm/trunk/test/CodeGen/AArch64/misched-fusion.ll (original)
> +++ llvm/trunk/test/CodeGen/AArch64/misched-fusion.ll Tue Jan 31 20:54:34 2017
> @@ -1,22 +1,14 @@
> ; RUN: llc -o - %s -mattr=+arith-cbz-fusion | FileCheck %s
> ; RUN: llc -o - %s -mcpu=cyclone | FileCheck %s
>
> -target triple = "arm64-apple-ios"
> +target triple = "aarch64-unknown"
>
> declare void @foobar(i32 %v0, i32 %v1)
>
> ; Make sure sub is scheduled in front of cbnz
> ; CHECK-LABEL: test_sub_cbz:
> -; CHECK: add w[[ADDRES:[0-9]+]], w1, #7
> ; CHECK: sub w[[SUBRES:[0-9]+]], w0, #13
> -; CHECK-NEXT: cbnz w[[SUBRES]], [[SKIPBLOCK:LBB[0-9_]+]]
> -; CHECK: mov [[REGTY:[x,w]]]0, [[REGTY]][[ADDRES]]
> -; CHECK: mov [[REGTY]]1, [[REGTY]][[SUBRES]]
> -; CHECK: bl _foobar
> -; CHECK: [[SKIPBLOCK]]:
> -; CHECK: mov [[REGTY]]0, [[REGTY]][[SUBRES]]
> -; CHECK: mov [[REGTY]]1, [[REGTY]][[ADDRES]]
> -; CHECK: bl _foobar
> +; CHECK-NEXT: cbnz w[[SUBRES]], {{.?LBB[0-9_]+}}
> define void @test_sub_cbz(i32 %a0, i32 %a1) {
> entry:
> ; except for the fusion opportunity the sub/add should be equal so the
>
>
> _______________________________________________
> llvm-commits mailing list
> llvm-commits at lists.llvm.org
> http://lists.llvm.org/cgi-bin/mailman/listinfo/llvm-commits
--
Davide
"There are no solved problems; there are only problems that are more
or less solved" -- Henri Poincare
More information about the llvm-commits
mailing list