[llvm] r234513 - Refactoring and enhancement to FMA combine.

Hal Finkel hfinkel at anl.gov
Fri Apr 10 08:16:23 PDT 2015


----- Original Message -----
> From: "Olivier Sallenave" <ol.sall at gmail.com>
> To: "Rafael EspĂ­ndola" <rafael.espindola at gmail.com>
> Cc: llvm-commits at cs.uiuc.edu
> Sent: Friday, April 10, 2015 10:02:06 AM
> Subject: Re: [llvm] r234513 - Refactoring and enhancement to FMA combine.
> 
> 
> 
> Sorry about that. Not sure to understand though, amcgcn seem like an
> out-of-tree target, why are regression tests in the tree in the
> first place?

No, the test that is failing is test/CodeGen/R600/mad-combine.ll -- amdgcn is Triple::amdgcn, which is handled by the R600 backend.

 -Hal

> 
> 
> Thanks,
> 
> 
> Olivier
> 
> 
> 2015-04-09 14:33 GMT-04:00 Rafael EspĂ­ndola <
> rafael.espindola at gmail.com > :
> 
> 
> Reverted since this was failing on the bots:
> 
> http://lab.llvm.org:8011/builders/clang-x86_64-debian-fast/builds/25489
> 
> 
> 
> On 9 April 2015 at 13:55, Olivier Sallenave < ohsallen at us.ibm.com >
> wrote:
> > Author: ohsallen
> > Date: Thu Apr 9 12:55:26 2015
> > New Revision: 234513
> > 
> > URL: http://llvm.org/viewvc/llvm-project?rev=234513&view=rev
> > Log:
> > Refactoring and enhancement to FMA combine.
> > 
> > Modified:
> > llvm/trunk/lib/CodeGen/SelectionDAG/DAGCombiner.cpp
> > llvm/trunk/test/CodeGen/PowerPC/fma-assoc.ll
> > llvm/trunk/test/CodeGen/PowerPC/fma-ext.ll
> > 
> > Modified: llvm/trunk/lib/CodeGen/SelectionDAG/DAGCombiner.cpp
> > URL:
> > http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/CodeGen/SelectionDAG/DAGCombiner.cpp?rev=234513&r1=234512&r2=234513&view=diff
> > ==============================================================================
> > --- llvm/trunk/lib/CodeGen/SelectionDAG/DAGCombiner.cpp (original)
> > +++ llvm/trunk/lib/CodeGen/SelectionDAG/DAGCombiner.cpp Thu Apr 9
> > 12:55:26 2015
> > @@ -308,6 +308,9 @@ namespace {
> > SDValue visitMLOAD(SDNode *N);
> > SDValue visitMSTORE(SDNode *N);
> > 
> > + SDValue visitFADDForFMACombine(SDNode *N);
> > + SDValue visitFSUBForFMACombine(SDNode *N);
> > +
> > SDValue XformToShuffleWithZero(SDNode *N);
> > SDValue ReassociateOps(unsigned Opc, SDLoc DL, SDValue LHS, SDValue
> > RHS);
> > 
> > @@ -7057,20 +7060,40 @@ ConstantFoldBITCASTofBUILD_VECTOR(SDNode
> > return DAG.getNode(ISD::BUILD_VECTOR, SDLoc(BV), VT, Ops);
> > }
> > 
> > -// Attempt different variants of (fadd (fmul a, b), c) -> fma or
> > fmad
> > -static SDValue performFaddFmulCombines(unsigned FusedOpcode,
> > - bool Aggressive,
> > - SDNode *N,
> > - const TargetLowering &TLI,
> > - SelectionDAG &DAG) {
> > +/// Try to perform FMA combining on a given FADD node.
> > +SDValue DAGCombiner::visitFADDForFMACombine(SDNode *N) {
> > SDValue N0 = N->getOperand(0);
> > SDValue N1 = N->getOperand(1);
> > EVT VT = N->getValueType(0);
> > + SDLoc SL(N);
> > +
> > + const TargetOptions &Options = DAG.getTarget().Options;
> > + bool UnsafeFPMath = (Options.AllowFPOpFusion == FPOpFusion::Fast
> > ||
> > + Options.UnsafeFPMath);
> > +
> > + // Floating-point multiply-add with intermediate rounding.
> > + bool HasFMAD = (LegalOperations &&
> > + TLI.isOperationLegal(ISD::FMAD, VT));
> > +
> > + // Floating-point multiply-add without intermediate rounding.
> > + bool HasFMA = ((!LegalOperations ||
> > + TLI.isOperationLegalOrCustom(ISD::FMA, VT)) &&
> > + TLI.isFMAFasterThanFMulAndFAdd(VT) &&
> > + UnsafeFPMath);
> > +
> > + // No valid opcode, do not combine.
> > + if (!HasFMAD && !HasFMA)
> > + return SDValue();
> > +
> > + // Always prefer FMAD to FMA for precision.
> > + unsigned int PreferredFusedOpcode = HasFMAD ? ISD::FMAD :
> > ISD::FMA;
> > + bool Aggressive = TLI.enableAggressiveFMAFusion(VT);
> > + bool LookThroughFPExt = TLI.isFPExtFree(VT);
> > 
> > // fold (fadd (fmul x, y), z) -> (fma x, y, z)
> > if (N0.getOpcode() == ISD::FMUL &&
> > (Aggressive || N0->hasOneUse())) {
> > - return DAG.getNode(FusedOpcode, SDLoc(N), VT,
> > + return DAG.getNode(PreferredFusedOpcode, SL, VT,
> > N0.getOperand(0), N0.getOperand(1), N1);
> > }
> > 
> > @@ -7078,53 +7101,176 @@ static SDValue performFaddFmulCombines(u
> > // Note: Commutes FADD operands.
> > if (N1.getOpcode() == ISD::FMUL &&
> > (Aggressive || N1->hasOneUse())) {
> > - return DAG.getNode(FusedOpcode, SDLoc(N), VT,
> > + return DAG.getNode(PreferredFusedOpcode, SL, VT,
> > N1.getOperand(0), N1.getOperand(1), N0);
> > }
> > 
> > + // Look through FP_EXTEND nodes to do more combining.
> > + if (UnsafeFPMath && LookThroughFPExt) {
> > + // fold (fadd (fpext (fmul x, y)), z) -> (fma (fpext x), (fpext
> > y), z)
> > + if (N0.getOpcode() == ISD::FP_EXTEND) {
> > + SDValue N00 = N0.getOperand(0);
> > + if (N00.getOpcode() == ISD::FMUL)
> > + return DAG.getNode(PreferredFusedOpcode, SL, VT,
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT,
> > + N00.getOperand(0)),
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT,
> > + N00.getOperand(1)), N1);
> > + }
> > +
> > + // fold (fadd x, (fpext (fmul y, z))) -> (fma (fpext y), (fpext
> > z), x)
> > + // Note: Commutes FADD operands.
> > + if (N1.getOpcode() == ISD::FP_EXTEND) {
> > + SDValue N10 = N1.getOperand(0);
> > + if (N10.getOpcode() == ISD::FMUL)
> > + return DAG.getNode(PreferredFusedOpcode, SL, VT,
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT,
> > + N10.getOperand(0)),
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT,
> > + N10.getOperand(1)), N0);
> > + }
> > + }
> > +
> > // More folding opportunities when target permits.
> > - if (Aggressive) {
> > + if (UnsafeFPMath && Aggressive) {
> > // fold (fadd (fma x, y, (fmul u, v)), z) -> (fma x, y (fma u, v,
> > z))
> > - if (N0.getOpcode() == ISD::FMA &&
> > + if (N0.getOpcode() == PreferredFusedOpcode &&
> > N0.getOperand(2).getOpcode() == ISD::FMUL) {
> > - return DAG.getNode(FusedOpcode, SDLoc(N), VT,
> > + return DAG.getNode(PreferredFusedOpcode, SL, VT,
> > N0.getOperand(0), N0.getOperand(1),
> > - DAG.getNode(FusedOpcode, SDLoc(N), VT,
> > + DAG.getNode(PreferredFusedOpcode, SL, VT,
> > N0.getOperand(2).getOperand(0),
> > N0.getOperand(2).getOperand(1),
> > N1));
> > }
> > 
> > // fold (fadd x, (fma y, z, (fmul u, v)) -> (fma y, z (fma u, v,
> > x))
> > - if (N1->getOpcode() == ISD::FMA &&
> > + if (N1->getOpcode() == PreferredFusedOpcode &&
> > N1.getOperand(2).getOpcode() == ISD::FMUL) {
> > - return DAG.getNode(FusedOpcode, SDLoc(N), VT,
> > + return DAG.getNode(PreferredFusedOpcode, SL, VT,
> > N1.getOperand(0), N1.getOperand(1),
> > - DAG.getNode(FusedOpcode, SDLoc(N), VT,
> > + DAG.getNode(PreferredFusedOpcode, SL, VT,
> > N1.getOperand(2).getOperand(0),
> > N1.getOperand(2).getOperand(1),
> > N0));
> > }
> > +
> > + if (LookThroughFPExt) {
> > + // fold (fadd (fma x, y, (fpext (fmul u, v))), z)
> > + // -> (fma x, y, (fma (fpext u), (fpext v), z))
> > + auto FoldFAddFMAFPExtFMul = [&] (
> > + SDValue X, SDValue Y, SDValue U, SDValue V, SDValue Z) {
> > + return DAG.getNode(PreferredFusedOpcode, SL, VT, X, Y,
> > + DAG.getNode(PreferredFusedOpcode, SL, VT,
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT, U),
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT, V),
> > + Z));
> > + };
> > + if (N0.getOpcode() == PreferredFusedOpcode) {
> > + SDValue N02 = N0.getOperand(2);
> > + if (N02.getOpcode() == ISD::FP_EXTEND) {
> > + SDValue N020 = N02.getOperand(0);
> > + if (N020.getOpcode() == ISD::FMUL)
> > + return FoldFAddFMAFPExtFMul(N0.getOperand(0), N0.getOperand(1),
> > + N020.getOperand(0), N020.getOperand(1),
> > + N1);
> > + }
> > + }
> > +
> > + // fold (fadd (fpext (fma x, y, (fmul u, v))), z)
> > + // -> (fma (fpext x), (fpext y), (fma (fpext u), (fpext v), z))
> > + // FIXME: This turns two single-precision and one
> > double-precision
> > + // operation into two double-precision operations, which might
> > not be
> > + // interesting for all targets, especially GPUs.
> > + auto FoldFAddFPExtFMAFMul = [&] (
> > + SDValue X, SDValue Y, SDValue U, SDValue V, SDValue Z) {
> > + return DAG.getNode(PreferredFusedOpcode, SL, VT,
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT, X),
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT, Y),
> > + DAG.getNode(PreferredFusedOpcode, SL, VT,
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT, U),
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT, V),
> > + Z));
> > + };
> > + if (N0.getOpcode() == ISD::FP_EXTEND) {
> > + SDValue N00 = N0.getOperand(0);
> > + if (N00.getOpcode() == PreferredFusedOpcode) {
> > + SDValue N002 = N00.getOperand(2);
> > + if (N002.getOpcode() == ISD::FMUL)
> > + return FoldFAddFPExtFMAFMul(N00.getOperand(0), N00.getOperand(1),
> > + N002.getOperand(0), N002.getOperand(1),
> > + N1);
> > + }
> > + }
> > +
> > + // fold (fadd x, (fma y, z, (fpext (fmul u, v)))
> > + // -> (fma y, z, (fma (fpext u), (fpext v), x))
> > + if (N1.getOpcode() == PreferredFusedOpcode) {
> > + SDValue N12 = N1.getOperand(2);
> > + if (N12.getOpcode() == ISD::FP_EXTEND) {
> > + SDValue N120 = N12.getOperand(0);
> > + if (N120.getOpcode() == ISD::FMUL)
> > + return FoldFAddFMAFPExtFMul(N1.getOperand(0), N1.getOperand(1),
> > + N120.getOperand(0), N120.getOperand(1),
> > + N0);
> > + }
> > + }
> > +
> > + // fold (fadd x, (fpext (fma y, z, (fmul u, v)))
> > + // -> (fma (fpext y), (fpext z), (fma (fpext u), (fpext v), x))
> > + // FIXME: This turns two single-precision and one
> > double-precision
> > + // operation into two double-precision operations, which might
> > not be
> > + // interesting for all targets, especially GPUs.
> > + if (N1.getOpcode() == ISD::FP_EXTEND) {
> > + SDValue N10 = N1.getOperand(0);
> > + if (N10.getOpcode() == PreferredFusedOpcode) {
> > + SDValue N102 = N10.getOperand(2);
> > + if (N102.getOpcode() == ISD::FMUL)
> > + return FoldFAddFPExtFMAFMul(N10.getOperand(0), N10.getOperand(1),
> > + N102.getOperand(0), N102.getOperand(1),
> > + N0);
> > + }
> > + }
> > + }
> > }
> > 
> > return SDValue();
> > }
> > 
> > -static SDValue performFsubFmulCombines(unsigned FusedOpcode,
> > - bool Aggressive,
> > - SDNode *N,
> > - const TargetLowering &TLI,
> > - SelectionDAG &DAG) {
> > +/// Try to perform FMA combining on a given FSUB node.
> > +SDValue DAGCombiner::visitFSUBForFMACombine(SDNode *N) {
> > SDValue N0 = N->getOperand(0);
> > SDValue N1 = N->getOperand(1);
> > EVT VT = N->getValueType(0);
> > -
> > SDLoc SL(N);
> > 
> > + const TargetOptions &Options = DAG.getTarget().Options;
> > + bool UnsafeFPMath = (Options.AllowFPOpFusion == FPOpFusion::Fast
> > ||
> > + Options.UnsafeFPMath);
> > +
> > + // Floating-point multiply-add with intermediate rounding.
> > + bool HasFMAD = (LegalOperations &&
> > + TLI.isOperationLegal(ISD::FMAD, VT));
> > +
> > + // Floating-point multiply-add without intermediate rounding.
> > + bool HasFMA = ((!LegalOperations ||
> > + TLI.isOperationLegalOrCustom(ISD::FMA, VT)) &&
> > + TLI.isFMAFasterThanFMulAndFAdd(VT) &&
> > + UnsafeFPMath);
> > +
> > + // No valid opcode, do not combine.
> > + if (!HasFMAD && !HasFMA)
> > + return SDValue();
> > +
> > + // Always prefer FMAD to FMA for precision.
> > + unsigned int PreferredFusedOpcode = HasFMAD ? ISD::FMAD :
> > ISD::FMA;
> > + bool Aggressive = TLI.enableAggressiveFMAFusion(VT);
> > + bool LookThroughFPExt = TLI.isFPExtFree(VT);
> > +
> > // fold (fsub (fmul x, y), z) -> (fma x, y, (fneg z))
> > if (N0.getOpcode() == ISD::FMUL &&
> > (Aggressive || N0->hasOneUse())) {
> > - return DAG.getNode(FusedOpcode, SL, VT,
> > + return DAG.getNode(PreferredFusedOpcode, SL, VT,
> > N0.getOperand(0), N0.getOperand(1),
> > DAG.getNode(ISD::FNEG, SL, VT, N1));
> > }
> > @@ -7133,7 +7279,7 @@ static SDValue performFsubFmulCombines(u
> > // Note: Commutes FSUB operands.
> > if (N1.getOpcode() == ISD::FMUL &&
> > (Aggressive || N1->hasOneUse()))
> > - return DAG.getNode(FusedOpcode, SL, VT,
> > + return DAG.getNode(PreferredFusedOpcode, SL, VT,
> > DAG.getNode(ISD::FNEG, SL, VT,
> > N1.getOperand(0)),
> > N1.getOperand(1), N0);
> > @@ -7144,41 +7290,213 @@ static SDValue performFsubFmulCombines(u
> > (Aggressive || (N0->hasOneUse() && N0.getOperand(0).hasOneUse())))
> > {
> > SDValue N00 = N0.getOperand(0).getOperand(0);
> > SDValue N01 = N0.getOperand(0).getOperand(1);
> > - return DAG.getNode(FusedOpcode, SL, VT,
> > + return DAG.getNode(PreferredFusedOpcode, SL, VT,
> > DAG.getNode(ISD::FNEG, SL, VT, N00), N01,
> > DAG.getNode(ISD::FNEG, SL, VT, N1));
> > }
> > 
> > + // Look through FP_EXTEND nodes to do more combining.
> > + if (UnsafeFPMath && LookThroughFPExt) {
> > + // fold (fsub (fpext (fmul x, y)), z)
> > + // -> (fma (fpext x), (fpext y), (fneg z))
> > + if (N0.getOpcode() == ISD::FP_EXTEND) {
> > + SDValue N00 = N0.getOperand(0);
> > + if (N00.getOpcode() == ISD::FMUL)
> > + return DAG.getNode(PreferredFusedOpcode, SL, VT,
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT,
> > + N00.getOperand(0)),
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT,
> > + N00.getOperand(1)),
> > + DAG.getNode(ISD::FNEG, SL, VT, N1));
> > + }
> > +
> > + // fold (fsub x, (fpext (fmul y, z)))
> > + // -> (fma (fneg (fpext y)), (fpext z), x)
> > + // Note: Commutes FSUB operands.
> > + if (N1.getOpcode() == ISD::FP_EXTEND) {
> > + SDValue N10 = N1.getOperand(0);
> > + if (N10.getOpcode() == ISD::FMUL)
> > + return DAG.getNode(PreferredFusedOpcode, SL, VT,
> > + DAG.getNode(ISD::FNEG, SL, VT,
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT,
> > + N10.getOperand(0))),
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT,
> > + N10.getOperand(1)),
> > + N0);
> > + }
> > +
> > + // fold (fsub (fpext (fneg (fmul, x, y))), z)
> > + // -> (fneg (fma (fpext x), (fpext y), z))
> > + // Note: This could be removed with appropriate canonicalization
> > of the
> > + // input expression into (fneg (fadd (fpext (fmul, x, y)), z).
> > However, the
> > + // orthogonal flags -fp-contract=fast and -enable-unsafe-fp-math
> > prevent
> > + // from implementing the canonicalization in visitFSUB.
> > + if (N0.getOpcode() == ISD::FP_EXTEND) {
> > + SDValue N00 = N0.getOperand(0);
> > + if (N00.getOpcode() == ISD::FNEG) {
> > + SDValue N000 = N00.getOperand(0);
> > + if (N000.getOpcode() == ISD::FMUL) {
> > + return DAG.getNode(ISD::FNEG, SL, VT,
> > + DAG.getNode(PreferredFusedOpcode, SL, VT,
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT,
> > + N000.getOperand(0)),
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT,
> > + N000.getOperand(1)),
> > + N1));
> > + }
> > + }
> > + }
> > +
> > + // fold (fsub (fneg (fpext (fmul, x, y))), z)
> > + // -> (fneg (fma (fpext x)), (fpext y), z)
> > + // Note: This could be removed with appropriate canonicalization
> > of the
> > + // input expression into (fneg (fadd (fpext (fmul, x, y)), z).
> > However, the
> > + // orthogonal flags -fp-contract=fast and -enable-unsafe-fp-math
> > prevent
> > + // from implementing the canonicalization in visitFSUB.
> > + if (N0.getOpcode() == ISD::FNEG) {
> > + SDValue N00 = N0.getOperand(0);
> > + if (N00.getOpcode() == ISD::FP_EXTEND) {
> > + SDValue N000 = N00.getOperand(0);
> > + if (N000.getOpcode() == ISD::FMUL) {
> > + return DAG.getNode(ISD::FNEG, SL, VT,
> > + DAG.getNode(PreferredFusedOpcode, SL, VT,
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT,
> > + N000.getOperand(0)),
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT,
> > + N000.getOperand(1)),
> > + N1));
> > + }
> > + }
> > + }
> > +
> > + }
> > +
> > // More folding opportunities when target permits.
> > - if (Aggressive) {
> > + if (UnsafeFPMath && Aggressive) {
> > // fold (fsub (fma x, y, (fmul u, v)), z)
> > // -> (fma x, y (fma u, v, (fneg z)))
> > - if (N0.getOpcode() == FusedOpcode &&
> > + if (N0.getOpcode() == PreferredFusedOpcode &&
> > N0.getOperand(2).getOpcode() == ISD::FMUL) {
> > - return DAG.getNode(FusedOpcode, SDLoc(N), VT,
> > + return DAG.getNode(PreferredFusedOpcode, SL, VT,
> > N0.getOperand(0), N0.getOperand(1),
> > - DAG.getNode(FusedOpcode, SDLoc(N), VT,
> > + DAG.getNode(PreferredFusedOpcode, SL, VT,
> > N0.getOperand(2).getOperand(0),
> > N0.getOperand(2).getOperand(1),
> > - DAG.getNode(ISD::FNEG, SDLoc(N), VT,
> > + DAG.getNode(ISD::FNEG, SL, VT,
> > N1)));
> > }
> > 
> > // fold (fsub x, (fma y, z, (fmul u, v)))
> > // -> (fma (fneg y), z, (fma (fneg u), v, x))
> > - if (N1.getOpcode() == FusedOpcode &&
> > + if (N1.getOpcode() == PreferredFusedOpcode &&
> > N1.getOperand(2).getOpcode() == ISD::FMUL) {
> > SDValue N20 = N1.getOperand(2).getOperand(0);
> > SDValue N21 = N1.getOperand(2).getOperand(1);
> > - return DAG.getNode(FusedOpcode, SDLoc(N), VT,
> > - DAG.getNode(ISD::FNEG, SDLoc(N), VT,
> > + return DAG.getNode(PreferredFusedOpcode, SL, VT,
> > + DAG.getNode(ISD::FNEG, SL, VT,
> > N1.getOperand(0)),
> > N1.getOperand(1),
> > - DAG.getNode(FusedOpcode, SDLoc(N), VT,
> > - DAG.getNode(ISD::FNEG, SDLoc(N), VT,
> > - N20),
> > + DAG.getNode(PreferredFusedOpcode, SL, VT,
> > + DAG.getNode(ISD::FNEG, SL, VT, N20),
> > N21, N0));
> > }
> > +
> > + if (LookThroughFPExt) {
> > + // fold (fsub (fma x, y, (fpext (fmul u, v))), z)
> > + // -> (fma x, y (fma (fpext u), (fpext v), (fneg z)))
> > + if (N0.getOpcode() == PreferredFusedOpcode) {
> > + SDValue N02 = N0.getOperand(2);
> > + if (N02.getOpcode() == ISD::FP_EXTEND) {
> > + SDValue N020 = N02.getOperand(0);
> > + if (N020.getOpcode() == ISD::FMUL)
> > + return DAG.getNode(PreferredFusedOpcode, SL, VT,
> > + N0.getOperand(0), N0.getOperand(1),
> > + DAG.getNode(PreferredFusedOpcode, SL, VT,
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT,
> > + N020.getOperand(0)),
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT,
> > + N020.getOperand(1)),
> > + DAG.getNode(ISD::FNEG, SL, VT,
> > + N1)));
> > + }
> > + }
> > +
> > + // fold (fsub (fpext (fma x, y, (fmul u, v))), z)
> > + // -> (fma (fpext x), (fpext y),
> > + // (fma (fpext u), (fpext v), (fneg z)))
> > + // FIXME: This turns two single-precision and one
> > double-precision
> > + // operation into two double-precision operations, which might
> > not be
> > + // interesting for all targets, especially GPUs.
> > + if (N0.getOpcode() == ISD::FP_EXTEND) {
> > + SDValue N00 = N0.getOperand(0);
> > + if (N00.getOpcode() == PreferredFusedOpcode) {
> > + SDValue N002 = N00.getOperand(2);
> > + if (N002.getOpcode() == ISD::FMUL)
> > + return DAG.getNode(PreferredFusedOpcode, SL, VT,
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT,
> > + N00.getOperand(0)),
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT,
> > + N00.getOperand(1)),
> > + DAG.getNode(PreferredFusedOpcode, SL, VT,
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT,
> > + N002.getOperand(0)),
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT,
> > + N002.getOperand(1)),
> > + DAG.getNode(ISD::FNEG, SL, VT,
> > + N1)));
> > + }
> > + }
> > +
> > + // fold (fsub x, (fma y, z, (fpext (fmul u, v))))
> > + // -> (fma (fneg y), z, (fma (fneg (fpext u)), (fpext v), x))
> > + if (N1.getOpcode() == PreferredFusedOpcode &&
> > + N1.getOperand(2).getOpcode() == ISD::FP_EXTEND) {
> > + SDValue N120 = N1.getOperand(2).getOperand(0);
> > + if (N120.getOpcode() == ISD::FMUL) {
> > + SDValue N1200 = N120.getOperand(0);
> > + SDValue N1201 = N120.getOperand(1);
> > + return DAG.getNode(PreferredFusedOpcode, SL, VT,
> > + DAG.getNode(ISD::FNEG, SL, VT, N1.getOperand(0)),
> > + N1.getOperand(1),
> > + DAG.getNode(PreferredFusedOpcode, SL, VT,
> > + DAG.getNode(ISD::FNEG, SL, VT,
> > + DAG.getNode(ISD::FP_EXTEND, SL,
> > + VT, N1200)),
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT,
> > + N1201),
> > + N0));
> > + }
> > + }
> > +
> > + // fold (fsub x, (fpext (fma y, z, (fmul u, v))))
> > + // -> (fma (fneg (fpext y)), (fpext z),
> > + // (fma (fneg (fpext u)), (fpext v), x))
> > + // FIXME: This turns two single-precision and one
> > double-precision
> > + // operation into two double-precision operations, which might
> > not be
> > + // interesting for all targets, especially GPUs.
> > + if (N1.getOpcode() == ISD::FP_EXTEND &&
> > + N1.getOperand(0).getOpcode() == PreferredFusedOpcode) {
> > + SDValue N100 = N1.getOperand(0).getOperand(0);
> > + SDValue N101 = N1.getOperand(0).getOperand(1);
> > + SDValue N102 = N1.getOperand(0).getOperand(2);
> > + if (N102.getOpcode() == ISD::FMUL) {
> > + SDValue N1020 = N102.getOperand(0);
> > + SDValue N1021 = N102.getOperand(1);
> > + return DAG.getNode(PreferredFusedOpcode, SL, VT,
> > + DAG.getNode(ISD::FNEG, SL, VT,
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT,
> > + N100)),
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT, N101),
> > + DAG.getNode(PreferredFusedOpcode, SL, VT,
> > + DAG.getNode(ISD::FNEG, SL, VT,
> > + DAG.getNode(ISD::FP_EXTEND, SL,
> > + VT, N1020)),
> > + DAG.getNode(ISD::FP_EXTEND, SL, VT,
> > + N1021),
> > + N0));
> > + }
> > + }
> > + }
> > }
> > 
> > return SDValue();
> > @@ -7322,55 +7640,11 @@ SDValue DAGCombiner::visitFADD(SDNode *N
> > }
> > } // enable-unsafe-fp-math
> > 
> > - if (LegalOperations && TLI.isOperationLegal(ISD::FMAD, VT)) {
> > - // Assume if there is an fmad instruction that it should be
> > aggressively
> > - // used.
> > - if (SDValue Fused = performFaddFmulCombines(ISD::FMAD, true, N,
> > TLI, DAG))
> > - return Fused;
> > - }
> > -
> > // FADD -> FMA combines:
> > - if ((Options.AllowFPOpFusion == FPOpFusion::Fast ||
> > Options.UnsafeFPMath) &&
> > - TLI.isFMAFasterThanFMulAndFAdd(VT) &&
> > - (!LegalOperations || TLI.isOperationLegalOrCustom(ISD::FMA, VT)))
> > {
> > -
> > - if (!TLI.isOperationLegal(ISD::FMAD, VT)) {
> > - // Don't form FMA if we are preferring FMAD.
> > - if (SDValue Fused
> > - = performFaddFmulCombines(ISD::FMA,
> > - TLI.enableAggressiveFMAFusion(VT),
> > - N, TLI, DAG)) {
> > - return Fused;
> > - }
> > - }
> > -
> > - // When FP_EXTEND nodes are free on the target, and there is an
> > opportunity
> > - // to combine into FMA, arrange such nodes accordingly.
> > - if (TLI.isFPExtFree(VT)) {
> > -
> > - // fold (fadd (fpext (fmul x, y)), z) -> (fma (fpext x), (fpext
> > y), z)
> > - if (N0.getOpcode() == ISD::FP_EXTEND) {
> > - SDValue N00 = N0.getOperand(0);
> > - if (N00.getOpcode() == ISD::FMUL)
> > - return DAG.getNode(ISD::FMA, SDLoc(N), VT,
> > - DAG.getNode(ISD::FP_EXTEND, SDLoc(N), VT,
> > - N00.getOperand(0)),
> > - DAG.getNode(ISD::FP_EXTEND, SDLoc(N), VT,
> > - N00.getOperand(1)), N1);
> > - }
> > -
> > - // fold (fadd x, (fpext (fmul y, z)), z) -> (fma (fpext y),
> > (fpext z), x)
> > - // Note: Commutes FADD operands.
> > - if (N1.getOpcode() == ISD::FP_EXTEND) {
> > - SDValue N10 = N1.getOperand(0);
> > - if (N10.getOpcode() == ISD::FMUL)
> > - return DAG.getNode(ISD::FMA, SDLoc(N), VT,
> > - DAG.getNode(ISD::FP_EXTEND, SDLoc(N), VT,
> > - N10.getOperand(0)),
> > - DAG.getNode(ISD::FP_EXTEND, SDLoc(N), VT,
> > - N10.getOperand(1)), N0);
> > - }
> > - }
> > + SDValue Fused = visitFADDForFMACombine(N);
> > + if (Fused) {
> > + AddToWorklist(Fused.getNode());
> > + return Fused;
> > }
> > 
> > return SDValue();
> > @@ -7431,96 +7705,11 @@ SDValue DAGCombiner::visitFSUB(SDNode *N
> > }
> > }
> > 
> > - if (LegalOperations && TLI.isOperationLegal(ISD::FMAD, VT)) {
> > - // Assume if there is an fmad instruction that it should be
> > aggressively
> > - // used.
> > - if (SDValue Fused = performFsubFmulCombines(ISD::FMAD, true, N,
> > TLI, DAG))
> > - return Fused;
> > - }
> > -
> > // FSUB -> FMA combines:
> > - if ((Options.AllowFPOpFusion == FPOpFusion::Fast ||
> > Options.UnsafeFPMath) &&
> > - TLI.isFMAFasterThanFMulAndFAdd(VT) &&
> > - (!LegalOperations || TLI.isOperationLegalOrCustom(ISD::FMA, VT)))
> > {
> > -
> > - if (!TLI.isOperationLegal(ISD::FMAD, VT)) {
> > - // Don't form FMA if we are preferring FMAD.
> > -
> > - if (SDValue Fused
> > - = performFsubFmulCombines(ISD::FMA,
> > - TLI.enableAggressiveFMAFusion(VT),
> > - N, TLI, DAG)) {
> > - return Fused;
> > - }
> > - }
> > -
> > - // When FP_EXTEND nodes are free on the target, and there is an
> > opportunity
> > - // to combine into FMA, arrange such nodes accordingly.
> > - if (TLI.isFPExtFree(VT)) {
> > - // fold (fsub (fpext (fmul x, y)), z)
> > - // -> (fma (fpext x), (fpext y), (fneg z))
> > - if (N0.getOpcode() == ISD::FP_EXTEND) {
> > - SDValue N00 = N0.getOperand(0);
> > - if (N00.getOpcode() == ISD::FMUL)
> > - return DAG.getNode(ISD::FMA, SDLoc(N), VT,
> > - DAG.getNode(ISD::FP_EXTEND, SDLoc(N), VT,
> > - N00.getOperand(0)),
> > - DAG.getNode(ISD::FP_EXTEND, SDLoc(N), VT,
> > - N00.getOperand(1)),
> > - DAG.getNode(ISD::FNEG, SDLoc(N), VT, N1));
> > - }
> > -
> > - // fold (fsub x, (fpext (fmul y, z)))
> > - // -> (fma (fneg (fpext y)), (fpext z), x)
> > - // Note: Commutes FSUB operands.
> > - if (N1.getOpcode() == ISD::FP_EXTEND) {
> > - SDValue N10 = N1.getOperand(0);
> > - if (N10.getOpcode() == ISD::FMUL)
> > - return DAG.getNode(ISD::FMA, SDLoc(N), VT,
> > - DAG.getNode(ISD::FNEG, SDLoc(N), VT,
> > - DAG.getNode(ISD::FP_EXTEND, SDLoc(N),
> > - VT, N10.getOperand(0))),
> > - DAG.getNode(ISD::FP_EXTEND, SDLoc(N), VT,
> > - N10.getOperand(1)),
> > - N0);
> > - }
> > -
> > - // fold (fsub (fpext (fneg (fmul, x, y))), z)
> > - // -> (fma (fneg (fpext x)), (fpext y), (fneg z))
> > - if (N0.getOpcode() == ISD::FP_EXTEND) {
> > - SDValue N00 = N0.getOperand(0);
> > - if (N00.getOpcode() == ISD::FNEG) {
> > - SDValue N000 = N00.getOperand(0);
> > - if (N000.getOpcode() == ISD::FMUL) {
> > - return DAG.getNode(ISD::FMA, dl, VT,
> > - DAG.getNode(ISD::FNEG, dl, VT,
> > - DAG.getNode(ISD::FP_EXTEND, SDLoc(N),
> > - VT, N000.getOperand(0))),
> > - DAG.getNode(ISD::FP_EXTEND, SDLoc(N), VT,
> > - N000.getOperand(1)),
> > - DAG.getNode(ISD::FNEG, dl, VT, N1));
> > - }
> > - }
> > - }
> > -
> > - // fold (fsub (fneg (fpext (fmul, x, y))), z)
> > - // -> (fma (fneg (fpext x)), (fpext y), (fneg z))
> > - if (N0.getOpcode() == ISD::FNEG) {
> > - SDValue N00 = N0.getOperand(0);
> > - if (N00.getOpcode() == ISD::FP_EXTEND) {
> > - SDValue N000 = N00.getOperand(0);
> > - if (N000.getOpcode() == ISD::FMUL) {
> > - return DAG.getNode(ISD::FMA, dl, VT,
> > - DAG.getNode(ISD::FNEG, dl, VT,
> > - DAG.getNode(ISD::FP_EXTEND, SDLoc(N),
> > - VT, N000.getOperand(0))),
> > - DAG.getNode(ISD::FP_EXTEND, SDLoc(N), VT,
> > - N000.getOperand(1)),
> > - DAG.getNode(ISD::FNEG, dl, VT, N1));
> > - }
> > - }
> > - }
> > - }
> > + SDValue Fused = visitFSUBForFMACombine(N);
> > + if (Fused) {
> > + AddToWorklist(Fused.getNode());
> > + return Fused;
> > }
> > 
> > return SDValue();
> > 
> > Modified: llvm/trunk/test/CodeGen/PowerPC/fma-assoc.ll
> > URL:
> > http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/PowerPC/fma-assoc.ll?rev=234513&r1=234512&r2=234513&view=diff
> > ==============================================================================
> > --- llvm/trunk/test/CodeGen/PowerPC/fma-assoc.ll (original)
> > +++ llvm/trunk/test/CodeGen/PowerPC/fma-assoc.ll Thu Apr 9 12:55:26
> > 2015
> > @@ -77,3 +77,159 @@ define double @test_FMSUB_ASSOC2(double
> > ; CHECK-VSX-NEXT: blr
> > }
> > 
> > +define double @test_FMADD_ASSOC_EXT1(float %A, float %B, double
> > %C,
> > + double %D, double %E) {
> > + %F = fmul float %A, %B ; <float> [#uses=1]
> > + %G = fpext float %F to double ; <double> [#uses=1]
> > + %H = fmul double %C, %D ; <double> [#uses=1]
> > + %I = fadd double %H, %G ; <double> [#uses=1]
> > + %J = fadd double %I, %E ; <double> [#uses=1]
> > + ret double %J
> > +; CHECK-LABEL: test_FMADD_ASSOC_EXT1:
> > +; CHECK: fmadd
> > +; CHECK-NEXT: fmadd
> > +; CHECK-NEXT: blr
> > +
> > +; CHECK-VSX-LABEL: test_FMADD_ASSOC_EXT1:
> > +; CHECK-VSX: xsmaddmdp
> > +; CHECK-VSX-NEXT: xsmaddadp
> > +; CHECK-VSX-NEXT: blr
> > +}
> > +
> > +define double @test_FMADD_ASSOC_EXT2(float %A, float %B, float %C,
> > + float %D, double %E) {
> > + %F = fmul float %A, %B ; <float> [#uses=1]
> > + %G = fmul float %C, %D ; <float> [#uses=1]
> > + %H = fadd float %F, %G ; <float> [#uses=1]
> > + %I = fpext float %H to double ; <double> [#uses=1]
> > + %J = fadd double %I, %E ; <double> [#uses=1]
> > + ret double %J
> > +; CHECK-LABEL: test_FMADD_ASSOC_EXT2:
> > +; CHECK: fmadd
> > +; CHECK-NEXT: fmadd
> > +; CHECK-NEXT: blr
> > +
> > +; CHECK-VSX-LABEL: test_FMADD_ASSOC_EXT2:
> > +; CHECK-VSX: xsmaddmdp
> > +; CHECK-VSX-NEXT: xsmaddadp
> > +; CHECK-VSX-NEXT: fmr
> > +; CHECK-VSX-NEXT: blr
> > +}
> > +
> > +define double @test_FMADD_ASSOC_EXT3(float %A, float %B, double
> > %C,
> > + double %D, double %E) {
> > + %F = fmul float %A, %B ; <float> [#uses=1]
> > + %G = fpext float %F to double ; <double> [#uses=1]
> > + %H = fmul double %C, %D ; <double> [#uses=1]
> > + %I = fadd double %H, %G ; <double> [#uses=1]
> > + %J = fadd double %E, %I ; <double> [#uses=1]
> > + ret double %J
> > +; CHECK-LABEL: test_FMADD_ASSOC_EXT3:
> > +; CHECK: fmadd
> > +; CHECK-NEXT: fmadd
> > +; CHECK-NEXT: blr
> > +
> > +; CHECK-VSX-LABEL: test_FMADD_ASSOC_EXT3:
> > +; CHECK-VSX: xsmaddmdp
> > +; CHECK-VSX-NEXT: xsmaddadp
> > +; CHECK-VSX-NEXT: blr
> > +}
> > +
> > +define double @test_FMADD_ASSOC_EXT4(float %A, float %B, float %C,
> > + float %D, double %E) {
> > + %F = fmul float %A, %B ; <float> [#uses=1]
> > + %G = fmul float %C, %D ; <float> [#uses=1]
> > + %H = fadd float %F, %G ; <float> [#uses=1]
> > + %I = fpext float %H to double ; <double> [#uses=1]
> > + %J = fadd double %E, %I ; <double> [#uses=1]
> > + ret double %J
> > +; CHECK-LABEL: test_FMADD_ASSOC_EXT4:
> > +; CHECK: fmadd
> > +; CHECK-NEXT: fmadd
> > +; CHECK-NEXT: blr
> > +
> > +; CHECK-VSX-LABEL: test_FMADD_ASSOC_EXT4:
> > +; CHECK-VSX: xsmaddmdp
> > +; CHECK-VSX-NEXT: xsmaddadp
> > +; CHECK-VSX-NEXT: fmr
> > +; CHECK-VSX-NEXT: blr
> > +}
> > +
> > +define double @test_FMSUB_ASSOC_EXT1(float %A, float %B, double
> > %C,
> > + double %D, double %E) {
> > + %F = fmul float %A, %B ; <float> [#uses=1]
> > + %G = fpext float %F to double ; <double> [#uses=1]
> > + %H = fmul double %C, %D ; <double> [#uses=1]
> > + %I = fadd double %H, %G ; <double> [#uses=1]
> > + %J = fsub double %I, %E ; <double> [#uses=1]
> > + ret double %J
> > +; CHECK-LABEL: test_FMSUB_ASSOC_EXT1:
> > +; CHECK: fmsub
> > +; CHECK-NEXT: fmadd
> > +; CHECK-NEXT: blr
> > +
> > +; CHECK-VSX-LABEL: test_FMSUB_ASSOC_EXT1:
> > +; CHECK-VSX: xsmsubmdp
> > +; CHECK-VSX-NEXT: xsmaddadp
> > +; CHECK-VSX-NEXT: blr
> > +}
> > +
> > +define double @test_FMSUB_ASSOC_EXT2(float %A, float %B, float %C,
> > + float %D, double %E) {
> > + %F = fmul float %A, %B ; <float> [#uses=1]
> > + %G = fmul float %C, %D ; <float> [#uses=1]
> > + %H = fadd float %F, %G ; <float> [#uses=1]
> > + %I = fpext float %H to double ; <double> [#uses=1]
> > + %J = fsub double %I, %E ; <double> [#uses=1]
> > + ret double %J
> > +; CHECK-LABEL: test_FMSUB_ASSOC_EXT2:
> > +; CHECK: fmsub
> > +; CHECK-NEXT: fmadd
> > +; CHECK-NEXT: blr
> > +
> > +; CHECK-VSX-LABEL: test_FMSUB_ASSOC_EXT2:
> > +; CHECK-VSX: xsmsubmdp
> > +; CHECK-VSX-NEXT: xsmaddadp
> > +; CHECK-VSX-NEXT: fmr
> > +; CHECK-VSX-NEXT: blr
> > +}
> > +
> > +define double @test_FMSUB_ASSOC_EXT3(float %A, float %B, double
> > %C,
> > + double %D, double %E) {
> > + %F = fmul float %A, %B ; <float> [#uses=1]
> > + %G = fpext float %F to double ; <double> [#uses=1]
> > + %H = fmul double %C, %D ; <double> [#uses=1]
> > + %I = fadd double %H, %G ; <double> [#uses=1]
> > + %J = fsub double %E, %I ; <double> [#uses=1]
> > + ret double %J
> > +; CHECK-LABEL: test_FMSUB_ASSOC_EXT3:
> > +; CHECK: fnmsub
> > +; CHECK-NEXT: fnmsub
> > +; CHECK-NEXT: blr
> > +
> > +; CHECK-VSX-LABEL: test_FMSUB_ASSOC_EXT3:
> > +; CHECK-VSX: xsnmsubmdp
> > +; CHECK-VSX-NEXT: xsnmsubadp
> > +; CHECK-VSX-NEXT: fmr
> > +; CHECK-VSX-NEXT: blr
> > +}
> > +
> > +define double @test_FMSUB_ASSOC_EXT4(float %A, float %B, float %C,
> > + float %D, double %E) {
> > + %F = fmul float %A, %B ; <float> [#uses=1]
> > + %G = fmul float %C, %D ; <float> [#uses=1]
> > + %H = fadd float %F, %G ; <float> [#uses=1]
> > + %I = fpext float %H to double ; <double> [#uses=1]
> > + %J = fsub double %E, %I ; <double> [#uses=1]
> > + ret double %J
> > +; CHECK-LABEL: test_FMSUB_ASSOC_EXT4:
> > +; CHECK: fnmsub
> > +; CHECK-NEXT: fnmsub
> > +; CHECK-NEXT: blr
> > +
> > +; CHECK-VSX-LABEL: test_FMSUB_ASSOC_EXT4:
> > +; CHECK-VSX: xsnmsubmdp
> > +; CHECK-VSX-NEXT: xsnmsubadp
> > +; CHECK-VSX-NEXT: fmr
> > +; CHECK-VSX-NEXT: blr
> > +}
> > \ No newline at end of file
> > 
> > Modified: llvm/trunk/test/CodeGen/PowerPC/fma-ext.ll
> > URL:
> > http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/PowerPC/fma-ext.ll?rev=234513&r1=234512&r2=234513&view=diff
> > ==============================================================================
> > --- llvm/trunk/test/CodeGen/PowerPC/fma-ext.ll (original)
> > +++ llvm/trunk/test/CodeGen/PowerPC/fma-ext.ll Thu Apr 9 12:55:26
> > 2015
> > @@ -65,13 +65,11 @@ define double @test_FMSUB_EXT3(float %A,
> > %G = fsub double %F, %C ; <double> [#uses=1]
> > ret double %G
> > ; CHECK-LABEL: test_FMSUB_EXT3:
> > -; CHECK: fneg
> > -; CHECK-NEXT: fmsub
> > +; CHECK: fnmadd
> > ; CHECK-NEXT: blr
> > 
> > ; CHECK-VSX-LABEL: test_FMSUB_EXT3:
> > -; CHECK-VSX: xsnegdp
> > -; CHECK-VSX-NEXT: xsmsubmdp
> > +; CHECK-VSX: xsnmaddmdp
> > ; CHECK-VSX-NEXT: blr
> > }
> > 
> > @@ -82,12 +80,10 @@ define double @test_FMSUB_EXT4(float %A,
> > %G = fsub double %F, %C ; <double> [#uses=1]
> > ret double %G
> > ; CHECK-LABEL: test_FMSUB_EXT4:
> > -; CHECK: fneg
> > -; CHECK-NEXT: fmsub
> > +; CHECK: fnmadd
> > ; CHECK-NEXT: blr
> > 
> > ; CHECK-VSX-LABEL: test_FMSUB_EXT4:
> > -; CHECK-VSX: xsnegdp
> > -; CHECK-VSX-NEXT: xsmsubmdp
> > +; CHECK-VSX: xsnmaddmdp
> > ; CHECK-VSX-NEXT: blr
> > -}
> > \ No newline at end of file
> > +}
> > 
> > 
> > _______________________________________________
> > llvm-commits mailing list
> > llvm-commits at cs.uiuc.edu
> > http://lists.cs.uiuc.edu/mailman/listinfo/llvm-commits
> _______________________________________________
> llvm-commits mailing list
> llvm-commits at cs.uiuc.edu
> http://lists.cs.uiuc.edu/mailman/listinfo/llvm-commits
> 
> 
> _______________________________________________
> llvm-commits mailing list
> llvm-commits at cs.uiuc.edu
> http://lists.cs.uiuc.edu/mailman/listinfo/llvm-commits
> 

-- 
Hal Finkel
Assistant Computational Scientist
Leadership Computing Facility
Argonne National Laboratory




More information about the llvm-commits mailing list