[llvm] 352fef3 - [InstCombine] Negator - sink sinkable negations

Kostya Serebryany via llvm-commits llvm-commits at lists.llvm.org
Wed Apr 22 16:41:31 PDT 2020


Hi,

This change causes our build bots to go red, indicating a potential
performance regression.
I've provided details in https://reviews.llvm.org/D68408
Please take a look.

--kcc

On Tue, Apr 21, 2020 at 12:00 PM Roman Lebedev via llvm-commits <
llvm-commits at lists.llvm.org> wrote:

>
> Author: Roman Lebedev
> Date: 2020-04-21T22:00:23+03:00
> New Revision: 352fef3f11f5ccb2ddc8e16cecb7302a54721e9f
>
> URL:
> https://github.com/llvm/llvm-project/commit/352fef3f11f5ccb2ddc8e16cecb7302a54721e9f
> DIFF:
> https://github.com/llvm/llvm-project/commit/352fef3f11f5ccb2ddc8e16cecb7302a54721e9f.diff
>
> LOG: [InstCombine] Negator - sink sinkable negations
>
> Summary:
> As we have discussed previously (e.g. in D63992 / D64090 / [[
> https://bugs.llvm.org/show_bug.cgi?id=42457 | PR42457 ]]), `sub`
> instruction
> can almost be considered non-canonical. While we do convert `sub %x, C` ->
> `add %x, -C`,
> we sparsely do that for non-constants. But we should.
>
> Here, i propose to interpret `sub %x, %y` as `add (sub 0, %y), %x` IFF the
> negation can be sinked into the `%y`
>
> This has some potential to cause endless combine loops (either around
> PHI's, or if there are some opposite transforms).
> For former there's `-instcombine-negator-max-depth` option to mitigate it,
> should this expose any such issues
> For latter, if there are still any such opposing folds, we'd need to
> remove the colliding fold.
> In any case, reproducers welcomed!
>
> Reviewers: spatel, nikic, efriedma, xbolva00
>
> Reviewed By: spatel
>
> Subscribers: xbolva00, mgorny, hiraditya, reames, llvm-commits
>
> Tags: #llvm
>
> Differential Revision: https://reviews.llvm.org/D68408
>
> Added:
>     llvm/lib/Transforms/InstCombine/InstCombineNegator.cpp
>
> Modified:
>     llvm/lib/Transforms/InstCombine/CMakeLists.txt
>     llvm/lib/Transforms/InstCombine/InstCombineAddSub.cpp
>     llvm/lib/Transforms/InstCombine/InstCombineInternal.h
>     llvm/lib/Transforms/InstCombine/InstructionCombining.cpp
>     llvm/test/Transforms/InstCombine/and-or-icmps.ll
>     llvm/test/Transforms/InstCombine/fold-sub-of-not-to-inc-of-add.ll
>     llvm/test/Transforms/InstCombine/high-bit-signmask-with-trunc.ll
>     llvm/test/Transforms/InstCombine/high-bit-signmask.ll
>     llvm/test/Transforms/InstCombine/sub-of-negatible.ll
>     llvm/test/Transforms/InstCombine/sub.ll
>     llvm/test/Transforms/InstCombine/zext-bool-add-sub.ll
>
> Removed:
>
>
>
>
> ################################################################################
> diff  --git a/llvm/lib/Transforms/InstCombine/CMakeLists.txt
> b/llvm/lib/Transforms/InstCombine/CMakeLists.txt
> index 2f19882c3316..1a34f22f064f 100644
> --- a/llvm/lib/Transforms/InstCombine/CMakeLists.txt
> +++ b/llvm/lib/Transforms/InstCombine/CMakeLists.txt
> @@ -12,6 +12,7 @@ add_llvm_component_library(LLVMInstCombine
>    InstCombineCompares.cpp
>    InstCombineLoadStoreAlloca.cpp
>    InstCombineMulDivRem.cpp
> +  InstCombineNegator.cpp
>    InstCombinePHI.cpp
>    InstCombineSelect.cpp
>    InstCombineShifts.cpp
>
> diff  --git a/llvm/lib/Transforms/InstCombine/InstCombineAddSub.cpp
> b/llvm/lib/Transforms/InstCombine/InstCombineAddSub.cpp
> index 7ca287f07a11..16666fe9430e 100644
> --- a/llvm/lib/Transforms/InstCombine/InstCombineAddSub.cpp
> +++ b/llvm/lib/Transforms/InstCombine/InstCombineAddSub.cpp
> @@ -1682,12 +1682,10 @@ Instruction *InstCombiner::visitSub(BinaryOperator
> &I) {
>    if (Instruction *X = foldVectorBinop(I))
>      return X;
>
> -  // (A*B)-(A*C) -> A*(B-C) etc
> -  if (Value *V = SimplifyUsingDistributiveLaws(I))
> -    return replaceInstUsesWith(I, V);
> +  Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
>
>    // If this is a 'B = x-(-A)', change to B = x+A.
> -  Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
> +  // We deal with this without involving Negator to preserve NSW flag.
>    if (Value *V = dyn_castNegVal(Op1)) {
>      BinaryOperator *Res = BinaryOperator::CreateAdd(Op0, V);
>
> @@ -1704,6 +1702,45 @@ Instruction *InstCombiner::visitSub(BinaryOperator
> &I) {
>      return Res;
>    }
>
> +  auto TryToNarrowDeduceFlags = [this, &I, &Op0, &Op1]() -> Instruction *
> {
> +    if (Instruction *Ext = narrowMathIfNoOverflow(I))
> +      return Ext;
> +
> +    bool Changed = false;
> +    if (!I.hasNoSignedWrap() && willNotOverflowSignedSub(Op0, Op1, I)) {
> +      Changed = true;
> +      I.setHasNoSignedWrap(true);
> +    }
> +    if (!I.hasNoUnsignedWrap() && willNotOverflowUnsignedSub(Op0, Op1,
> I)) {
> +      Changed = true;
> +      I.setHasNoUnsignedWrap(true);
> +    }
> +
> +    return Changed ? &I : nullptr;
> +  };
> +
> +  // First, let's try to interpret `sub a, b` as `add a, (sub 0, b)`,
> +  // and let's try to sink `(sub 0, b)` into `b` itself. But only if this
> isn't
> +  // a pure negation used by a select that looks like abs/nabs.
> +  bool IsNegation = match(Op0, m_ZeroInt());
> +  if (!IsNegation || none_of(I.users(), [&I, Op1](const User *U) {
> +        const Instruction *UI = dyn_cast<Instruction>(U);
> +        if (!UI)
> +          return false;
> +        return match(UI,
> +                     m_Select(m_Value(), m_Specific(Op1),
> m_Specific(&I))) ||
> +               match(UI, m_Select(m_Value(), m_Specific(&I),
> m_Specific(Op1)));
> +      })) {
> +    if (Value *NegOp1 = Negator::Negate(IsNegation, Op1, *this))
> +      return BinaryOperator::CreateAdd(NegOp1, Op0);
> +  }
> +  if (IsNegation)
> +    return TryToNarrowDeduceFlags(); // Should have been handled in
> Negator!
> +
> +  // (A*B)-(A*C) -> A*(B-C) etc
> +  if (Value *V = SimplifyUsingDistributiveLaws(I))
> +    return replaceInstUsesWith(I, V);
> +
>    if (I.getType()->isIntOrIntVectorTy(1))
>      return BinaryOperator::CreateXor(Op0, Op1);
>
> @@ -1720,22 +1757,7 @@ Instruction *InstCombiner::visitSub(BinaryOperator
> &I) {
>    if (match(Op0, m_OneUse(m_Add(m_Value(X), m_AllOnes()))))
>      return BinaryOperator::CreateAdd(Builder.CreateNot(Op1), X);
>
> -  // Y - (X + 1) --> ~X + Y
> -  if (match(Op1, m_OneUse(m_Add(m_Value(X), m_One()))))
> -    return BinaryOperator::CreateAdd(Builder.CreateNot(X), Op0);
> -
> -  // Y - ~X --> (X + 1) + Y
> -  if (match(Op1, m_OneUse(m_Not(m_Value(X))))) {
> -    return BinaryOperator::CreateAdd(
> -        Builder.CreateAdd(Op0, ConstantInt::get(I.getType(), 1)), X);
> -  }
> -
>    if (Constant *C = dyn_cast<Constant>(Op0)) {
> -    // -f(x) -> f(-x) if possible.
> -    if (match(C, m_Zero()))
> -      if (Value *Neg = freelyNegateValue(Op1))
> -        return replaceInstUsesWith(I, Neg);
> -
>      Value *X;
>      if (match(Op1, m_ZExt(m_Value(X))) &&
> X->getType()->isIntOrIntVectorTy(1))
>        // C - (zext bool) --> bool ? C - 1 : C
> @@ -1770,26 +1792,12 @@ Instruction *InstCombiner::visitSub(BinaryOperator
> &I) {
>    }
>
>    const APInt *Op0C;
> -  if (match(Op0, m_APInt(Op0C))) {
> -    if (Op0C->isNullValue() && Op1->hasOneUse()) {
> -      Value *LHS, *RHS;
> -      SelectPatternFlavor SPF = matchSelectPattern(Op1, LHS, RHS).Flavor;
> -      if (SPF == SPF_ABS || SPF == SPF_NABS) {
> -        // This is a negate of an ABS/NABS pattern. Just swap the operands
> -        // of the select.
> -        cast<SelectInst>(Op1)->swapValues();
> -        // Don't swap prof metadata, we didn't change the branch behavior.
> -        return replaceInstUsesWith(I, Op1);
> -      }
> -    }
> -
> +  if (match(Op0, m_APInt(Op0C)) && Op0C->isMask()) {
>      // Turn this into a xor if LHS is 2^n-1 and the remaining bits are
> known
>      // zero.
> -    if (Op0C->isMask()) {
> -      KnownBits RHSKnown = computeKnownBits(Op1, 0, &I);
> -      if ((*Op0C | RHSKnown.Zero).isAllOnesValue())
> -        return BinaryOperator::CreateXor(Op1, Op0);
> -    }
> +    KnownBits RHSKnown = computeKnownBits(Op1, 0, &I);
> +    if ((*Op0C | RHSKnown.Zero).isAllOnesValue())
> +      return BinaryOperator::CreateXor(Op1, Op0);
>    }
>
>    {
> @@ -1919,49 +1927,6 @@ Instruction *InstCombiner::visitSub(BinaryOperator
> &I) {
>      return BinaryOperator::CreateAnd(
>          Op0, Builder.CreateNot(Y, Y->getName() + ".not"));
>
> -  if (Op1->hasOneUse()) {
> -    Value *Y = nullptr, *Z = nullptr;
> -    Constant *C = nullptr;
> -
> -    // (X - (Y - Z))  -->  (X + (Z - Y)).
> -    if (match(Op1, m_Sub(m_Value(Y), m_Value(Z))))
> -      return BinaryOperator::CreateAdd(Op0,
> -                                      Builder.CreateSub(Z, Y,
> Op1->getName()));
> -
> -    // Subtracting -1/0 is the same as adding 1/0:
> -    // sub [nsw] Op0, sext(bool Y) -> add [nsw] Op0, zext(bool Y)
> -    // 'nuw' is dropped in favor of the canonical form.
> -    if (match(Op1, m_SExt(m_Value(Y))) &&
> -        Y->getType()->getScalarSizeInBits() == 1) {
> -      Value *Zext = Builder.CreateZExt(Y, I.getType());
> -      BinaryOperator *Add = BinaryOperator::CreateAdd(Op0, Zext);
> -      Add->setHasNoSignedWrap(I.hasNoSignedWrap());
> -      return Add;
> -    }
> -    // sub [nsw] X, zext(bool Y) -> add [nsw] X, sext(bool Y)
> -    // 'nuw' is dropped in favor of the canonical form.
> -    if (match(Op1, m_ZExt(m_Value(Y))) &&
> Y->getType()->isIntOrIntVectorTy(1)) {
> -      Value *Sext = Builder.CreateSExt(Y, I.getType());
> -      BinaryOperator *Add = BinaryOperator::CreateAdd(Op0, Sext);
> -      Add->setHasNoSignedWrap(I.hasNoSignedWrap());
> -      return Add;
> -    }
> -
> -    // X - A*-B -> X + A*B
> -    // X - -A*B -> X + A*B
> -    Value *A, *B;
> -    if (match(Op1, m_c_Mul(m_Value(A), m_Neg(m_Value(B)))))
> -      return BinaryOperator::CreateAdd(Op0, Builder.CreateMul(A, B));
> -
> -    // X - A*C -> X + A*-C
> -    // No need to handle commuted multiply because multiply handling will
> -    // ensure constant will be move to the right hand side.
> -    if (match(Op1, m_Mul(m_Value(A), m_Constant(C))) &&
> !isa<ConstantExpr>(C)) {
> -      Value *NewMul = Builder.CreateMul(A, ConstantExpr::getNeg(C));
> -      return BinaryOperator::CreateAdd(Op0, NewMul);
> -    }
> -  }
> -
>    {
>      // ~A - Min/Max(~A, O) -> Max/Min(A, ~O) - A
>      // ~A - Min/Max(O, ~A) -> Max/Min(A, ~O) - A
> @@ -2036,20 +2001,7 @@ Instruction *InstCombiner::visitSub(BinaryOperator
> &I) {
>
>  canonicalizeCondSignextOfHighBitExtractToSignextHighBitExtract(I))
>      return V;
>
> -  if (Instruction *Ext = narrowMathIfNoOverflow(I))
> -    return Ext;
> -
> -  bool Changed = false;
> -  if (!I.hasNoSignedWrap() && willNotOverflowSignedSub(Op0, Op1, I)) {
> -    Changed = true;
> -    I.setHasNoSignedWrap(true);
> -  }
> -  if (!I.hasNoUnsignedWrap() && willNotOverflowUnsignedSub(Op0, Op1, I)) {
> -    Changed = true;
> -    I.setHasNoUnsignedWrap(true);
> -  }
> -
> -  return Changed ? &I : nullptr;
> +  return TryToNarrowDeduceFlags();
>  }
>
>  /// This eliminates floating-point negation in either 'fneg(X)' or
>
> diff  --git a/llvm/lib/Transforms/InstCombine/InstCombineInternal.h
> b/llvm/lib/Transforms/InstCombine/InstCombineInternal.h
> index 6544fd4ee8da..a908349eaff1 100644
> --- a/llvm/lib/Transforms/InstCombine/InstCombineInternal.h
> +++ b/llvm/lib/Transforms/InstCombine/InstCombineInternal.h
> @@ -16,6 +16,7 @@
>  #define LLVM_LIB_TRANSFORMS_INSTCOMBINE_INSTCOMBINEINTERNAL_H
>
>  #include "llvm/ADT/ArrayRef.h"
> +#include "llvm/ADT/Statistic.h"
>  #include "llvm/Analysis/AliasAnalysis.h"
>  #include "llvm/Analysis/InstructionSimplify.h"
>  #include "llvm/Analysis/TargetFolder.h"
> @@ -471,7 +472,6 @@ class LLVM_LIBRARY_VISIBILITY InstCombiner
>    bool shouldChangeType(unsigned FromBitWidth, unsigned ToBitWidth) const;
>    bool shouldChangeType(Type *From, Type *To) const;
>    Value *dyn_castNegVal(Value *V) const;
> -  Value *freelyNegateValue(Value *V);
>    Type *FindElementAtOffset(PointerType *PtrTy, int64_t Offset,
>                              SmallVectorImpl<Value *> &NewIndices);
>
> @@ -513,7 +513,7 @@ class LLVM_LIBRARY_VISIBILITY InstCombiner
>    Instruction *simplifyMaskedStore(IntrinsicInst &II);
>    Instruction *simplifyMaskedGather(IntrinsicInst &II);
>    Instruction *simplifyMaskedScatter(IntrinsicInst &II);
> -
> +
>    /// Transform (zext icmp) to bitwise / integer operations in order to
>    /// eliminate it.
>    ///
> @@ -1014,6 +1014,55 @@ class LLVM_LIBRARY_VISIBILITY InstCombiner
>    Value *Descale(Value *Val, APInt Scale, bool &NoSignedWrap);
>  };
>
> +namespace {
> +
> +// As a default, let's assume that we want to be aggressive,
> +// and attempt to traverse with no limits in attempt to sink negation.
> +static constexpr unsigned NegatorDefaultMaxDepth = ~0U;
> +
> +// Let's guesstimate that most often we will end up visiting/producing
> +// fairly small number of new instructions.
> +static constexpr unsigned NegatorMaxNodesSSO = 16;
> +
> +} // namespace
> +
> +class Negator final {
> +  /// Top-to-bottom, def-to-use negated instruction tree we produced.
> +  SmallVector<Instruction *, NegatorMaxNodesSSO> NewInstructions;
> +
> +  using BuilderTy = IRBuilder<TargetFolder, IRBuilderCallbackInserter>;
> +  BuilderTy Builder;
> +
> +  const bool IsTrulyNegation;
> +
> +  Negator(LLVMContext &C, const DataLayout &DL, bool IsTrulyNegation);
> +
> +#if LLVM_ENABLE_STATS
> +  unsigned NumValuesVisitedInThisNegator = 0;
> +  ~Negator();
> +#endif
> +
> +  using Result = std::pair<ArrayRef<Instruction *> /*NewInstructions*/,
> +                           Value * /*NegatedRoot*/>;
> +
> +  LLVM_NODISCARD Value *visit(Value *V, unsigned Depth);
> +
> +  /// Recurse depth-first and attempt to sink the negation.
> +  /// FIXME: use worklist?
> +  LLVM_NODISCARD Optional<Result> run(Value *Root);
> +
> +  Negator(const Negator &) = delete;
> +  Negator(Negator &&) = delete;
> +  Negator &operator=(const Negator &) = delete;
> +  Negator &operator=(Negator &&) = delete;
> +
> +public:
> +  /// Attempt to negate \p Root. Retuns nullptr if negation can't be
> performed,
> +  /// otherwise returns negated value.
> +  LLVM_NODISCARD static Value *Negate(bool LHSIsZero, Value *Root,
> +                                      InstCombiner &IC);
> +};
> +
>  } // end namespace llvm
>
>  #undef DEBUG_TYPE
>
> diff  --git a/llvm/lib/Transforms/InstCombine/InstCombineNegator.cpp
> b/llvm/lib/Transforms/InstCombine/InstCombineNegator.cpp
> new file mode 100644
> index 000000000000..2655ef304787
> --- /dev/null
> +++ b/llvm/lib/Transforms/InstCombine/InstCombineNegator.cpp
> @@ -0,0 +1,377 @@
> +//===- InstCombineNegator.cpp -----------------------------------*- C++
> -*-===//
> +//
> +// Part of the LLVM Project, under the Apache License v2.0 with LLVM
> Exceptions.
> +// See https://llvm.org/LICENSE.txt for license information.
> +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
> +//
>
> +//===----------------------------------------------------------------------===//
> +//
> +// This file implements sinking of negation into expression trees,
> +// as long as that can be done without increasing instruction count.
> +//
>
> +//===----------------------------------------------------------------------===//
> +
> +#include "InstCombineInternal.h"
> +#include "llvm/ADT/APInt.h"
> +#include "llvm/ADT/ArrayRef.h"
> +#include "llvm/ADT/None.h"
> +#include "llvm/ADT/Optional.h"
> +#include "llvm/ADT/STLExtras.h"
> +#include "llvm/ADT/SmallVector.h"
> +#include "llvm/ADT/Statistic.h"
> +#include "llvm/ADT/StringRef.h"
> +#include "llvm/ADT/Twine.h"
> +#include "llvm/ADT/iterator_range.h"
> +#include "llvm/Analysis/TargetFolder.h"
> +#include "llvm/Analysis/ValueTracking.h"
> +#include "llvm/IR/Constant.h"
> +#include "llvm/IR/Constants.h"
> +#include "llvm/IR/DebugLoc.h"
> +#include "llvm/IR/DerivedTypes.h"
> +#include "llvm/IR/IRBuilder.h"
> +#include "llvm/IR/Instruction.h"
> +#include "llvm/IR/Instructions.h"
> +#include "llvm/IR/PatternMatch.h"
> +#include "llvm/IR/Type.h"
> +#include "llvm/IR/Use.h"
> +#include "llvm/IR/User.h"
> +#include "llvm/IR/Value.h"
> +#include "llvm/Support/Casting.h"
> +#include "llvm/Support/CommandLine.h"
> +#include "llvm/Support/Compiler.h"
> +#include "llvm/Support/DebugCounter.h"
> +#include "llvm/Support/ErrorHandling.h"
> +#include "llvm/Support/raw_ostream.h"
> +#include <functional>
> +#include <tuple>
> +#include <utility>
> +
> +using namespace llvm;
> +
> +#define DEBUG_TYPE "instcombine"
> +
> +STATISTIC(NegatorTotalNegationsAttempted,
> +          "Negator: Number of negations attempted to be sinked");
> +STATISTIC(NegatorNumTreesNegated,
> +          "Negator: Number of negations successfully sinked");
> +STATISTIC(NegatorMaxDepthVisited, "Negator: Maximal traversal depth ever "
> +                                  "reached while attempting to sink
> negation");
> +STATISTIC(NegatorTimesDepthLimitReached,
> +          "Negator: How many times did the traversal depth limit was
> reached "
> +          "during sinking");
> +STATISTIC(
> +    NegatorNumValuesVisited,
> +    "Negator: Total number of values visited during attempts to sink
> negation");
> +STATISTIC(NegatorMaxTotalValuesVisited,
> +          "Negator: Maximal number of values ever visited while
> attempting to "
> +          "sink negation");
> +STATISTIC(NegatorNumInstructionsCreatedTotal,
> +          "Negator: Number of new negated instructions created, total");
> +STATISTIC(NegatorMaxInstructionsCreated,
> +          "Negator: Maximal number of new instructions created during
> negation "
> +          "attempt");
> +STATISTIC(NegatorNumInstructionsNegatedSuccess,
> +          "Negator: Number of new negated instructions created in
> successful "
> +          "negation sinking attempts");
> +
> +DEBUG_COUNTER(NegatorCounter, "instcombine-negator",
> +              "Controls Negator transformations in InstCombine pass");
> +
> +static cl::opt<bool>
> +    NegatorEnabled("instcombine-negator-enabled", cl::init(true),
> +                   cl::desc("Should we attempt to sink negations?"));
> +
> +static cl::opt<unsigned>
> +    NegatorMaxDepth("instcombine-negator-max-depth",
> +                    cl::init(NegatorDefaultMaxDepth),
> +                    cl::desc("What is the maximal lookup depth when
> trying to "
> +                             "check for viability of negation sinking."));
> +
> +Negator::Negator(LLVMContext &C, const DataLayout &DL, bool
> IsTrulyNegation_)
> +    : Builder(C, TargetFolder(DL),
> +              IRBuilderCallbackInserter([&](Instruction *I) {
> +                ++NegatorNumInstructionsCreatedTotal;
> +                NewInstructions.push_back(I);
> +              })),
> +      IsTrulyNegation(IsTrulyNegation_) {}
> +
> +#if LLVM_ENABLE_STATS
> +Negator::~Negator() {
> +  NegatorMaxTotalValuesVisited.updateMax(NumValuesVisitedInThisNegator);
> +}
> +#endif
> +
> +// FIXME: can this be reworked into a worklist-based algorithm while
> preserving
> +// the depth-first, early bailout traversal?
> +LLVM_NODISCARD Value *Negator::visit(Value *V, unsigned Depth) {
> +  NegatorMaxDepthVisited.updateMax(Depth);
> +  ++NegatorNumValuesVisited;
> +
> +#if LLVM_ENABLE_STATS
> +  ++NumValuesVisitedInThisNegator;
> +#endif
> +
> +  // In i1, negation can simply be ignored.
> +  if (V->getType()->isIntOrIntVectorTy(1))
> +    return V;
> +
> +  Value *X;
> +
> +  // -(-(X)) -> X.
> +  if (match(V, m_Neg(m_Value(X))))
> +    return X;
> +
> +  // Integral constants can be freely negated.
> +  if (match(V, m_AnyIntegralConstant()))
> +    return ConstantExpr::getNeg(cast<Constant>(V), /*HasNUW=*/false,
> +                                /*HasNSW=*/false);
> +
> +  // If we have a non-instruction, then give up.
> +  if (!isa<Instruction>(V))
> +    return nullptr;
> +
> +  // If we have started with a true negation (i.e. `sub 0, %y`), then if
> we've
> +  // got instruction that does not require recursive reasoning, we can
> still
> +  // negate it even if it has other uses, without increasing instruction
> count.
> +  if (!V->hasOneUse() && !IsTrulyNegation)
> +    return nullptr;
> +
> +  auto *I = cast<Instruction>(V);
> +  unsigned BitWidth = I->getType()->getScalarSizeInBits();
> +
> +  // We must preserve the insertion point and debug info that is set in
> the
> +  // builder at the time this function is called.
> +  InstCombiner::BuilderTy::InsertPointGuard Guard(Builder);
> +  // And since we are trying to negate instruction I, that tells us about
> the
> +  // insertion point and the debug info that we need to keep.
> +  Builder.SetInsertPoint(I);
> +
> +  // In some cases we can give the answer without further recursion.
> +  switch (I->getOpcode()) {
> +  case Instruction::Sub:
> +    // `sub` is always negatible.
> +    return Builder.CreateSub(I->getOperand(1), I->getOperand(0),
> +                             I->getName() + ".neg");
> +  case Instruction::Add:
> +    // `inc` is always negatible.
> +    if (match(I->getOperand(1), m_One()))
> +      return Builder.CreateNot(I->getOperand(0), I->getName() + ".neg");
> +    break;
> +  case Instruction::Xor:
> +    // `not` is always negatible.
> +    if (match(I, m_Not(m_Value(X))))
> +      return Builder.CreateAdd(X, ConstantInt::get(X->getType(), 1),
> +                               I->getName() + ".neg");
> +    break;
> +  case Instruction::AShr:
> +  case Instruction::LShr: {
> +    // Right-shift sign bit smear is negatible.
> +    const APInt *Op1Val;
> +    if (match(I->getOperand(1), m_APInt(Op1Val)) && *Op1Val == BitWidth -
> 1) {
> +      Value *BO = I->getOpcode() == Instruction::AShr
> +                      ? Builder.CreateLShr(I->getOperand(0),
> I->getOperand(1))
> +                      : Builder.CreateAShr(I->getOperand(0),
> I->getOperand(1));
> +      if (auto *NewInstr = dyn_cast<Instruction>(BO)) {
> +        NewInstr->copyIRFlags(I);
> +        NewInstr->setName(I->getName() + ".neg");
> +      }
> +      return BO;
> +    }
> +    break;
> +  }
> +  case Instruction::SDiv:
> +    // `sdiv` is negatible if divisor is not undef/INT_MIN/1.
> +    // While this is normally not behind a use-check,
> +    // let's consider division to be special since it's costly.
> +    if (!I->hasOneUse())
> +      break;
> +    if (auto *Op1C = dyn_cast<Constant>(I->getOperand(1))) {
> +      if (!Op1C->containsUndefElement() && Op1C->isNotMinSignedValue() &&
> +          Op1C->isNotOneValue()) {
> +        Value *BO =
> +            Builder.CreateSDiv(I->getOperand(0),
> ConstantExpr::getNeg(Op1C),
> +                               I->getName() + ".neg");
> +        if (auto *NewInstr = dyn_cast<Instruction>(BO))
> +          NewInstr->setIsExact(I->isExact());
> +        return BO;
> +      }
> +    }
> +    break;
> +  case Instruction::SExt:
> +  case Instruction::ZExt:
> +    // `*ext` of i1 is always negatible
> +    if (I->getOperand(0)->getType()->isIntOrIntVectorTy(1))
> +      return I->getOpcode() == Instruction::SExt
> +                 ? Builder.CreateZExt(I->getOperand(0), I->getType(),
> +                                      I->getName() + ".neg")
> +                 : Builder.CreateSExt(I->getOperand(0), I->getType(),
> +                                      I->getName() + ".neg");
> +    break;
> +  default:
> +    break; // Other instructions require recursive reasoning.
> +  }
> +
> +  // Rest of the logic is recursive, and if either the current instruction
> +  // has other uses or if it's time to give up then it's time.
> +  if (!V->hasOneUse())
> +    return nullptr;
> +  if (Depth > NegatorMaxDepth) {
> +    LLVM_DEBUG(dbgs() << "Negator: reached maximal allowed traversal
> depth in "
> +                      << *V << ". Giving up.\n");
> +    ++NegatorTimesDepthLimitReached;
> +    return nullptr;
> +  }
> +
> +  switch (I->getOpcode()) {
> +  case Instruction::PHI: {
> +    // `phi` is negatible if all the incoming values are negatible.
> +    PHINode *PHI = cast<PHINode>(I);
> +    SmallVector<Value *, 4> NegatedIncomingValues(PHI->getNumOperands());
> +    for (auto I : zip(PHI->incoming_values(), NegatedIncomingValues)) {
> +      if (!(std::get<1>(I) = visit(std::get<0>(I), Depth + 1))) // Early
> return.
> +        return nullptr;
> +    }
> +    // All incoming values are indeed negatible. Create negated PHI node.
> +    PHINode *NegatedPHI = Builder.CreatePHI(
> +        PHI->getType(), PHI->getNumOperands(), PHI->getName() + ".neg");
> +    for (auto I : zip(NegatedIncomingValues, PHI->blocks()))
> +      NegatedPHI->addIncoming(std::get<0>(I), std::get<1>(I));
> +    return NegatedPHI;
> +  }
> +  case Instruction::Select: {
> +    {
> +      // `abs`/`nabs` is always negatible.
> +      Value *LHS, *RHS;
> +      SelectPatternFlavor SPF =
> +          matchSelectPattern(I, LHS, RHS, /*CastOp=*/nullptr,
> Depth).Flavor;
> +      if (SPF == SPF_ABS || SPF == SPF_NABS) {
> +        auto *NewSelect = cast<SelectInst>(I->clone());
> +        // Just swap the operands of the select.
> +        NewSelect->swapValues();
> +        // Don't swap prof metadata, we didn't change the branch behavior.
> +        NewSelect->setName(I->getName() + ".neg");
> +        Builder.Insert(NewSelect);
> +        return NewSelect;
> +      }
> +    }
> +    // `select` is negatible if both hands of `select` are negatible.
> +    Value *NegOp1 = visit(I->getOperand(1), Depth + 1);
> +    if (!NegOp1) // Early return.
> +      return nullptr;
> +    Value *NegOp2 = visit(I->getOperand(2), Depth + 1);
> +    if (!NegOp2)
> +      return nullptr;
> +    // Do preserve the metadata!
> +    return Builder.CreateSelect(I->getOperand(0), NegOp1, NegOp2,
> +                                I->getName() + ".neg", /*MDFrom=*/I);
> +  }
> +  case Instruction::Trunc: {
> +    // `trunc` is negatible if its operand is negatible.
> +    Value *NegOp = visit(I->getOperand(0), Depth + 1);
> +    if (!NegOp) // Early return.
> +      return nullptr;
> +    return Builder.CreateTrunc(NegOp, I->getType(), I->getName() +
> ".neg");
> +  }
> +  case Instruction::Shl: {
> +    // `shl` is negatible if the first operand is negatible.
> +    Value *NegOp0 = visit(I->getOperand(0), Depth + 1);
> +    if (!NegOp0) // Early return.
> +      return nullptr;
> +    return Builder.CreateShl(NegOp0, I->getOperand(1), I->getName() +
> ".neg");
> +  }
> +  case Instruction::Add: {
> +    // `add` is negatible if both of its operands are negatible.
> +    Value *NegOp0 = visit(I->getOperand(0), Depth + 1);
> +    if (!NegOp0) // Early return.
> +      return nullptr;
> +    Value *NegOp1 = visit(I->getOperand(1), Depth + 1);
> +    if (!NegOp1)
> +      return nullptr;
> +    return Builder.CreateAdd(NegOp0, NegOp1, I->getName() + ".neg");
> +  }
> +  case Instruction::Xor:
> +    // `xor` is negatible if one of its operands is invertible.
> +    // FIXME: InstCombineInverter? But how to connect Inverter and
> Negator?
> +    if (auto *C = dyn_cast<Constant>(I->getOperand(1))) {
> +      Value *Xor = Builder.CreateXor(I->getOperand(0),
> ConstantExpr::getNot(C));
> +      return Builder.CreateAdd(Xor, ConstantInt::get(Xor->getType(), 1),
> +                               I->getName() + ".neg");
> +    }
> +    return nullptr;
> +  case Instruction::Mul: {
> +    // `mul` is negatible if one of its operands is negatible.
> +    Value *NegatedOp, *OtherOp;
> +    // First try the second operand, in case it's a constant it will be
> best to
> +    // just invert it instead of sinking the `neg` deeper.
> +    if (Value *NegOp1 = visit(I->getOperand(1), Depth + 1)) {
> +      NegatedOp = NegOp1;
> +      OtherOp = I->getOperand(0);
> +    } else if (Value *NegOp0 = visit(I->getOperand(0), Depth + 1)) {
> +      NegatedOp = NegOp0;
> +      OtherOp = I->getOperand(1);
> +    } else
> +      // Can't negate either of them.
> +      return nullptr;
> +    return Builder.CreateMul(NegatedOp, OtherOp, I->getName() + ".neg");
> +  }
> +  default:
> +    return nullptr; // Don't know, likely not negatible for free.
> +  }
> +
> +  llvm_unreachable("Can't get here. We always return from switch.");
> +};
> +
> +LLVM_NODISCARD Optional<Negator::Result> Negator::run(Value *Root) {
> +  Value *Negated = visit(Root, /*Depth=*/0);
> +  if (!Negated) {
> +    // We must cleanup newly-inserted instructions, to avoid any potential
> +    // endless combine looping.
> +    llvm::for_each(llvm::reverse(NewInstructions),
> +                   [&](Instruction *I) { I->eraseFromParent(); });
> +    return llvm::None;
> +  }
> +  return std::make_pair(ArrayRef<Instruction *>(NewInstructions),
> Negated);
> +};
> +
> +LLVM_NODISCARD Value *Negator::Negate(bool LHSIsZero, Value *Root,
> +                                      InstCombiner &IC) {
> +  ++NegatorTotalNegationsAttempted;
> +  LLVM_DEBUG(dbgs() << "Negator: attempting to sink negation into " <<
> *Root
> +                    << "\n");
> +
> +  if (!NegatorEnabled || !DebugCounter::shouldExecute(NegatorCounter))
> +    return nullptr;
> +
> +  Negator N(Root->getContext(), IC.getDataLayout(), LHSIsZero);
> +  Optional<Result> Res = N.run(Root);
> +  if (!Res) { // Negation failed.
> +    LLVM_DEBUG(dbgs() << "Negator: failed to sink negation into " << *Root
> +                      << "\n");
> +    return nullptr;
> +  }
> +
> +  LLVM_DEBUG(dbgs() << "Negator: successfully sunk negation into " <<
> *Root
> +                    << "\n         NEW: " << *Res->second << "\n");
> +  ++NegatorNumTreesNegated;
> +
> +  // We must temporarily unset the 'current' insertion point and DebugLoc
> of the
> +  // InstCombine's IRBuilder so that it won't interfere with the ones we
> have
> +  // already specified when producing negated instructions.
> +  InstCombiner::BuilderTy::InsertPointGuard Guard(IC.Builder);
> +  IC.Builder.ClearInsertionPoint();
> +  IC.Builder.SetCurrentDebugLocation(DebugLoc());
> +
> +  // And finally, we must add newly-created instructions into the
> InstCombine's
> +  // worklist (in a proper order!) so it can attempt to combine them.
> +  LLVM_DEBUG(dbgs() << "Negator: Propagating " << Res->first.size()
> +                    << " instrs to InstCombine\n");
> +  NegatorMaxInstructionsCreated.updateMax(Res->first.size());
> +  NegatorNumInstructionsNegatedSuccess += Res->first.size();
> +
> +  // They are in def-use order, so nothing fancy, just insert them in
> order.
> +  llvm::for_each(Res->first,
> +                 [&](Instruction *I) { IC.Builder.Insert(I,
> I->getName()); });
> +
> +  // And return the new root.
> +  return Res->second;
> +};
>
> diff  --git a/llvm/lib/Transforms/InstCombine/InstructionCombining.cpp
> b/llvm/lib/Transforms/InstCombine/InstructionCombining.cpp
> index 4d286f19f919..88cb8d5300b9 100644
> --- a/llvm/lib/Transforms/InstCombine/InstructionCombining.cpp
> +++ b/llvm/lib/Transforms/InstCombine/InstructionCombining.cpp
> @@ -853,120 +853,6 @@ Value *InstCombiner::dyn_castNegVal(Value *V) const {
>    return nullptr;
>  }
>
> -/// Get negated V (that is 0-V) without increasing instruction count,
> -/// assuming that the original V will become unused.
> -Value *InstCombiner::freelyNegateValue(Value *V) {
> -  if (Value *NegV = dyn_castNegVal(V))
> -    return NegV;
> -
> -  Instruction *I = dyn_cast<Instruction>(V);
> -  if (!I)
> -    return nullptr;
> -
> -  unsigned BitWidth = I->getType()->getScalarSizeInBits();
> -  switch (I->getOpcode()) {
> -  // 0-(zext i1 A)  =>  sext i1 A
> -  case Instruction::ZExt:
> -    if (I->getOperand(0)->getType()->isIntOrIntVectorTy(1))
> -      return Builder.CreateSExtOrBitCast(
> -          I->getOperand(0), I->getType(), I->getName() + ".neg");
> -    return nullptr;
> -
> -  // 0-(sext i1 A)  =>  zext i1 A
> -  case Instruction::SExt:
> -    if (I->getOperand(0)->getType()->isIntOrIntVectorTy(1))
> -      return Builder.CreateZExtOrBitCast(
> -          I->getOperand(0), I->getType(), I->getName() + ".neg");
> -    return nullptr;
> -
> -  // 0-(A lshr (BW-1))  =>  A ashr (BW-1)
> -  case Instruction::LShr:
> -    if (match(I->getOperand(1), m_SpecificInt(BitWidth - 1)))
> -      return Builder.CreateAShr(
> -          I->getOperand(0), I->getOperand(1),
> -          I->getName() + ".neg", cast<BinaryOperator>(I)->isExact());
> -    return nullptr;
> -
> -  // 0-(A ashr (BW-1))  =>  A lshr (BW-1)
> -  case Instruction::AShr:
> -    if (match(I->getOperand(1), m_SpecificInt(BitWidth - 1)))
> -      return Builder.CreateLShr(
> -          I->getOperand(0), I->getOperand(1),
> -          I->getName() + ".neg", cast<BinaryOperator>(I)->isExact());
> -    return nullptr;
> -
> -  // Negation is equivalent to bitwise-not + 1.
> -  case Instruction::Xor: {
> -    // Special case for negate of 'not' - replace with increment:
> -    // 0 - (~A)  =>  ((A ^ -1) ^ -1) + 1  =>  A + 1
> -    Value *A;
> -    if (match(I, m_Not(m_Value(A))))
> -      return Builder.CreateAdd(A, ConstantInt::get(A->getType(), 1),
> -                               I->getName() + ".neg");
> -
> -    // General case xor (not a 'not') requires creating a new xor, so
> this has a
> -    // one-use limitation:
> -    // 0 - (A ^ C)  =>  ((A ^ C) ^ -1) + 1  =>  A ^ ~C + 1
> -    Constant *C;
> -    if (match(I, m_OneUse(m_Xor(m_Value(A), m_Constant(C))))) {
> -      Value *Xor = Builder.CreateXor(A, ConstantExpr::getNot(C));
> -      return Builder.CreateAdd(Xor, ConstantInt::get(Xor->getType(), 1),
> -                               I->getName() + ".neg");
> -    }
> -    return nullptr;
> -  }
> -
> -  default:
> -    break;
> -  }
> -
> -  // TODO: The "sub" pattern below could also be applied without the
> one-use
> -  // restriction. Not allowing it for now in line with existing behavior.
> -  if (!I->hasOneUse())
> -    return nullptr;
> -
> -  switch (I->getOpcode()) {
> -  // 0-(A-B)  =>  B-A
> -  case Instruction::Sub:
> -    return Builder.CreateSub(
> -        I->getOperand(1), I->getOperand(0), I->getName() + ".neg");
> -
> -  // 0-(A sdiv C)  =>  A sdiv (0-C)  provided the negation doesn't
> overflow.
> -  case Instruction::SDiv: {
> -    Constant *C = dyn_cast<Constant>(I->getOperand(1));
> -    if (C && !C->containsUndefElement() && C->isNotMinSignedValue() &&
> -        C->isNotOneValue())
> -      return Builder.CreateSDiv(I->getOperand(0), ConstantExpr::getNeg(C),
> -          I->getName() + ".neg", cast<BinaryOperator>(I)->isExact());
> -    return nullptr;
> -  }
> -
> -  // 0-(A<<B)  =>  (0-A)<<B
> -  case Instruction::Shl:
> -    if (Value *NegA = freelyNegateValue(I->getOperand(0)))
> -      return Builder.CreateShl(NegA, I->getOperand(1), I->getName() +
> ".neg");
> -    return nullptr;
> -
> -  // 0-(trunc A)  =>  trunc (0-A)
> -  case Instruction::Trunc:
> -    if (Value *NegA = freelyNegateValue(I->getOperand(0)))
> -      return Builder.CreateTrunc(NegA, I->getType(), I->getName() +
> ".neg");
> -    return nullptr;
> -
> -  // 0-(A*B)  =>  (0-A)*B
> -  // 0-(A*B)  =>  A*(0-B)
> -  case Instruction::Mul:
> -    if (Value *NegA = freelyNegateValue(I->getOperand(0)))
> -      return Builder.CreateMul(NegA, I->getOperand(1), V->getName() +
> ".neg");
> -    if (Value *NegB = freelyNegateValue(I->getOperand(1)))
> -      return Builder.CreateMul(I->getOperand(0), NegB, V->getName() +
> ".neg");
> -    return nullptr;
> -
> -  default:
> -    return nullptr;
> -  }
> -}
> -
>  static Value *foldOperationIntoSelectOperand(Instruction &I, Value *SO,
>                                               InstCombiner::BuilderTy
> &Builder) {
>    if (auto *Cast = dyn_cast<CastInst>(&I))
>
> diff  --git a/llvm/test/Transforms/InstCombine/and-or-icmps.ll
> b/llvm/test/Transforms/InstCombine/and-or-icmps.ll
> index 175103886ff4..efc22068d9ee 100644
> --- a/llvm/test/Transforms/InstCombine/and-or-icmps.ll
> +++ b/llvm/test/Transforms/InstCombine/and-or-icmps.ll
> @@ -210,12 +210,20 @@ define void @simplify_before_foldAndOfICmps() {
>  ; CHECK-LABEL: @simplify_before_foldAndOfICmps(
>  ; CHECK-NEXT:    [[A8:%.*]] = alloca i16, align 2
>  ; CHECK-NEXT:    [[L7:%.*]] = load i16, i16* [[A8]], align 2
> -; CHECK-NEXT:    [[C10:%.*]] = icmp ult i16 [[L7]], 2
> +; CHECK-NEXT:    [[TMP1:%.*]] = icmp eq i16 [[L7]], -1
> +; CHECK-NEXT:    [[B11:%.*]] = zext i1 [[TMP1]] to i16
> +; CHECK-NEXT:    [[C10:%.*]] = icmp ugt i16 [[L7]], [[B11]]
> +; CHECK-NEXT:    [[C5:%.*]] = icmp slt i16 [[L7]], 1
> +; CHECK-NEXT:    [[C11:%.*]] = icmp ne i16 [[L7]], 0
>  ; CHECK-NEXT:    [[C7:%.*]] = icmp slt i16 [[L7]], 0
> -; CHECK-NEXT:    [[C18:%.*]] = or i1 [[C7]], [[C10]]
> -; CHECK-NEXT:    [[L7_LOBIT:%.*]] = ashr i16 [[L7]], 15
> -; CHECK-NEXT:    [[TMP1:%.*]] = sext i16 [[L7_LOBIT]] to i64
> -; CHECK-NEXT:    [[G26:%.*]] = getelementptr i1, i1* null, i64 [[TMP1]]
> +; CHECK-NEXT:    [[B15:%.*]] = xor i1 [[C7]], [[C10]]
> +; CHECK-NEXT:    [[B19:%.*]] = xor i1 [[C11]], [[B15]]
> +; CHECK-NEXT:    [[TMP2:%.*]] = and i1 [[C10]], [[C5]]
> +; CHECK-NEXT:    [[C3:%.*]] = and i1 [[TMP2]], [[B19]]
> +; CHECK-NEXT:    [[TMP3:%.*]] = xor i1 [[C10]], true
> +; CHECK-NEXT:    [[C18:%.*]] = or i1 [[C7]], [[TMP3]]
> +; CHECK-NEXT:    [[TMP4:%.*]] = sext i1 [[C3]] to i64
> +; CHECK-NEXT:    [[G26:%.*]] = getelementptr i1, i1* null, i64 [[TMP4]]
>  ; CHECK-NEXT:    store i16 [[L7]], i16* undef, align 2
>  ; CHECK-NEXT:    store i1 [[C18]], i1* undef, align 1
>  ; CHECK-NEXT:    store i1* [[G26]], i1** undef, align 8
>
> diff  --git
> a/llvm/test/Transforms/InstCombine/fold-sub-of-not-to-inc-of-add.ll
> b/llvm/test/Transforms/InstCombine/fold-sub-of-not-to-inc-of-add.ll
> index cf9d051d125b..49d4ec43768b 100644
> --- a/llvm/test/Transforms/InstCombine/fold-sub-of-not-to-inc-of-add.ll
> +++ b/llvm/test/Transforms/InstCombine/fold-sub-of-not-to-inc-of-add.ll
> @@ -13,8 +13,8 @@
>
>  define i32 @p0_scalar(i32 %x, i32 %y) {
>  ; CHECK-LABEL: @p0_scalar(
> -; CHECK-NEXT:    [[TMP1:%.*]] = add i32 [[Y:%.*]], 1
> -; CHECK-NEXT:    [[T1:%.*]] = add i32 [[TMP1]], [[X:%.*]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = add i32 [[X:%.*]], 1
> +; CHECK-NEXT:    [[T1:%.*]] = add i32 [[T0_NEG]], [[Y:%.*]]
>  ; CHECK-NEXT:    ret i32 [[T1]]
>  ;
>    %t0 = xor i32 %x, -1
> @@ -28,8 +28,8 @@ define i32 @p0_scalar(i32 %x, i32 %y) {
>
>  define <4 x i32> @p1_vector_splat(<4 x i32> %x, <4 x i32> %y) {
>  ; CHECK-LABEL: @p1_vector_splat(
> -; CHECK-NEXT:    [[TMP1:%.*]] = add <4 x i32> [[Y:%.*]], <i32 1, i32 1,
> i32 1, i32 1>
> -; CHECK-NEXT:    [[T1:%.*]] = add <4 x i32> [[TMP1]], [[X:%.*]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = add <4 x i32> [[X:%.*]], <i32 1, i32 1,
> i32 1, i32 1>
> +; CHECK-NEXT:    [[T1:%.*]] = add <4 x i32> [[T0_NEG]], [[Y:%.*]]
>  ; CHECK-NEXT:    ret <4 x i32> [[T1]]
>  ;
>    %t0 = xor <4 x i32> %x, <i32 -1, i32 -1, i32 -1, i32 -1>
> @@ -39,8 +39,8 @@ define <4 x i32> @p1_vector_splat(<4 x i32> %x, <4 x
> i32> %y) {
>
>  define <4 x i32> @p2_vector_undef(<4 x i32> %x, <4 x i32> %y) {
>  ; CHECK-LABEL: @p2_vector_undef(
> -; CHECK-NEXT:    [[TMP1:%.*]] = add <4 x i32> [[Y:%.*]], <i32 1, i32 1,
> i32 1, i32 1>
> -; CHECK-NEXT:    [[T1:%.*]] = add <4 x i32> [[TMP1]], [[X:%.*]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = add <4 x i32> [[X:%.*]], <i32 1, i32 1,
> i32 1, i32 1>
> +; CHECK-NEXT:    [[T1:%.*]] = add <4 x i32> [[T0_NEG]], [[Y:%.*]]
>  ; CHECK-NEXT:    ret <4 x i32> [[T1]]
>  ;
>    %t0 = xor <4 x i32> %x, <i32 -1, i32 -1, i32 undef, i32 -1>
> @@ -85,8 +85,8 @@ define i32 @n4(i32 %x, i32 %y) {
>
>  define i32 @n5_is_not_not(i32 %x, i32 %y) {
>  ; CHECK-LABEL: @n5_is_not_not(
> -; CHECK-NEXT:    [[T0:%.*]] = xor i32 [[X:%.*]], 2147483647
> -; CHECK-NEXT:    [[T1:%.*]] = sub i32 [[Y:%.*]], [[T0]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = add i32 [[X:%.*]], -2147483647
> +; CHECK-NEXT:    [[T1:%.*]] = add i32 [[T0_NEG]], [[Y:%.*]]
>  ; CHECK-NEXT:    ret i32 [[T1]]
>  ;
>    %t0 = xor i32 %x, 2147483647 ; not -1
>
> diff  --git
> a/llvm/test/Transforms/InstCombine/high-bit-signmask-with-trunc.ll
> b/llvm/test/Transforms/InstCombine/high-bit-signmask-with-trunc.ll
> index 034c285dd1ff..b9bce627d2ab 100644
> --- a/llvm/test/Transforms/InstCombine/high-bit-signmask-with-trunc.ll
> +++ b/llvm/test/Transforms/InstCombine/high-bit-signmask-with-trunc.ll
> @@ -3,9 +3,9 @@
>
>  define i32 @t0(i64 %x) {
>  ; CHECK-LABEL: @t0(
> -; CHECK-NEXT:    [[TMP1:%.*]] = ashr i64 [[X:%.*]], 63
> -; CHECK-NEXT:    [[R:%.*]] = trunc i64 [[TMP1]] to i32
> -; CHECK-NEXT:    ret i32 [[R]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = ashr i64 [[X:%.*]], 63
> +; CHECK-NEXT:    [[T1_NEG:%.*]] = trunc i64 [[T0_NEG]] to i32
> +; CHECK-NEXT:    ret i32 [[T1_NEG]]
>  ;
>    %t0 = lshr i64 %x, 63
>    %t1 = trunc i64 %t0 to i32
> @@ -14,9 +14,9 @@ define i32 @t0(i64 %x) {
>  }
>  define i32 @t1_exact(i64 %x) {
>  ; CHECK-LABEL: @t1_exact(
> -; CHECK-NEXT:    [[TMP1:%.*]] = ashr exact i64 [[X:%.*]], 63
> -; CHECK-NEXT:    [[R:%.*]] = trunc i64 [[TMP1]] to i32
> -; CHECK-NEXT:    ret i32 [[R]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = ashr exact i64 [[X:%.*]], 63
> +; CHECK-NEXT:    [[T1_NEG:%.*]] = trunc i64 [[T0_NEG]] to i32
> +; CHECK-NEXT:    ret i32 [[T1_NEG]]
>  ;
>    %t0 = lshr exact i64 %x, 63
>    %t1 = trunc i64 %t0 to i32
> @@ -25,9 +25,9 @@ define i32 @t1_exact(i64 %x) {
>  }
>  define i32 @t2(i64 %x) {
>  ; CHECK-LABEL: @t2(
> -; CHECK-NEXT:    [[TMP1:%.*]] = lshr i64 [[X:%.*]], 63
> -; CHECK-NEXT:    [[R:%.*]] = trunc i64 [[TMP1]] to i32
> -; CHECK-NEXT:    ret i32 [[R]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = lshr i64 [[X:%.*]], 63
> +; CHECK-NEXT:    [[T1_NEG:%.*]] = trunc i64 [[T0_NEG]] to i32
> +; CHECK-NEXT:    ret i32 [[T1_NEG]]
>  ;
>    %t0 = ashr i64 %x, 63
>    %t1 = trunc i64 %t0 to i32
> @@ -36,9 +36,9 @@ define i32 @t2(i64 %x) {
>  }
>  define i32 @t3_exact(i64 %x) {
>  ; CHECK-LABEL: @t3_exact(
> -; CHECK-NEXT:    [[TMP1:%.*]] = lshr exact i64 [[X:%.*]], 63
> -; CHECK-NEXT:    [[R:%.*]] = trunc i64 [[TMP1]] to i32
> -; CHECK-NEXT:    ret i32 [[R]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = lshr exact i64 [[X:%.*]], 63
> +; CHECK-NEXT:    [[T1_NEG:%.*]] = trunc i64 [[T0_NEG]] to i32
> +; CHECK-NEXT:    ret i32 [[T1_NEG]]
>  ;
>    %t0 = ashr exact i64 %x, 63
>    %t1 = trunc i64 %t0 to i32
> @@ -48,9 +48,9 @@ define i32 @t3_exact(i64 %x) {
>
>  define <2 x i32> @t4(<2 x i64> %x) {
>  ; CHECK-LABEL: @t4(
> -; CHECK-NEXT:    [[TMP1:%.*]] = ashr <2 x i64> [[X:%.*]], <i64 63, i64 63>
> -; CHECK-NEXT:    [[R:%.*]] = trunc <2 x i64> [[TMP1]] to <2 x i32>
> -; CHECK-NEXT:    ret <2 x i32> [[R]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = ashr <2 x i64> [[X:%.*]], <i64 63, i64
> 63>
> +; CHECK-NEXT:    [[T1_NEG:%.*]] = trunc <2 x i64> [[T0_NEG]] to <2 x i32>
> +; CHECK-NEXT:    ret <2 x i32> [[T1_NEG]]
>  ;
>    %t0 = lshr <2 x i64> %x, <i64 63, i64 63>
>    %t1 = trunc <2 x i64> %t0 to <2 x i32>
> @@ -76,11 +76,11 @@ declare void @use32(i32)
>
>  define i32 @t6(i64 %x) {
>  ; CHECK-LABEL: @t6(
> -; CHECK-NEXT:    [[T0:%.*]] = lshr i64 [[X:%.*]], 63
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = ashr i64 [[X:%.*]], 63
> +; CHECK-NEXT:    [[T0:%.*]] = lshr i64 [[X]], 63
>  ; CHECK-NEXT:    call void @use64(i64 [[T0]])
> -; CHECK-NEXT:    [[TMP1:%.*]] = ashr i64 [[X]], 63
> -; CHECK-NEXT:    [[R:%.*]] = trunc i64 [[TMP1]] to i32
> -; CHECK-NEXT:    ret i32 [[R]]
> +; CHECK-NEXT:    [[T1_NEG:%.*]] = trunc i64 [[T0_NEG]] to i32
> +; CHECK-NEXT:    ret i32 [[T1_NEG]]
>  ;
>    %t0 = lshr i64 %x, 63
>    call void @use64(i64 %t0)
> @@ -136,9 +136,9 @@ define i32 @n9(i64 %x) {
>
>  define i32 @n10(i64 %x) {
>  ; CHECK-LABEL: @n10(
> -; CHECK-NEXT:    [[T0:%.*]] = lshr i64 [[X:%.*]], 63
> -; CHECK-NEXT:    [[T1:%.*]] = trunc i64 [[T0]] to i32
> -; CHECK-NEXT:    [[R:%.*]] = xor i32 [[T1]], 1
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = ashr i64 [[X:%.*]], 63
> +; CHECK-NEXT:    [[T1_NEG:%.*]] = trunc i64 [[T0_NEG]] to i32
> +; CHECK-NEXT:    [[R:%.*]] = add i32 [[T1_NEG]], 1
>  ; CHECK-NEXT:    ret i32 [[R]]
>  ;
>    %t0 = lshr i64 %x, 63
>
> diff  --git a/llvm/test/Transforms/InstCombine/high-bit-signmask.ll
> b/llvm/test/Transforms/InstCombine/high-bit-signmask.ll
> index 4a1b395ca35f..18a87273c021 100644
> --- a/llvm/test/Transforms/InstCombine/high-bit-signmask.ll
> +++ b/llvm/test/Transforms/InstCombine/high-bit-signmask.ll
> @@ -3,8 +3,8 @@
>
>  define i64 @t0(i64 %x) {
>  ; CHECK-LABEL: @t0(
> -; CHECK-NEXT:    [[R:%.*]] = ashr i64 [[X:%.*]], 63
> -; CHECK-NEXT:    ret i64 [[R]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = ashr i64 [[X:%.*]], 63
> +; CHECK-NEXT:    ret i64 [[T0_NEG]]
>  ;
>    %t0 = lshr i64 %x, 63
>    %r = sub i64 0, %t0
> @@ -12,8 +12,8 @@ define i64 @t0(i64 %x) {
>  }
>  define i64 @t0_exact(i64 %x) {
>  ; CHECK-LABEL: @t0_exact(
> -; CHECK-NEXT:    [[R:%.*]] = ashr exact i64 [[X:%.*]], 63
> -; CHECK-NEXT:    ret i64 [[R]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = ashr exact i64 [[X:%.*]], 63
> +; CHECK-NEXT:    ret i64 [[T0_NEG]]
>  ;
>    %t0 = lshr exact i64 %x, 63
>    %r = sub i64 0, %t0
> @@ -21,8 +21,8 @@ define i64 @t0_exact(i64 %x) {
>  }
>  define i64 @t2(i64 %x) {
>  ; CHECK-LABEL: @t2(
> -; CHECK-NEXT:    [[R:%.*]] = lshr i64 [[X:%.*]], 63
> -; CHECK-NEXT:    ret i64 [[R]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = lshr i64 [[X:%.*]], 63
> +; CHECK-NEXT:    ret i64 [[T0_NEG]]
>  ;
>    %t0 = ashr i64 %x, 63
>    %r = sub i64 0, %t0
> @@ -30,8 +30,8 @@ define i64 @t2(i64 %x) {
>  }
>  define i64 @t3_exact(i64 %x) {
>  ; CHECK-LABEL: @t3_exact(
> -; CHECK-NEXT:    [[R:%.*]] = lshr exact i64 [[X:%.*]], 63
> -; CHECK-NEXT:    ret i64 [[R]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = lshr exact i64 [[X:%.*]], 63
> +; CHECK-NEXT:    ret i64 [[T0_NEG]]
>  ;
>    %t0 = ashr exact i64 %x, 63
>    %r = sub i64 0, %t0
> @@ -40,8 +40,8 @@ define i64 @t3_exact(i64 %x) {
>
>  define <2 x i64> @t4(<2 x i64> %x) {
>  ; CHECK-LABEL: @t4(
> -; CHECK-NEXT:    [[R:%.*]] = ashr <2 x i64> [[X:%.*]], <i64 63, i64 63>
> -; CHECK-NEXT:    ret <2 x i64> [[R]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = ashr <2 x i64> [[X:%.*]], <i64 63, i64
> 63>
> +; CHECK-NEXT:    ret <2 x i64> [[T0_NEG]]
>  ;
>    %t0 = lshr <2 x i64> %x, <i64 63, i64 63>
>    %r = sub <2 x i64> zeroinitializer, %t0
> @@ -64,10 +64,10 @@ declare void @use32(i64)
>
>  define i64 @t6(i64 %x) {
>  ; CHECK-LABEL: @t6(
> -; CHECK-NEXT:    [[T0:%.*]] = lshr i64 [[X:%.*]], 63
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = ashr i64 [[X:%.*]], 63
> +; CHECK-NEXT:    [[T0:%.*]] = lshr i64 [[X]], 63
>  ; CHECK-NEXT:    call void @use64(i64 [[T0]])
> -; CHECK-NEXT:    [[R:%.*]] = ashr i64 [[X]], 63
> -; CHECK-NEXT:    ret i64 [[R]]
> +; CHECK-NEXT:    ret i64 [[T0_NEG]]
>  ;
>    %t0 = lshr i64 %x, 63
>    call void @use64(i64 %t0)
> @@ -77,10 +77,10 @@ define i64 @t6(i64 %x) {
>
>  define i64 @n7(i64 %x) {
>  ; CHECK-LABEL: @n7(
> -; CHECK-NEXT:    [[T0:%.*]] = lshr i64 [[X:%.*]], 63
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = ashr i64 [[X:%.*]], 63
> +; CHECK-NEXT:    [[T0:%.*]] = lshr i64 [[X]], 63
>  ; CHECK-NEXT:    call void @use32(i64 [[T0]])
> -; CHECK-NEXT:    [[R:%.*]] = ashr i64 [[X]], 63
> -; CHECK-NEXT:    ret i64 [[R]]
> +; CHECK-NEXT:    ret i64 [[T0_NEG]]
>  ;
>    %t0 = lshr i64 %x, 63
>    call void @use32(i64 %t0)
> @@ -90,11 +90,11 @@ define i64 @n7(i64 %x) {
>
>  define i64 @n8(i64 %x) {
>  ; CHECK-LABEL: @n8(
> -; CHECK-NEXT:    [[T0:%.*]] = lshr i64 [[X:%.*]], 63
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = ashr i64 [[X:%.*]], 63
> +; CHECK-NEXT:    [[T0:%.*]] = lshr i64 [[X]], 63
>  ; CHECK-NEXT:    call void @use64(i64 [[T0]])
>  ; CHECK-NEXT:    call void @use32(i64 [[T0]])
> -; CHECK-NEXT:    [[R:%.*]] = ashr i64 [[X]], 63
> -; CHECK-NEXT:    ret i64 [[R]]
> +; CHECK-NEXT:    ret i64 [[T0_NEG]]
>  ;
>    %t0 = lshr i64 %x, 63
>    call void @use64(i64 %t0)
> @@ -116,8 +116,8 @@ define i64 @n9(i64 %x) {
>
>  define i64 @n10(i64 %x) {
>  ; CHECK-LABEL: @n10(
> -; CHECK-NEXT:    [[T0:%.*]] = lshr i64 [[X:%.*]], 63
> -; CHECK-NEXT:    [[R:%.*]] = xor i64 [[T0]], 1
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = ashr i64 [[X:%.*]], 63
> +; CHECK-NEXT:    [[R:%.*]] = add nsw i64 [[T0_NEG]], 1
>  ; CHECK-NEXT:    ret i64 [[R]]
>  ;
>    %t0 = lshr i64 %x, 63
>
> diff  --git a/llvm/test/Transforms/InstCombine/sub-of-negatible.ll
> b/llvm/test/Transforms/InstCombine/sub-of-negatible.ll
> index c256bd1ed4dc..40cb4aa29446 100644
> --- a/llvm/test/Transforms/InstCombine/sub-of-negatible.ll
> +++ b/llvm/test/Transforms/InstCombine/sub-of-negatible.ll
> @@ -31,8 +31,8 @@ define i8 @t1(i8 %x, i8 %y) {
>  ; Shift-left can be negated if all uses can be updated
>  define i8 @t2(i8 %x, i8 %y) {
>  ; CHECK-LABEL: @t2(
> -; CHECK-NEXT:    [[T0:%.*]] = shl i8 -42, [[Y:%.*]]
> -; CHECK-NEXT:    [[T1:%.*]] = sub i8 [[X:%.*]], [[T0]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = shl i8 42, [[Y:%.*]]
> +; CHECK-NEXT:    [[T1:%.*]] = add i8 [[T0_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret i8 [[T1]]
>  ;
>    %t0 = shl i8 -42, %y
> @@ -55,8 +55,8 @@ define i8 @t3(i8 %x, i8 %y, i8 %z) {
>  ; CHECK-LABEL: @t3(
>  ; CHECK-NEXT:    [[T0:%.*]] = sub i8 0, [[Z:%.*]]
>  ; CHECK-NEXT:    call void @use8(i8 [[T0]])
> -; CHECK-NEXT:    [[T1:%.*]] = shl i8 [[T0]], [[Y:%.*]]
> -; CHECK-NEXT:    [[T2:%.*]] = sub i8 [[X:%.*]], [[T1]]
> +; CHECK-NEXT:    [[T1_NEG:%.*]] = shl i8 [[Z]], [[Y:%.*]]
> +; CHECK-NEXT:    [[T2:%.*]] = add i8 [[T1_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret i8 [[T2]]
>  ;
>    %t0 = sub i8 0, %z
> @@ -85,8 +85,8 @@ define i8 @n3(i8 %x, i8 %y, i8 %z) {
>  ; Select can be negated if all it's operands can be negated and all the
> users of select can be updated
>  define i8 @t4(i8 %x, i1 %y) {
>  ; CHECK-LABEL: @t4(
> -; CHECK-NEXT:    [[T0:%.*]] = select i1 [[Y:%.*]], i8 -42, i8 44
> -; CHECK-NEXT:    [[T1:%.*]] = sub i8 [[X:%.*]], [[T0]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = select i1 [[Y:%.*]], i8 42, i8 -44
> +; CHECK-NEXT:    [[T1:%.*]] = add i8 [[T0_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret i8 [[T1]]
>  ;
>    %t0 = select i1 %y, i8 -42, i8 44
> @@ -119,8 +119,8 @@ define i8 @t6(i8 %x, i1 %y, i8 %z) {
>  ; CHECK-LABEL: @t6(
>  ; CHECK-NEXT:    [[T0:%.*]] = sub i8 0, [[Z:%.*]]
>  ; CHECK-NEXT:    call void @use8(i8 [[T0]])
> -; CHECK-NEXT:    [[T1:%.*]] = select i1 [[Y:%.*]], i8 -42, i8 [[T0]]
> -; CHECK-NEXT:    [[T2:%.*]] = sub i8 [[X:%.*]], [[T1]]
> +; CHECK-NEXT:    [[T1_NEG:%.*]] = select i1 [[Y:%.*]], i8 42, i8 [[Z]]
> +; CHECK-NEXT:    [[T2:%.*]] = add i8 [[T1_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret i8 [[T2]]
>  ;
>    %t0 = sub i8 0, %z
> @@ -131,9 +131,9 @@ define i8 @t6(i8 %x, i1 %y, i8 %z) {
>  }
>  define i8 @t7(i8 %x, i1 %y, i8 %z) {
>  ; CHECK-LABEL: @t7(
> -; CHECK-NEXT:    [[T0:%.*]] = shl i8 1, [[Z:%.*]]
> -; CHECK-NEXT:    [[T1:%.*]] = select i1 [[Y:%.*]], i8 0, i8 [[T0]]
> -; CHECK-NEXT:    [[T2:%.*]] = sub i8 [[X:%.*]], [[T1]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = shl i8 -1, [[Z:%.*]]
> +; CHECK-NEXT:    [[T1_NEG:%.*]] = select i1 [[Y:%.*]], i8 0, i8 [[T0_NEG]]
> +; CHECK-NEXT:    [[T2:%.*]] = add i8 [[T1_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret i8 [[T2]]
>  ;
>    %t0 = shl i8 1, %z
> @@ -169,10 +169,10 @@ define i8 @t9(i8 %x, i8 %y) {
>  }
>  define i8 @n10(i8 %x, i8 %y, i8 %z) {
>  ; CHECK-LABEL: @n10(
> -; CHECK-NEXT:    [[T0:%.*]] = sub i8 [[Y:%.*]], [[X:%.*]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = sub i8 [[X:%.*]], [[Y:%.*]]
> +; CHECK-NEXT:    [[T0:%.*]] = sub i8 [[Y]], [[X]]
>  ; CHECK-NEXT:    call void @use8(i8 [[T0]])
> -; CHECK-NEXT:    [[T1:%.*]] = sub i8 0, [[T0]]
> -; CHECK-NEXT:    ret i8 [[T1]]
> +; CHECK-NEXT:    ret i8 [[T0_NEG]]
>  ;
>    %t0 = sub i8 %y, %x
>    call void @use8(i8 %t0)
> @@ -204,8 +204,8 @@ define i8 @n13(i8 %x, i8 %y, i8 %z) {
>  ; CHECK-LABEL: @n13(
>  ; CHECK-NEXT:    [[T0:%.*]] = sub i8 0, [[Y:%.*]]
>  ; CHECK-NEXT:    call void @use8(i8 [[T0]])
> -; CHECK-NEXT:    [[T11:%.*]] = sub i8 [[Y]], [[Z:%.*]]
> -; CHECK-NEXT:    [[T2:%.*]] = add i8 [[T11]], [[X:%.*]]
> +; CHECK-NEXT:    [[T1_NEG:%.*]] = sub i8 [[Y]], [[Z:%.*]]
> +; CHECK-NEXT:    [[T2:%.*]] = add i8 [[T1_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret i8 [[T2]]
>  ;
>    %t0 = sub i8 0, %y
> @@ -242,8 +242,8 @@ define i8 @t15(i8 %x, i8 %y, i8 %z) {
>  ; CHECK-LABEL: @t15(
>  ; CHECK-NEXT:    [[T0:%.*]] = sub i8 0, [[Y:%.*]]
>  ; CHECK-NEXT:    call void @use8(i8 [[T0]])
> -; CHECK-NEXT:    [[TMP1:%.*]] = mul i8 [[Z:%.*]], [[Y]]
> -; CHECK-NEXT:    [[T2:%.*]] = add i8 [[TMP1]], [[X:%.*]]
> +; CHECK-NEXT:    [[T1_NEG:%.*]] = mul i8 [[Y]], [[Z:%.*]]
> +; CHECK-NEXT:    [[T2:%.*]] = add i8 [[T1_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret i8 [[T2]]
>  ;
>    %t0 = sub i8 0, %y
> @@ -279,8 +279,8 @@ define i8 @t16(i1 %c, i8 %x) {
>  ; CHECK:       else:
>  ; CHECK-NEXT:    br label [[END]]
>  ; CHECK:       end:
> -; CHECK-NEXT:    [[Z:%.*]] = phi i8 [ [[X:%.*]], [[THEN]] ], [ 42,
> [[ELSE]] ]
> -; CHECK-NEXT:    ret i8 [[Z]]
> +; CHECK-NEXT:    [[Z_NEG:%.*]] = phi i8 [ [[X:%.*]], [[THEN]] ], [ 42,
> [[ELSE]] ]
> +; CHECK-NEXT:    ret i8 [[Z_NEG]]
>  ;
>  begin:
>    br i1 %c, label %then, label %else
> @@ -352,9 +352,9 @@ end:
>  ; truncation can be negated if it's operand can be negated
>  define i8 @t20(i8 %x, i16 %y) {
>  ; CHECK-LABEL: @t20(
> -; CHECK-NEXT:    [[T0:%.*]] = shl i16 -42, [[Y:%.*]]
> -; CHECK-NEXT:    [[T1:%.*]] = trunc i16 [[T0]] to i8
> -; CHECK-NEXT:    [[T2:%.*]] = sub i8 [[X:%.*]], [[T1]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = shl i16 42, [[Y:%.*]]
> +; CHECK-NEXT:    [[T1_NEG:%.*]] = trunc i16 [[T0_NEG]] to i8
> +; CHECK-NEXT:    [[T2:%.*]] = add i8 [[T1_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret i8 [[T2]]
>  ;
>    %t0 = shl i16 -42, %y
> @@ -427,9 +427,9 @@ define i4 @negate_shl_xor(i4 %x, i4 %y) {
>
>  define i8 @negate_shl_not_uses(i8 %x, i8 %y) {
>  ; CHECK-LABEL: @negate_shl_not_uses(
> -; CHECK-NEXT:    [[O:%.*]] = xor i8 [[X:%.*]], -1
> +; CHECK-NEXT:    [[O_NEG:%.*]] = add i8 [[X:%.*]], 1
> +; CHECK-NEXT:    [[O:%.*]] = xor i8 [[X]], -1
>  ; CHECK-NEXT:    call void @use8(i8 [[O]])
> -; CHECK-NEXT:    [[O_NEG:%.*]] = add i8 [[X]], 1
>  ; CHECK-NEXT:    [[S_NEG:%.*]] = shl i8 [[O_NEG]], [[Y:%.*]]
>  ; CHECK-NEXT:    ret i8 [[S_NEG]]
>  ;
> @@ -442,9 +442,9 @@ define i8 @negate_shl_not_uses(i8 %x, i8 %y) {
>
>  define <2 x i4> @negate_mul_not_uses_vec(<2 x i4> %x, <2 x i4> %y) {
>  ; CHECK-LABEL: @negate_mul_not_uses_vec(
> -; CHECK-NEXT:    [[O:%.*]] = xor <2 x i4> [[X:%.*]], <i4 -1, i4 -1>
> +; CHECK-NEXT:    [[O_NEG:%.*]] = add <2 x i4> [[X:%.*]], <i4 1, i4 1>
> +; CHECK-NEXT:    [[O:%.*]] = xor <2 x i4> [[X]], <i4 -1, i4 -1>
>  ; CHECK-NEXT:    call void @use_v2i4(<2 x i4> [[O]])
> -; CHECK-NEXT:    [[O_NEG:%.*]] = add <2 x i4> [[X]], <i4 1, i4 1>
>  ; CHECK-NEXT:    [[S_NEG:%.*]] = mul <2 x i4> [[O_NEG]], [[Y:%.*]]
>  ; CHECK-NEXT:    ret <2 x i4> [[S_NEG]]
>  ;
> @@ -458,8 +458,8 @@ define <2 x i4> @negate_mul_not_uses_vec(<2 x i4> %x,
> <2 x i4> %y) {
>  ; signed division can be negated if divisor can be negated and is not 1/-1
>  define i8 @negate_sdiv(i8 %x, i8 %y) {
>  ; CHECK-LABEL: @negate_sdiv(
> -; CHECK-NEXT:    [[T0:%.*]] = sdiv i8 [[Y:%.*]], 42
> -; CHECK-NEXT:    [[T1:%.*]] = sub i8 [[X:%.*]], [[T0]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = sdiv i8 [[Y:%.*]], -42
> +; CHECK-NEXT:    [[T1:%.*]] = add i8 [[T0_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret i8 [[T1]]
>  ;
>    %t0 = sdiv i8 %y, 42
> @@ -478,12 +478,24 @@ define i8 @negate_sdiv_extrause(i8 %x, i8 %y) {
>    %t1 = sub i8 %x, %t0
>    ret i8 %t1
>  }
> +define i8 @negate_sdiv_extrause2(i8 %x, i8 %y) {
> +; CHECK-LABEL: @negate_sdiv_extrause2(
> +; CHECK-NEXT:    [[T0:%.*]] = sdiv i8 [[Y:%.*]], 42
> +; CHECK-NEXT:    call void @use8(i8 [[T0]])
> +; CHECK-NEXT:    [[T1:%.*]] = sub nsw i8 0, [[T0]]
> +; CHECK-NEXT:    ret i8 [[T1]]
> +;
> +  %t0 = sdiv i8 %y, 42
> +  call void @use8(i8 %t0)
> +  %t1 = sub i8 0, %t0
> +  ret i8 %t1
> +}
>
>  ; Right-shift sign bit smear is negatible.
>  define i8 @negate_ashr(i8 %x, i8 %y) {
>  ; CHECK-LABEL: @negate_ashr(
> -; CHECK-NEXT:    [[T0:%.*]] = ashr i8 [[Y:%.*]], 7
> -; CHECK-NEXT:    [[T1:%.*]] = sub i8 [[X:%.*]], [[T0]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = lshr i8 [[Y:%.*]], 7
> +; CHECK-NEXT:    [[T1:%.*]] = add i8 [[T0_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret i8 [[T1]]
>  ;
>    %t0 = ashr i8 %y, 7
> @@ -492,8 +504,8 @@ define i8 @negate_ashr(i8 %x, i8 %y) {
>  }
>  define i8 @negate_lshr(i8 %x, i8 %y) {
>  ; CHECK-LABEL: @negate_lshr(
> -; CHECK-NEXT:    [[T0:%.*]] = lshr i8 [[Y:%.*]], 7
> -; CHECK-NEXT:    [[T1:%.*]] = sub i8 [[X:%.*]], [[T0]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = ashr i8 [[Y:%.*]], 7
> +; CHECK-NEXT:    [[T1:%.*]] = add i8 [[T0_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret i8 [[T1]]
>  ;
>    %t0 = lshr i8 %y, 7
> @@ -548,8 +560,8 @@ define i8 @negate_lshr_wrongshift(i8 %x, i8 %y) {
>  ; *ext of i1 is always negatible
>  define i8 @negate_sext(i8 %x, i1 %y) {
>  ; CHECK-LABEL: @negate_sext(
> -; CHECK-NEXT:    [[TMP1:%.*]] = zext i1 [[Y:%.*]] to i8
> -; CHECK-NEXT:    [[T1:%.*]] = add i8 [[TMP1]], [[X:%.*]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = zext i1 [[Y:%.*]] to i8
> +; CHECK-NEXT:    [[T1:%.*]] = add i8 [[T0_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret i8 [[T1]]
>  ;
>    %t0 = sext i1 %y to i8
> @@ -558,8 +570,8 @@ define i8 @negate_sext(i8 %x, i1 %y) {
>  }
>  define i8 @negate_zext(i8 %x, i1 %y) {
>  ; CHECK-LABEL: @negate_zext(
> -; CHECK-NEXT:    [[TMP1:%.*]] = sext i1 [[Y:%.*]] to i8
> -; CHECK-NEXT:    [[T1:%.*]] = add i8 [[TMP1]], [[X:%.*]]
> +; CHECK-NEXT:    [[T0_NEG:%.*]] = sext i1 [[Y:%.*]] to i8
> +; CHECK-NEXT:    [[T1:%.*]] = add i8 [[T0_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret i8 [[T1]]
>  ;
>    %t0 = zext i1 %y to i8
>
> diff  --git a/llvm/test/Transforms/InstCombine/sub.ll
> b/llvm/test/Transforms/InstCombine/sub.ll
> index 6bda5b0a3019..1b4149291951 100644
> --- a/llvm/test/Transforms/InstCombine/sub.ll
> +++ b/llvm/test/Transforms/InstCombine/sub.ll
> @@ -218,8 +218,8 @@ define <2 x i8> @notnotsub_vec_undef_elts(<2 x i8> %x,
> <2 x i8> %y) {
>
>  define i32 @test5(i32 %A, i32 %B, i32 %C) {
>  ; CHECK-LABEL: @test5(
> -; CHECK-NEXT:    [[D1:%.*]] = sub i32 [[C:%.*]], [[B:%.*]]
> -; CHECK-NEXT:    [[E:%.*]] = add i32 [[D1]], [[A:%.*]]
> +; CHECK-NEXT:    [[D_NEG:%.*]] = sub i32 [[C:%.*]], [[B:%.*]]
> +; CHECK-NEXT:    [[E:%.*]] = add i32 [[D_NEG]], [[A:%.*]]
>  ; CHECK-NEXT:    ret i32 [[E]]
>  ;
>    %D = sub i32 %B, %C
> @@ -555,11 +555,11 @@ define i64 @test_neg_shl_sub(i64 %a, i64 %b) {
>
>  define i64 @test_neg_shl_sub_extra_use1(i64 %a, i64 %b, i64* %p) {
>  ; CHECK-LABEL: @test_neg_shl_sub_extra_use1(
> -; CHECK-NEXT:    [[SUB:%.*]] = sub i64 [[A:%.*]], [[B:%.*]]
> +; CHECK-NEXT:    [[SUB_NEG:%.*]] = sub i64 [[B:%.*]], [[A:%.*]]
> +; CHECK-NEXT:    [[SUB:%.*]] = sub i64 [[A]], [[B]]
>  ; CHECK-NEXT:    store i64 [[SUB]], i64* [[P:%.*]], align 8
> -; CHECK-NEXT:    [[MUL:%.*]] = shl i64 [[SUB]], 2
> -; CHECK-NEXT:    [[NEG:%.*]] = sub i64 0, [[MUL]]
> -; CHECK-NEXT:    ret i64 [[NEG]]
> +; CHECK-NEXT:    [[MUL_NEG:%.*]] = shl i64 [[SUB_NEG]], 2
> +; CHECK-NEXT:    ret i64 [[MUL_NEG]]
>  ;
>    %sub = sub i64 %a, %b
>    store i64 %sub, i64* %p
> @@ -621,8 +621,8 @@ define i64 @test_neg_shl_sext_i1(i1 %a, i64 %b) {
>
>  define i64 @test_neg_zext_i1_extra_use(i1 %a, i64 %b, i64* %p) {
>  ; CHECK-LABEL: @test_neg_zext_i1_extra_use(
> -; CHECK-NEXT:    [[EXT:%.*]] = zext i1 [[A:%.*]] to i64
> -; CHECK-NEXT:    [[EXT_NEG:%.*]] = sext i1 [[A]] to i64
> +; CHECK-NEXT:    [[EXT_NEG:%.*]] = sext i1 [[A:%.*]] to i64
> +; CHECK-NEXT:    [[EXT:%.*]] = zext i1 [[A]] to i64
>  ; CHECK-NEXT:    store i64 [[EXT]], i64* [[P:%.*]], align 8
>  ; CHECK-NEXT:    ret i64 [[EXT_NEG]]
>  ;
> @@ -634,8 +634,8 @@ define i64 @test_neg_zext_i1_extra_use(i1 %a, i64 %b,
> i64* %p) {
>
>  define i64 @test_neg_sext_i1_extra_use(i1 %a, i64 %b, i64* %p) {
>  ; CHECK-LABEL: @test_neg_sext_i1_extra_use(
> -; CHECK-NEXT:    [[EXT:%.*]] = sext i1 [[A:%.*]] to i64
> -; CHECK-NEXT:    [[EXT_NEG:%.*]] = zext i1 [[A]] to i64
> +; CHECK-NEXT:    [[EXT_NEG:%.*]] = zext i1 [[A:%.*]] to i64
> +; CHECK-NEXT:    [[EXT:%.*]] = sext i1 [[A]] to i64
>  ; CHECK-NEXT:    store i64 [[EXT]], i64* [[P:%.*]], align 8
>  ; CHECK-NEXT:    ret i64 [[EXT_NEG]]
>  ;
> @@ -703,7 +703,7 @@ define i64 @test_neg_mul_sub_commuted(i64 %a, i64 %b,
> i64 %c) {
>  ; CHECK-LABEL: @test_neg_mul_sub_commuted(
>  ; CHECK-NEXT:    [[COMPLEX:%.*]] = mul i64 [[C:%.*]], [[C]]
>  ; CHECK-NEXT:    [[SUB_NEG:%.*]] = sub i64 [[B:%.*]], [[A:%.*]]
> -; CHECK-NEXT:    [[MUL_NEG:%.*]] = mul i64 [[COMPLEX]], [[SUB_NEG]]
> +; CHECK-NEXT:    [[MUL_NEG:%.*]] = mul i64 [[SUB_NEG]], [[COMPLEX]]
>  ; CHECK-NEXT:    ret i64 [[MUL_NEG]]
>  ;
>    %complex = mul i64 %c, %c
> @@ -715,8 +715,8 @@ define i64 @test_neg_mul_sub_commuted(i64 %a, i64 %b,
> i64 %c) {
>
>  define i32 @test27(i32 %x, i32 %y) {
>  ; CHECK-LABEL: @test27(
> -; CHECK-NEXT:    [[TMP1:%.*]] = shl i32 [[Y:%.*]], 3
> -; CHECK-NEXT:    [[SUB:%.*]] = add i32 [[TMP1]], [[X:%.*]]
> +; CHECK-NEXT:    [[MUL_NEG:%.*]] = shl i32 [[Y:%.*]], 3
> +; CHECK-NEXT:    [[SUB:%.*]] = add i32 [[MUL_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret i32 [[SUB]]
>  ;
>    %mul = mul i32 %y, -8
> @@ -726,8 +726,8 @@ define i32 @test27(i32 %x, i32 %y) {
>
>  define <2 x i32> @test27vec(<2 x i32> %x, <2 x i32> %y) {
>  ; CHECK-LABEL: @test27vec(
> -; CHECK-NEXT:    [[TMP1:%.*]] = mul <2 x i32> [[Y:%.*]], <i32 8, i32 6>
> -; CHECK-NEXT:    [[SUB:%.*]] = add <2 x i32> [[TMP1]], [[X:%.*]]
> +; CHECK-NEXT:    [[MUL_NEG:%.*]] = mul <2 x i32> [[Y:%.*]], <i32 8, i32 6>
> +; CHECK-NEXT:    [[SUB:%.*]] = add <2 x i32> [[MUL_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret <2 x i32> [[SUB]]
>  ;
>    %mul = mul <2 x i32> %y, <i32 -8, i32 -6>
> @@ -737,8 +737,8 @@ define <2 x i32> @test27vec(<2 x i32> %x, <2 x i32>
> %y) {
>
>  define <2 x i32> @test27vecsplat(<2 x i32> %x, <2 x i32> %y) {
>  ; CHECK-LABEL: @test27vecsplat(
> -; CHECK-NEXT:    [[TMP1:%.*]] = shl <2 x i32> [[Y:%.*]], <i32 3, i32 3>
> -; CHECK-NEXT:    [[SUB:%.*]] = add <2 x i32> [[TMP1]], [[X:%.*]]
> +; CHECK-NEXT:    [[MUL_NEG:%.*]] = shl <2 x i32> [[Y:%.*]], <i32 3, i32 3>
> +; CHECK-NEXT:    [[SUB:%.*]] = add <2 x i32> [[MUL_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret <2 x i32> [[SUB]]
>  ;
>    %mul = mul <2 x i32> %y, <i32 -8, i32 -8>
> @@ -748,8 +748,8 @@ define <2 x i32> @test27vecsplat(<2 x i32> %x, <2 x
> i32> %y) {
>
>  define <2 x i32> @test27vecmixed(<2 x i32> %x, <2 x i32> %y) {
>  ; CHECK-LABEL: @test27vecmixed(
> -; CHECK-NEXT:    [[TMP1:%.*]] = mul <2 x i32> [[Y:%.*]], <i32 8, i32 -8>
> -; CHECK-NEXT:    [[SUB:%.*]] = add <2 x i32> [[TMP1]], [[X:%.*]]
> +; CHECK-NEXT:    [[MUL_NEG:%.*]] = mul <2 x i32> [[Y:%.*]], <i32 8, i32
> -8>
> +; CHECK-NEXT:    [[SUB:%.*]] = add <2 x i32> [[MUL_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret <2 x i32> [[SUB]]
>  ;
>    %mul = mul <2 x i32> %y, <i32 -8, i32 8>
> @@ -759,8 +759,8 @@ define <2 x i32> @test27vecmixed(<2 x i32> %x, <2 x
> i32> %y) {
>
>  define i32 @test27commuted(i32 %x, i32 %y) {
>  ; CHECK-LABEL: @test27commuted(
> -; CHECK-NEXT:    [[TMP1:%.*]] = shl i32 [[Y:%.*]], 3
> -; CHECK-NEXT:    [[SUB:%.*]] = add i32 [[TMP1]], [[X:%.*]]
> +; CHECK-NEXT:    [[MUL_NEG:%.*]] = shl i32 [[Y:%.*]], 3
> +; CHECK-NEXT:    [[SUB:%.*]] = add i32 [[MUL_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret i32 [[SUB]]
>  ;
>    %mul = mul i32 -8, %y
> @@ -770,8 +770,8 @@ define i32 @test27commuted(i32 %x, i32 %y) {
>
>  define <2 x i32> @test27commutedvec(<2 x i32> %x, <2 x i32> %y) {
>  ; CHECK-LABEL: @test27commutedvec(
> -; CHECK-NEXT:    [[TMP1:%.*]] = mul <2 x i32> [[Y:%.*]], <i32 8, i32 6>
> -; CHECK-NEXT:    [[SUB:%.*]] = add <2 x i32> [[TMP1]], [[X:%.*]]
> +; CHECK-NEXT:    [[MUL_NEG:%.*]] = mul <2 x i32> [[Y:%.*]], <i32 8, i32 6>
> +; CHECK-NEXT:    [[SUB:%.*]] = add <2 x i32> [[MUL_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret <2 x i32> [[SUB]]
>  ;
>    %mul = mul <2 x i32> <i32 -8, i32 -6>, %y
> @@ -781,8 +781,8 @@ define <2 x i32> @test27commutedvec(<2 x i32> %x, <2 x
> i32> %y) {
>
>  define <2 x i32> @test27commutedvecsplat(<2 x i32> %x, <2 x i32> %y) {
>  ; CHECK-LABEL: @test27commutedvecsplat(
> -; CHECK-NEXT:    [[TMP1:%.*]] = shl <2 x i32> [[Y:%.*]], <i32 3, i32 3>
> -; CHECK-NEXT:    [[SUB:%.*]] = add <2 x i32> [[TMP1]], [[X:%.*]]
> +; CHECK-NEXT:    [[MUL_NEG:%.*]] = shl <2 x i32> [[Y:%.*]], <i32 3, i32 3>
> +; CHECK-NEXT:    [[SUB:%.*]] = add <2 x i32> [[MUL_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret <2 x i32> [[SUB]]
>  ;
>    %mul = mul <2 x i32> <i32 -8, i32 -8>, %y
> @@ -792,8 +792,8 @@ define <2 x i32> @test27commutedvecsplat(<2 x i32> %x,
> <2 x i32> %y) {
>
>  define <2 x i32> @test27commutedvecmixed(<2 x i32> %x, <2 x i32> %y) {
>  ; CHECK-LABEL: @test27commutedvecmixed(
> -; CHECK-NEXT:    [[TMP1:%.*]] = mul <2 x i32> [[Y:%.*]], <i32 8, i32 -8>
> -; CHECK-NEXT:    [[SUB:%.*]] = add <2 x i32> [[TMP1]], [[X:%.*]]
> +; CHECK-NEXT:    [[MUL_NEG:%.*]] = mul <2 x i32> [[Y:%.*]], <i32 8, i32
> -8>
> +; CHECK-NEXT:    [[SUB:%.*]] = add <2 x i32> [[MUL_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret <2 x i32> [[SUB]]
>  ;
>    %mul = mul <2 x i32> <i32 -8, i32 8>, %y
> @@ -1105,7 +1105,7 @@ define i32 @test52(i32 %X) {
>
>  define <2 x i1> @test53(<2 x i1> %A, <2 x i1> %B) {
>  ; CHECK-LABEL: @test53(
> -; CHECK-NEXT:    [[SUB:%.*]] = xor <2 x i1> [[A:%.*]], [[B:%.*]]
> +; CHECK-NEXT:    [[SUB:%.*]] = xor <2 x i1> [[B:%.*]], [[A:%.*]]
>  ; CHECK-NEXT:    ret <2 x i1> [[SUB]]
>  ;
>    %sub = sub <2 x i1> %A, %B
> @@ -1149,8 +1149,8 @@ define i32 @test55(i1 %which) {
>  ; CHECK:       delay:
>  ; CHECK-NEXT:    br label [[FINAL]]
>  ; CHECK:       final:
> -; CHECK-NEXT:    [[A:%.*]] = phi i32 [ -877, [[ENTRY:%.*]] ], [ 113,
> [[DELAY]] ]
> -; CHECK-NEXT:    ret i32 [[A]]
> +; CHECK-NEXT:    [[A_NEG:%.*]] = phi i32 [ -877, [[ENTRY:%.*]] ], [ 113,
> [[DELAY]] ]
> +; CHECK-NEXT:    ret i32 [[A_NEG]]
>  ;
>  entry:
>    br i1 %which, label %final, label %delay
> @@ -1171,8 +1171,8 @@ define <2 x i32> @test55vec(i1 %which) {
>  ; CHECK:       delay:
>  ; CHECK-NEXT:    br label [[FINAL]]
>  ; CHECK:       final:
> -; CHECK-NEXT:    [[A:%.*]] = phi <2 x i32> [ <i32 -877, i32 -877>,
> [[ENTRY:%.*]] ], [ <i32 113, i32 113>, [[DELAY]] ]
> -; CHECK-NEXT:    ret <2 x i32> [[A]]
> +; CHECK-NEXT:    [[A_NEG:%.*]] = phi <2 x i32> [ <i32 -877, i32 -877>,
> [[ENTRY:%.*]] ], [ <i32 113, i32 113>, [[DELAY]] ]
> +; CHECK-NEXT:    ret <2 x i32> [[A_NEG]]
>  ;
>  entry:
>    br i1 %which, label %final, label %delay
> @@ -1193,8 +1193,8 @@ define <2 x i32> @test55vec2(i1 %which) {
>  ; CHECK:       delay:
>  ; CHECK-NEXT:    br label [[FINAL]]
>  ; CHECK:       final:
> -; CHECK-NEXT:    [[A:%.*]] = phi <2 x i32> [ <i32 -877, i32 -2167>,
> [[ENTRY:%.*]] ], [ <i32 113, i32 303>, [[DELAY]] ]
> -; CHECK-NEXT:    ret <2 x i32> [[A]]
> +; CHECK-NEXT:    [[A_NEG:%.*]] = phi <2 x i32> [ <i32 -877, i32 -2167>,
> [[ENTRY:%.*]] ], [ <i32 113, i32 303>, [[DELAY]] ]
> +; CHECK-NEXT:    ret <2 x i32> [[A_NEG]]
>  ;
>  entry:
>    br i1 %which, label %final, label %delay
> @@ -1356,8 +1356,8 @@ define i32 @test64(i32 %x) {
>  ; CHECK-LABEL: @test64(
>  ; CHECK-NEXT:    [[TMP1:%.*]] = icmp slt i32 [[X:%.*]], 255
>  ; CHECK-NEXT:    [[TMP2:%.*]] = select i1 [[TMP1]], i32 [[X]], i32 255
> -; CHECK-NEXT:    [[RES:%.*]] = add nsw i32 [[TMP2]], 1
> -; CHECK-NEXT:    ret i32 [[RES]]
> +; CHECK-NEXT:    [[DOTNEG:%.*]] = add nsw i32 [[TMP2]], 1
> +; CHECK-NEXT:    ret i32 [[DOTNEG]]
>  ;
>    %1 = xor i32 %x, -1
>    %2 = icmp sgt i32 %1, -256
> @@ -1370,8 +1370,8 @@ define i32 @test65(i32 %x) {
>  ; CHECK-LABEL: @test65(
>  ; CHECK-NEXT:    [[TMP1:%.*]] = icmp sgt i32 [[X:%.*]], -256
>  ; CHECK-NEXT:    [[TMP2:%.*]] = select i1 [[TMP1]], i32 [[X]], i32 -256
> -; CHECK-NEXT:    [[RES:%.*]] = add i32 [[TMP2]], 1
> -; CHECK-NEXT:    ret i32 [[RES]]
> +; CHECK-NEXT:    [[DOTNEG:%.*]] = add i32 [[TMP2]], 1
> +; CHECK-NEXT:    ret i32 [[DOTNEG]]
>  ;
>    %1 = xor i32 %x, -1
>    %2 = icmp slt i32 %1, 255
> @@ -1384,8 +1384,8 @@ define i32 @test66(i32 %x) {
>  ; CHECK-LABEL: @test66(
>  ; CHECK-NEXT:    [[TMP1:%.*]] = icmp ult i32 [[X:%.*]], -101
>  ; CHECK-NEXT:    [[TMP2:%.*]] = select i1 [[TMP1]], i32 [[X]], i32 -101
> -; CHECK-NEXT:    [[RES:%.*]] = add nuw i32 [[TMP2]], 1
> -; CHECK-NEXT:    ret i32 [[RES]]
> +; CHECK-NEXT:    [[DOTNEG:%.*]] = add nuw i32 [[TMP2]], 1
> +; CHECK-NEXT:    ret i32 [[DOTNEG]]
>  ;
>    %1 = xor i32 %x, -1
>    %2 = icmp ugt i32 %1, 100
> @@ -1398,8 +1398,8 @@ define i32 @test67(i32 %x) {
>  ; CHECK-LABEL: @test67(
>  ; CHECK-NEXT:    [[TMP1:%.*]] = icmp ugt i32 [[X:%.*]], 100
>  ; CHECK-NEXT:    [[TMP2:%.*]] = select i1 [[TMP1]], i32 [[X]], i32 100
> -; CHECK-NEXT:    [[RES:%.*]] = add i32 [[TMP2]], 1
> -; CHECK-NEXT:    ret i32 [[RES]]
> +; CHECK-NEXT:    [[DOTNEG:%.*]] = add i32 [[TMP2]], 1
> +; CHECK-NEXT:    ret i32 [[DOTNEG]]
>  ;
>    %1 = xor i32 %x, -1
>    %2 = icmp ult i32 %1, -101
> @@ -1413,8 +1413,8 @@ define <2 x i32> @test68(<2 x i32> %x) {
>  ; CHECK-LABEL: @test68(
>  ; CHECK-NEXT:    [[TMP1:%.*]] = icmp slt <2 x i32> [[X:%.*]], <i32 255,
> i32 255>
>  ; CHECK-NEXT:    [[TMP2:%.*]] = select <2 x i1> [[TMP1]], <2 x i32>
> [[X]], <2 x i32> <i32 255, i32 255>
> -; CHECK-NEXT:    [[RES:%.*]] = add nsw <2 x i32> [[TMP2]], <i32 1, i32 1>
> -; CHECK-NEXT:    ret <2 x i32> [[RES]]
> +; CHECK-NEXT:    [[DOTNEG:%.*]] = add nsw <2 x i32> [[TMP2]], <i32 1, i32
> 1>
> +; CHECK-NEXT:    ret <2 x i32> [[DOTNEG]]
>  ;
>    %1 = xor <2 x i32> %x, <i32 -1, i32 -1>
>    %2 = icmp sgt <2 x i32> %1, <i32 -256, i32 -256>
> @@ -1428,8 +1428,8 @@ define <2 x i32> @test69(<2 x i32> %x) {
>  ; CHECK-LABEL: @test69(
>  ; CHECK-NEXT:    [[TMP1:%.*]] = icmp slt <2 x i32> [[X:%.*]], <i32 255,
> i32 127>
>  ; CHECK-NEXT:    [[TMP2:%.*]] = select <2 x i1> [[TMP1]], <2 x i32>
> [[X]], <2 x i32> <i32 255, i32 127>
> -; CHECK-NEXT:    [[RES:%.*]] = add <2 x i32> [[TMP2]], <i32 1, i32 1>
> -; CHECK-NEXT:    ret <2 x i32> [[RES]]
> +; CHECK-NEXT:    [[DOTNEG:%.*]] = add <2 x i32> [[TMP2]], <i32 1, i32 1>
> +; CHECK-NEXT:    ret <2 x i32> [[DOTNEG]]
>  ;
>    %1 = xor <2 x i32> %x, <i32 -1, i32 -1>
>    %2 = icmp sgt <2 x i32> %1, <i32 -256, i32 -128>
>
> diff  --git a/llvm/test/Transforms/InstCombine/zext-bool-add-sub.ll
> b/llvm/test/Transforms/InstCombine/zext-bool-add-sub.ll
> index 71fa9a795735..7ad0655dcbe8 100644
> --- a/llvm/test/Transforms/InstCombine/zext-bool-add-sub.ll
> +++ b/llvm/test/Transforms/InstCombine/zext-bool-add-sub.ll
> @@ -5,9 +5,9 @@
>
>  define i32 @a(i1 zeroext %x, i1 zeroext %y) {
>  ; CHECK-LABEL: @a(
> -; CHECK-NEXT:    [[CONV3_NEG:%.*]] = sext i1 [[Y:%.*]] to i32
> +; CHECK-NEXT:    [[CONV3_NEG1:%.*]] = sext i1 [[Y:%.*]] to i32
>  ; CHECK-NEXT:    [[SUB:%.*]] = select i1 [[X:%.*]], i32 2, i32 1
> -; CHECK-NEXT:    [[ADD:%.*]] = add nsw i32 [[SUB]], [[CONV3_NEG]]
> +; CHECK-NEXT:    [[ADD:%.*]] = add nsw i32 [[SUB]], [[CONV3_NEG1]]
>  ; CHECK-NEXT:    ret i32 [[ADD]]
>  ;
>    %conv = zext i1 %x to i32
> @@ -95,8 +95,8 @@ declare void @use(i64)
>
>  define i64 @zext_negate(i1 %A) {
>  ; CHECK-LABEL: @zext_negate(
> -; CHECK-NEXT:    [[SUB:%.*]] = sext i1 [[A:%.*]] to i64
> -; CHECK-NEXT:    ret i64 [[SUB]]
> +; CHECK-NEXT:    [[EXT_NEG:%.*]] = sext i1 [[A:%.*]] to i64
> +; CHECK-NEXT:    ret i64 [[EXT_NEG]]
>  ;
>    %ext = zext i1 %A to i64
>    %sub = sub i64 0, %ext
> @@ -105,10 +105,10 @@ define i64 @zext_negate(i1 %A) {
>
>  define i64 @zext_negate_extra_use(i1 %A) {
>  ; CHECK-LABEL: @zext_negate_extra_use(
> -; CHECK-NEXT:    [[EXT:%.*]] = zext i1 [[A:%.*]] to i64
> -; CHECK-NEXT:    [[SUB:%.*]] = sext i1 [[A]] to i64
> +; CHECK-NEXT:    [[EXT_NEG:%.*]] = sext i1 [[A:%.*]] to i64
> +; CHECK-NEXT:    [[EXT:%.*]] = zext i1 [[A]] to i64
>  ; CHECK-NEXT:    call void @use(i64 [[EXT]])
> -; CHECK-NEXT:    ret i64 [[SUB]]
> +; CHECK-NEXT:    ret i64 [[EXT_NEG]]
>  ;
>    %ext = zext i1 %A to i64
>    %sub = sub i64 0, %ext
> @@ -118,8 +118,8 @@ define i64 @zext_negate_extra_use(i1 %A) {
>
>  define <2 x i64> @zext_negate_vec(<2 x i1> %A) {
>  ; CHECK-LABEL: @zext_negate_vec(
> -; CHECK-NEXT:    [[SUB:%.*]] = sext <2 x i1> [[A:%.*]] to <2 x i64>
> -; CHECK-NEXT:    ret <2 x i64> [[SUB]]
> +; CHECK-NEXT:    [[EXT_NEG:%.*]] = sext <2 x i1> [[A:%.*]] to <2 x i64>
> +; CHECK-NEXT:    ret <2 x i64> [[EXT_NEG]]
>  ;
>    %ext = zext <2 x i1> %A to <2 x i64>
>    %sub = sub <2 x i64> zeroinitializer, %ext
> @@ -128,8 +128,8 @@ define <2 x i64> @zext_negate_vec(<2 x i1> %A) {
>
>  define <2 x i64> @zext_negate_vec_undef_elt(<2 x i1> %A) {
>  ; CHECK-LABEL: @zext_negate_vec_undef_elt(
> -; CHECK-NEXT:    [[SUB:%.*]] = sext <2 x i1> [[A:%.*]] to <2 x i64>
> -; CHECK-NEXT:    ret <2 x i64> [[SUB]]
> +; CHECK-NEXT:    [[EXT_NEG:%.*]] = sext <2 x i1> [[A:%.*]] to <2 x i64>
> +; CHECK-NEXT:    ret <2 x i64> [[EXT_NEG]]
>  ;
>    %ext = zext <2 x i1> %A to <2 x i64>
>    %sub = sub <2 x i64> <i64 0, i64 undef>, %ext
> @@ -181,8 +181,8 @@ define <2 x i64> @zext_sub_const_vec_undef_elt(<2 x
> i1> %A) {
>
>  define i64 @sext_negate(i1 %A) {
>  ; CHECK-LABEL: @sext_negate(
> -; CHECK-NEXT:    [[SUB:%.*]] = zext i1 [[A:%.*]] to i64
> -; CHECK-NEXT:    ret i64 [[SUB]]
> +; CHECK-NEXT:    [[EXT_NEG:%.*]] = zext i1 [[A:%.*]] to i64
> +; CHECK-NEXT:    ret i64 [[EXT_NEG]]
>  ;
>    %ext = sext i1 %A to i64
>    %sub = sub i64 0, %ext
> @@ -191,10 +191,10 @@ define i64 @sext_negate(i1 %A) {
>
>  define i64 @sext_negate_extra_use(i1 %A) {
>  ; CHECK-LABEL: @sext_negate_extra_use(
> -; CHECK-NEXT:    [[EXT:%.*]] = sext i1 [[A:%.*]] to i64
> -; CHECK-NEXT:    [[SUB:%.*]] = zext i1 [[A]] to i64
> +; CHECK-NEXT:    [[EXT_NEG:%.*]] = zext i1 [[A:%.*]] to i64
> +; CHECK-NEXT:    [[EXT:%.*]] = sext i1 [[A]] to i64
>  ; CHECK-NEXT:    call void @use(i64 [[EXT]])
> -; CHECK-NEXT:    ret i64 [[SUB]]
> +; CHECK-NEXT:    ret i64 [[EXT_NEG]]
>  ;
>    %ext = sext i1 %A to i64
>    %sub = sub i64 0, %ext
> @@ -204,8 +204,8 @@ define i64 @sext_negate_extra_use(i1 %A) {
>
>  define <2 x i64> @sext_negate_vec(<2 x i1> %A) {
>  ; CHECK-LABEL: @sext_negate_vec(
> -; CHECK-NEXT:    [[SUB:%.*]] = zext <2 x i1> [[A:%.*]] to <2 x i64>
> -; CHECK-NEXT:    ret <2 x i64> [[SUB]]
> +; CHECK-NEXT:    [[EXT_NEG:%.*]] = zext <2 x i1> [[A:%.*]] to <2 x i64>
> +; CHECK-NEXT:    ret <2 x i64> [[EXT_NEG]]
>  ;
>    %ext = sext <2 x i1> %A to <2 x i64>
>    %sub = sub <2 x i64> zeroinitializer, %ext
> @@ -214,8 +214,8 @@ define <2 x i64> @sext_negate_vec(<2 x i1> %A) {
>
>  define <2 x i64> @sext_negate_vec_undef_elt(<2 x i1> %A) {
>  ; CHECK-LABEL: @sext_negate_vec_undef_elt(
> -; CHECK-NEXT:    [[SUB:%.*]] = zext <2 x i1> [[A:%.*]] to <2 x i64>
> -; CHECK-NEXT:    ret <2 x i64> [[SUB]]
> +; CHECK-NEXT:    [[EXT_NEG:%.*]] = zext <2 x i1> [[A:%.*]] to <2 x i64>
> +; CHECK-NEXT:    ret <2 x i64> [[EXT_NEG]]
>  ;
>    %ext = sext <2 x i1> %A to <2 x i64>
>    %sub = sub <2 x i64> <i64 0, i64 undef>, %ext
> @@ -267,8 +267,8 @@ define <2 x i64> @sext_sub_const_vec_undef_elt(<2 x
> i1> %A) {
>
>  define i8 @sext_sub(i8 %x, i1 %y) {
>  ; CHECK-LABEL: @sext_sub(
> -; CHECK-NEXT:    [[TMP1:%.*]] = zext i1 [[Y:%.*]] to i8
> -; CHECK-NEXT:    [[SUB:%.*]] = add i8 [[TMP1]], [[X:%.*]]
> +; CHECK-NEXT:    [[SEXT_NEG:%.*]] = zext i1 [[Y:%.*]] to i8
> +; CHECK-NEXT:    [[SUB:%.*]] = add i8 [[SEXT_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret i8 [[SUB]]
>  ;
>    %sext = sext i1 %y to i8
> @@ -280,8 +280,8 @@ define i8 @sext_sub(i8 %x, i1 %y) {
>
>  define <2 x i8> @sext_sub_vec(<2 x i8> %x, <2 x i1> %y) {
>  ; CHECK-LABEL: @sext_sub_vec(
> -; CHECK-NEXT:    [[TMP1:%.*]] = zext <2 x i1> [[Y:%.*]] to <2 x i8>
> -; CHECK-NEXT:    [[SUB:%.*]] = add <2 x i8> [[TMP1]], [[X:%.*]]
> +; CHECK-NEXT:    [[SEXT_NEG:%.*]] = zext <2 x i1> [[Y:%.*]] to <2 x i8>
> +; CHECK-NEXT:    [[SUB:%.*]] = add <2 x i8> [[SEXT_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret <2 x i8> [[SUB]]
>  ;
>    %sext = sext <2 x i1> %y to <2 x i8>
> @@ -293,8 +293,8 @@ define <2 x i8> @sext_sub_vec(<2 x i8> %x, <2 x i1>
> %y) {
>
>  define <2 x i8> @sext_sub_vec_nsw(<2 x i8> %x, <2 x i1> %y) {
>  ; CHECK-LABEL: @sext_sub_vec_nsw(
> -; CHECK-NEXT:    [[TMP1:%.*]] = zext <2 x i1> [[Y:%.*]] to <2 x i8>
> -; CHECK-NEXT:    [[SUB:%.*]] = add nsw <2 x i8> [[TMP1]], [[X:%.*]]
> +; CHECK-NEXT:    [[SEXT_NEG:%.*]] = zext <2 x i1> [[Y:%.*]] to <2 x i8>
> +; CHECK-NEXT:    [[SUB:%.*]] = add <2 x i8> [[SEXT_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret <2 x i8> [[SUB]]
>  ;
>    %sext = sext <2 x i1> %y to <2 x i8>
> @@ -306,8 +306,8 @@ define <2 x i8> @sext_sub_vec_nsw(<2 x i8> %x, <2 x
> i1> %y) {
>
>  define i8 @sext_sub_nuw(i8 %x, i1 %y) {
>  ; CHECK-LABEL: @sext_sub_nuw(
> -; CHECK-NEXT:    [[TMP1:%.*]] = zext i1 [[Y:%.*]] to i8
> -; CHECK-NEXT:    [[SUB:%.*]] = add i8 [[TMP1]], [[X:%.*]]
> +; CHECK-NEXT:    [[SEXT_NEG:%.*]] = zext i1 [[Y:%.*]] to i8
> +; CHECK-NEXT:    [[SUB:%.*]] = add i8 [[SEXT_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret i8 [[SUB]]
>  ;
>    %sext = sext i1 %y to i8
> @@ -393,8 +393,8 @@ define i32 @zextbool_sub_uses(i1 %c, i32 %x) {
>
>  define <4 x i32> @zextbool_sub_vector(<4 x i1> %c, <4 x i32> %x) {
>  ; CHECK-LABEL: @zextbool_sub_vector(
> -; CHECK-NEXT:    [[TMP1:%.*]] = sext <4 x i1> [[C:%.*]] to <4 x i32>
> -; CHECK-NEXT:    [[S:%.*]] = add <4 x i32> [[TMP1]], [[X:%.*]]
> +; CHECK-NEXT:    [[B_NEG:%.*]] = sext <4 x i1> [[C:%.*]] to <4 x i32>
> +; CHECK-NEXT:    [[S:%.*]] = add <4 x i32> [[B_NEG]], [[X:%.*]]
>  ; CHECK-NEXT:    ret <4 x i32> [[S]]
>  ;
>    %b = zext <4 x i1> %c to <4 x i32>
>
>
>
> _______________________________________________
> llvm-commits mailing list
> llvm-commits at lists.llvm.org
> https://lists.llvm.org/cgi-bin/mailman/listinfo/llvm-commits
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.llvm.org/pipermail/llvm-commits/attachments/20200422/ab35194d/attachment-0001.html>


More information about the llvm-commits mailing list