[llvm] r204071 - Make DAGCombiner work on vector bitshifts with constant splat vectors.

Matt Arsenault arsenm2 at gmail.com
Thu Apr 24 18:59:10 PDT 2014


On Apr 24, 2014, at 6:51 PM, Alex Rosenberg <alexr at ohmantics.com> wrote:

> I've just come across this and am wondering why this isn't handled in InstCombine. There's already a transform of (x << c1) << c2 into x << c1+c2 in there.
> 
> Alex
> 

It partially is, although many of the cases handled in instcombine currently don’t work for vectors, though I fixed some of those recently. When I ran into this, I was working on lowering extending loads of vectors, and this helped get rid of redundant operations after that.


> On Mar 17, 2014, at 11:58 AM, Matt Arsenault <Matthew.Arsenault at amd.com> wrote:
> 
>> Author: arsenm
>> Date: Mon Mar 17 13:58:01 2014
>> New Revision: 204071
>> 
>> URL: http://llvm.org/viewvc/llvm-project?rev=204071&view=rev
>> Log:
>> Make DAGCombiner work on vector bitshifts with constant splat vectors.
>> 
>> Modified:
>>   llvm/trunk/include/llvm/CodeGen/SelectionDAGNodes.h
>>   llvm/trunk/lib/CodeGen/SelectionDAG/DAGCombiner.cpp
>>   llvm/trunk/lib/CodeGen/SelectionDAG/SelectionDAG.cpp
>>   llvm/trunk/test/CodeGen/AArch64/neon-shl-ashr-lshr.ll
>>   llvm/trunk/test/CodeGen/X86/avx2-vector-shifts.ll
>>   llvm/trunk/test/CodeGen/X86/sse2-vector-shifts.ll
>> 
>> Modified: llvm/trunk/include/llvm/CodeGen/SelectionDAGNodes.h
>> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/include/llvm/CodeGen/SelectionDAGNodes.h?rev=204071&r1=204070&r2=204071&view=diff
>> ==============================================================================
>> --- llvm/trunk/include/llvm/CodeGen/SelectionDAGNodes.h (original)
>> +++ llvm/trunk/include/llvm/CodeGen/SelectionDAGNodes.h Mon Mar 17 13:58:01 2014
>> @@ -1522,6 +1522,11 @@ public:
>>                       unsigned MinSplatBits = 0,
>>                       bool isBigEndian = false) const;
>> 
>> +  /// isConstantSplat - Simpler form of isConstantSplat. Get the constant splat
>> +  /// when you only care about the value. Returns nullptr if this isn't a
>> +  /// constant splat vector.
>> +  ConstantSDNode *isConstantSplat() const;
>> +
>>  bool isConstant() const;
>> 
>>  static inline bool classof(const SDNode *N) {
>> 
>> Modified: llvm/trunk/lib/CodeGen/SelectionDAG/DAGCombiner.cpp
>> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/CodeGen/SelectionDAG/DAGCombiner.cpp?rev=204071&r1=204070&r2=204071&view=diff
>> ==============================================================================
>> --- llvm/trunk/lib/CodeGen/SelectionDAG/DAGCombiner.cpp (original)
>> +++ llvm/trunk/lib/CodeGen/SelectionDAG/DAGCombiner.cpp Mon Mar 17 13:58:01 2014
>> @@ -280,7 +280,7 @@ namespace {
>>    SDValue XformToShuffleWithZero(SDNode *N);
>>    SDValue ReassociateOps(unsigned Opc, SDLoc DL, SDValue LHS, SDValue RHS);
>> 
>> -    SDValue visitShiftByConstant(SDNode *N, unsigned Amt);
>> +    SDValue visitShiftByConstant(SDNode *N, ConstantSDNode *Amt);
>> 
>>    bool SimplifySelectOps(SDNode *SELECT, SDValue LHS, SDValue RHS);
>>    SDValue SimplifyBinOpWithSameOpcodeHands(SDNode *N);
>> @@ -634,7 +634,23 @@ static bool isOneUseSetCC(SDValue N) {
>>  return false;
>> }
>> 
>> -// \brief Returns the SDNode if it is a constant BuildVector or constant int.
>> +/// isConstantSplatVector - Returns true if N is a BUILD_VECTOR node whose
>> +/// elements are all the same constant or undefined.
>> +static bool isConstantSplatVector(SDNode *N, APInt& SplatValue) {
>> +  BuildVectorSDNode *C = dyn_cast<BuildVectorSDNode>(N);
>> +  if (!C)
>> +    return false;
>> +
>> +  APInt SplatUndef;
>> +  unsigned SplatBitSize;
>> +  bool HasAnyUndefs;
>> +  EVT EltVT = N->getValueType(0).getVectorElementType();
>> +  return (C->isConstantSplat(SplatValue, SplatUndef, SplatBitSize,
>> +                             HasAnyUndefs) &&
>> +          EltVT.getSizeInBits() >= SplatBitSize);
>> +}
>> +
>> +// \brief Returns the SDNode if it is a constant BuildVector or constant.
>> static SDNode *isConstantBuildVectorOrConstantInt(SDValue N) {
>>  if (isa<ConstantSDNode>(N))
>>    return N.getNode();
>> @@ -644,6 +660,18 @@ static SDNode *isConstantBuildVectorOrCo
>>  return NULL;
>> }
>> 
>> +// \brief Returns the SDNode if it is a constant splat BuildVector or constant
>> +// int.
>> +static ConstantSDNode *isConstOrConstSplat(SDValue N) {
>> +  if (ConstantSDNode *CN = dyn_cast<ConstantSDNode>(N))
>> +    return CN;
>> +
>> +  if (BuildVectorSDNode *BV = dyn_cast<BuildVectorSDNode>(N))
>> +    return BV->isConstantSplat();
>> +
>> +  return nullptr;
>> +}
>> +
>> SDValue DAGCombiner::ReassociateOps(unsigned Opc, SDLoc DL,
>>                                    SDValue N0, SDValue N1) {
>>  EVT VT = N0.getValueType();
>> @@ -1830,22 +1858,6 @@ SDValue DAGCombiner::visitSUBE(SDNode *N
>>  return SDValue();
>> }
>> 
>> -/// isConstantSplatVector - Returns true if N is a BUILD_VECTOR node whose
>> -/// elements are all the same constant or undefined.
>> -static bool isConstantSplatVector(SDNode *N, APInt& SplatValue) {
>> -  BuildVectorSDNode *C = dyn_cast<BuildVectorSDNode>(N);
>> -  if (!C)
>> -    return false;
>> -
>> -  APInt SplatUndef;
>> -  unsigned SplatBitSize;
>> -  bool HasAnyUndefs;
>> -  EVT EltVT = N->getValueType(0).getVectorElementType();
>> -  return (C->isConstantSplat(SplatValue, SplatUndef, SplatBitSize,
>> -                             HasAnyUndefs) &&
>> -          EltVT.getSizeInBits() >= SplatBitSize);
>> -}
>> -
>> SDValue DAGCombiner::visitMUL(SDNode *N) {
>>  SDValue N0 = N->getOperand(0);
>>  SDValue N1 = N->getOperand(1);
>> @@ -3805,11 +3817,9 @@ SDValue DAGCombiner::visitXOR(SDNode *N)
>> 
>> /// visitShiftByConstant - Handle transforms common to the three shifts, when
>> /// the shift amount is a constant.
>> -SDValue DAGCombiner::visitShiftByConstant(SDNode *N, unsigned Amt) {
>> -  assert(isa<ConstantSDNode>(N->getOperand(1)) &&
>> -         "Expected an ConstantSDNode operand.");
>> +SDValue DAGCombiner::visitShiftByConstant(SDNode *N, ConstantSDNode *Amt) {
>>  // We can't and shouldn't fold opaque constants.
>> -  if (cast<ConstantSDNode>(N->getOperand(1))->isOpaque())
>> +  if (Amt->isOpaque())
>>    return SDValue();
>> 
>>  SDNode *LHS = N->getOperand(0).getNode();
>> @@ -3888,11 +3898,11 @@ SDValue DAGCombiner::distributeTruncateT
>>  if (N->hasOneUse() && N->getOperand(0).hasOneUse()) {
>>    SDValue N01 = N->getOperand(0).getOperand(1);
>> 
>> -    if (ConstantSDNode *N01C = dyn_cast<ConstantSDNode>(N01)) {
>> +    if (ConstantSDNode *N01C = isConstOrConstSplat(N01)) {
>>      EVT TruncVT = N->getValueType(0);
>>      SDValue N00 = N->getOperand(0).getOperand(0);
>>      APInt TruncC = N01C->getAPIntValue();
>> -      TruncC = TruncC.trunc(TruncVT.getScalarType().getSizeInBits());
>> +      TruncC = TruncC.trunc(TruncVT.getScalarSizeInBits());
>> 
>>      return DAG.getNode(ISD::AND, SDLoc(N), TruncVT,
>>                         DAG.getNode(ISD::TRUNCATE, SDLoc(N), TruncVT, N00),
>> @@ -3921,7 +3931,7 @@ SDValue DAGCombiner::visitSHL(SDNode *N)
>>  ConstantSDNode *N0C = dyn_cast<ConstantSDNode>(N0);
>>  ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1);
>>  EVT VT = N0.getValueType();
>> -  unsigned OpSizeInBits = VT.getScalarType().getSizeInBits();
>> +  unsigned OpSizeInBits = VT.getScalarSizeInBits();
>> 
>>  // fold vector ops
>>  if (VT.isVector()) {
>> @@ -3931,18 +3941,21 @@ SDValue DAGCombiner::visitSHL(SDNode *N)
>>    BuildVectorSDNode *N1CV = dyn_cast<BuildVectorSDNode>(N1);
>>    // If setcc produces all-one true value then:
>>    // (shl (and (setcc) N01CV) N1CV) -> (and (setcc) N01CV<<N1CV)
>> -    if (N1CV && N1CV->isConstant() &&
>> -        TLI.getBooleanContents(true) ==
>> -          TargetLowering::ZeroOrNegativeOneBooleanContent &&
>> -        N0.getOpcode() == ISD::AND) {
>> -      SDValue N00 = N0->getOperand(0);
>> -      SDValue N01 = N0->getOperand(1);
>> -      BuildVectorSDNode *N01CV = dyn_cast<BuildVectorSDNode>(N01);
>> -
>> -      if (N01CV && N01CV->isConstant() && N00.getOpcode() == ISD::SETCC) {
>> -        SDValue C = DAG.FoldConstantArithmetic(ISD::SHL, VT, N01CV, N1CV);
>> -        if (C.getNode())
>> -          return DAG.getNode(ISD::AND, SDLoc(N), VT, N00, C);
>> +    if (N1CV && N1CV->isConstant()) {
>> +      if (N0.getOpcode() == ISD::AND &&
>> +          TLI.getBooleanContents(true) ==
>> +          TargetLowering::ZeroOrNegativeOneBooleanContent) {
>> +        SDValue N00 = N0->getOperand(0);
>> +        SDValue N01 = N0->getOperand(1);
>> +        BuildVectorSDNode *N01CV = dyn_cast<BuildVectorSDNode>(N01);
>> +
>> +        if (N01CV && N01CV->isConstant() && N00.getOpcode() == ISD::SETCC) {
>> +          SDValue C = DAG.FoldConstantArithmetic(ISD::SHL, VT, N01CV, N1CV);
>> +          if (C.getNode())
>> +            return DAG.getNode(ISD::AND, SDLoc(N), VT, N00, C);
>> +        }
>> +      } else {
>> +        N1C = isConstOrConstSplat(N1);
>>      }
>>    }
>>  }
>> @@ -3978,14 +3991,15 @@ SDValue DAGCombiner::visitSHL(SDNode *N)
>>    return SDValue(N, 0);
>> 
>>  // fold (shl (shl x, c1), c2) -> 0 or (shl x, (add c1, c2))
>> -  if (N1C && N0.getOpcode() == ISD::SHL &&
>> -      N0.getOperand(1).getOpcode() == ISD::Constant) {
>> -    uint64_t c1 = cast<ConstantSDNode>(N0.getOperand(1))->getZExtValue();
>> -    uint64_t c2 = N1C->getZExtValue();
>> -    if (c1 + c2 >= OpSizeInBits)
>> -      return DAG.getConstant(0, VT);
>> -    return DAG.getNode(ISD::SHL, SDLoc(N), VT, N0.getOperand(0),
>> -                       DAG.getConstant(c1 + c2, N1.getValueType()));
>> +  if (N1C && N0.getOpcode() == ISD::SHL) {
>> +    if (ConstantSDNode *N0C1 = isConstOrConstSplat(N0.getOperand(1))) {
>> +      uint64_t c1 = N0C1->getZExtValue();
>> +      uint64_t c2 = N1C->getZExtValue();
>> +      if (c1 + c2 >= OpSizeInBits)
>> +        return DAG.getConstant(0, VT);
>> +      return DAG.getNode(ISD::SHL, SDLoc(N), VT, N0.getOperand(0),
>> +                         DAG.getConstant(c1 + c2, N1.getValueType()));
>> +    }
>>  }
>> 
>>  // fold (shl (ext (shl x, c1)), c2) -> (ext (shl x, (add c1, c2)))
>> @@ -3996,20 +4010,21 @@ SDValue DAGCombiner::visitSHL(SDNode *N)
>>  if (N1C && (N0.getOpcode() == ISD::ZERO_EXTEND ||
>>              N0.getOpcode() == ISD::ANY_EXTEND ||
>>              N0.getOpcode() == ISD::SIGN_EXTEND) &&
>> -      N0.getOperand(0).getOpcode() == ISD::SHL &&
>> -      isa<ConstantSDNode>(N0.getOperand(0)->getOperand(1))) {
>> -    uint64_t c1 =
>> -      cast<ConstantSDNode>(N0.getOperand(0)->getOperand(1))->getZExtValue();
>> -    uint64_t c2 = N1C->getZExtValue();
>> -    EVT InnerShiftVT = N0.getOperand(0).getValueType();
>> -    uint64_t InnerShiftSize = InnerShiftVT.getScalarType().getSizeInBits();
>> -    if (c2 >= OpSizeInBits - InnerShiftSize) {
>> -      if (c1 + c2 >= OpSizeInBits)
>> -        return DAG.getConstant(0, VT);
>> -      return DAG.getNode(ISD::SHL, SDLoc(N0), VT,
>> -                         DAG.getNode(N0.getOpcode(), SDLoc(N0), VT,
>> -                                     N0.getOperand(0)->getOperand(0)),
>> -                         DAG.getConstant(c1 + c2, N1.getValueType()));
>> +      N0.getOperand(0).getOpcode() == ISD::SHL) {
>> +    SDValue N0Op0 = N0.getOperand(0);
>> +    if (ConstantSDNode *N0Op0C1 = isConstOrConstSplat(N0Op0.getOperand(1))) {
>> +      uint64_t c1 = N0Op0C1->getZExtValue();
>> +      uint64_t c2 = N1C->getZExtValue();
>> +      EVT InnerShiftVT = N0Op0.getValueType();
>> +      uint64_t InnerShiftSize = InnerShiftVT.getScalarSizeInBits();
>> +      if (c2 >= OpSizeInBits - InnerShiftSize) {
>> +        if (c1 + c2 >= OpSizeInBits)
>> +          return DAG.getConstant(0, VT);
>> +        return DAG.getNode(ISD::SHL, SDLoc(N0), VT,
>> +                           DAG.getNode(N0.getOpcode(), SDLoc(N0), VT,
>> +                                       N0Op0->getOperand(0)),
>> +                           DAG.getConstant(c1 + c2, N1.getValueType()));
>> +      }
>>    }
>>  }
>> 
>> @@ -4017,19 +4032,20 @@ SDValue DAGCombiner::visitSHL(SDNode *N)
>>  // Only fold this if the inner zext has no other uses to avoid increasing
>>  // the total number of instructions.
>>  if (N1C && N0.getOpcode() == ISD::ZERO_EXTEND && N0.hasOneUse() &&
>> -      N0.getOperand(0).getOpcode() == ISD::SRL &&
>> -      isa<ConstantSDNode>(N0.getOperand(0)->getOperand(1))) {
>> -    uint64_t c1 =
>> -      cast<ConstantSDNode>(N0.getOperand(0)->getOperand(1))->getZExtValue();
>> -    if (c1 < VT.getSizeInBits()) {
>> -      uint64_t c2 = N1C->getZExtValue();
>> -      if (c1 == c2) {
>> -        SDValue NewOp0 = N0.getOperand(0);
>> -        EVT CountVT = NewOp0.getOperand(1).getValueType();
>> -        SDValue NewSHL = DAG.getNode(ISD::SHL, SDLoc(N), NewOp0.getValueType(),
>> -                                     NewOp0, DAG.getConstant(c2, CountVT));
>> -        AddToWorkList(NewSHL.getNode());
>> -        return DAG.getNode(ISD::ZERO_EXTEND, SDLoc(N0), VT, NewSHL);
>> +      N0.getOperand(0).getOpcode() == ISD::SRL) {
>> +    SDValue N0Op0 = N0.getOperand(0);
>> +    if (ConstantSDNode *N0Op0C1 = isConstOrConstSplat(N0Op0.getOperand(1))) {
>> +      uint64_t c1 = N0Op0C1->getZExtValue();
>> +      if (c1 < VT.getScalarSizeInBits()) {
>> +        uint64_t c2 = N1C->getZExtValue();
>> +        if (c1 == c2) {
>> +          SDValue NewOp0 = N0.getOperand(0);
>> +          EVT CountVT = NewOp0.getOperand(1).getValueType();
>> +          SDValue NewSHL = DAG.getNode(ISD::SHL, SDLoc(N), NewOp0.getValueType(),
>> +                                       NewOp0, DAG.getConstant(c2, CountVT));
>> +          AddToWorkList(NewSHL.getNode());
>> +          return DAG.getNode(ISD::ZERO_EXTEND, SDLoc(N0), VT, NewSHL);
>> +        }
>>      }
>>    }
>>  }
>> @@ -4038,40 +4054,39 @@ SDValue DAGCombiner::visitSHL(SDNode *N)
>>  //                               (and (srl x, (sub c1, c2), MASK)
>>  // Only fold this if the inner shift has no other uses -- if it does, folding
>>  // this will increase the total number of instructions.
>> -  if (N1C && N0.getOpcode() == ISD::SRL && N0.hasOneUse() &&
>> -      N0.getOperand(1).getOpcode() == ISD::Constant) {
>> -    uint64_t c1 = cast<ConstantSDNode>(N0.getOperand(1))->getZExtValue();
>> -    if (c1 < VT.getSizeInBits()) {
>> -      uint64_t c2 = N1C->getZExtValue();
>> -      APInt Mask = APInt::getHighBitsSet(VT.getSizeInBits(),
>> -                                         VT.getSizeInBits() - c1);
>> -      SDValue Shift;
>> -      if (c2 > c1) {
>> -        Mask = Mask.shl(c2-c1);
>> -        Shift = DAG.getNode(ISD::SHL, SDLoc(N), VT, N0.getOperand(0),
>> -                            DAG.getConstant(c2-c1, N1.getValueType()));
>> -      } else {
>> -        Mask = Mask.lshr(c1-c2);
>> -        Shift = DAG.getNode(ISD::SRL, SDLoc(N), VT, N0.getOperand(0),
>> -                            DAG.getConstant(c1-c2, N1.getValueType()));
>> +  if (N1C && N0.getOpcode() == ISD::SRL && N0.hasOneUse()) {
>> +    if (ConstantSDNode *N0C1 = isConstOrConstSplat(N0.getOperand(1))) {
>> +      uint64_t c1 = N0C1->getZExtValue();
>> +      if (c1 < OpSizeInBits) {
>> +        uint64_t c2 = N1C->getZExtValue();
>> +        APInt Mask = APInt::getHighBitsSet(OpSizeInBits, OpSizeInBits - c1);
>> +        SDValue Shift;
>> +        if (c2 > c1) {
>> +          Mask = Mask.shl(c2 - c1);
>> +          Shift = DAG.getNode(ISD::SHL, SDLoc(N), VT, N0.getOperand(0),
>> +                              DAG.getConstant(c2 - c1, N1.getValueType()));
>> +        } else {
>> +          Mask = Mask.lshr(c1 - c2);
>> +          Shift = DAG.getNode(ISD::SRL, SDLoc(N), VT, N0.getOperand(0),
>> +                              DAG.getConstant(c1 - c2, N1.getValueType()));
>> +        }
>> +        return DAG.getNode(ISD::AND, SDLoc(N0), VT, Shift,
>> +                           DAG.getConstant(Mask, VT));
>>      }
>> -      return DAG.getNode(ISD::AND, SDLoc(N0), VT, Shift,
>> -                         DAG.getConstant(Mask, VT));
>>    }
>>  }
>>  // fold (shl (sra x, c1), c1) -> (and x, (shl -1, c1))
>>  if (N1C && N0.getOpcode() == ISD::SRA && N1 == N0.getOperand(1)) {
>> +    unsigned BitSize = VT.getScalarSizeInBits();
>>    SDValue HiBitsMask =
>> -      DAG.getConstant(APInt::getHighBitsSet(VT.getSizeInBits(),
>> -                                            VT.getSizeInBits() -
>> -                                              N1C->getZExtValue()),
>> -                      VT);
>> +      DAG.getConstant(APInt::getHighBitsSet(BitSize,
>> +                                            BitSize - N1C->getZExtValue()), VT);
>>    return DAG.getNode(ISD::AND, SDLoc(N), VT, N0.getOperand(0),
>>                       HiBitsMask);
>>  }
>> 
>>  if (N1C) {
>> -    SDValue NewSHL = visitShiftByConstant(N, N1C->getZExtValue());
>> +    SDValue NewSHL = visitShiftByConstant(N, N1C);
>>    if (NewSHL.getNode())
>>      return NewSHL;
>>  }
>> @@ -4091,6 +4106,8 @@ SDValue DAGCombiner::visitSRA(SDNode *N)
>>  if (VT.isVector()) {
>>    SDValue FoldedVOp = SimplifyVBinOp(N);
>>    if (FoldedVOp.getNode()) return FoldedVOp;
>> +
>> +    N1C = isConstOrConstSplat(N1);
>>  }
>> 
>>  // fold (sra c1, c2) -> (sra c1, c2)
>> @@ -4124,11 +4141,12 @@ SDValue DAGCombiner::visitSRA(SDNode *N)
>> 
>>  // fold (sra (sra x, c1), c2) -> (sra x, (add c1, c2))
>>  if (N1C && N0.getOpcode() == ISD::SRA) {
>> -    if (ConstantSDNode *C1 = dyn_cast<ConstantSDNode>(N0.getOperand(1))) {
>> +    if (ConstantSDNode *C1 = isConstOrConstSplat(N0.getOperand(1))) {
>>      unsigned Sum = N1C->getZExtValue() + C1->getZExtValue();
>> -      if (Sum >= OpSizeInBits) Sum = OpSizeInBits-1;
>> +      if (Sum >= OpSizeInBits)
>> +        Sum = OpSizeInBits - 1;
>>      return DAG.getNode(ISD::SRA, SDLoc(N), VT, N0.getOperand(0),
>> -                         DAG.getConstant(Sum, N1C->getValueType(0)));
>> +                         DAG.getConstant(Sum, N1.getValueType()));
>>    }
>>  }
>> 
>> @@ -4137,14 +4155,17 @@ SDValue DAGCombiner::visitSRA(SDNode *N)
>>  // result_size - n != m.
>>  // If truncate is free for the target sext(shl) is likely to result in better
>>  // code.
>> -  if (N0.getOpcode() == ISD::SHL) {
>> +  if (N0.getOpcode() == ISD::SHL && N1C) {
>>    // Get the two constanst of the shifts, CN0 = m, CN = n.
>> -    const ConstantSDNode *N01C = dyn_cast<ConstantSDNode>(N0.getOperand(1));
>> -    if (N01C && N1C) {
>> +    const ConstantSDNode *N01C = isConstOrConstSplat(N0.getOperand(1));
>> +    if (N01C) {
>> +      LLVMContext &Ctx = *DAG.getContext();
>>      // Determine what the truncate's result bitsize and type would be.
>> -      EVT TruncVT =
>> -        EVT::getIntegerVT(*DAG.getContext(),
>> -                          OpSizeInBits - N1C->getZExtValue());
>> +      EVT TruncVT = EVT::getIntegerVT(Ctx, OpSizeInBits - N1C->getZExtValue());
>> +
>> +      if (VT.isVector())
>> +        TruncVT = EVT::getVectorVT(Ctx, TruncVT, VT.getVectorNumElements());
>> +
>>      // Determine the residual right-shift amount.
>>      signed ShiftAmt = N1C->getZExtValue() - N01C->getZExtValue();
>> 
>> @@ -4177,26 +4198,27 @@ SDValue DAGCombiner::visitSRA(SDNode *N)
>>      return DAG.getNode(ISD::SRA, SDLoc(N), VT, N0, NewOp1);
>>  }
>> 
>> -  // fold (sra (trunc (sr x, c1)), c2) -> (trunc (sra x, c1+c2))
>> +  // fold (sra (trunc (srl x, c1)), c2) -> (trunc (sra x, c1 + c2))
>>  //      if c1 is equal to the number of bits the trunc removes
>>  if (N0.getOpcode() == ISD::TRUNCATE &&
>>      (N0.getOperand(0).getOpcode() == ISD::SRL ||
>>       N0.getOperand(0).getOpcode() == ISD::SRA) &&
>>      N0.getOperand(0).hasOneUse() &&
>>      N0.getOperand(0).getOperand(1).hasOneUse() &&
>> -      N1C && isa<ConstantSDNode>(N0.getOperand(0).getOperand(1))) {
>> -    EVT LargeVT = N0.getOperand(0).getValueType();
>> -    ConstantSDNode *LargeShiftAmt =
>> -      cast<ConstantSDNode>(N0.getOperand(0).getOperand(1));
>> -
>> -    if (LargeVT.getScalarType().getSizeInBits() - OpSizeInBits ==
>> -        LargeShiftAmt->getZExtValue()) {
>> -      SDValue Amt =
>> -        DAG.getConstant(LargeShiftAmt->getZExtValue() + N1C->getZExtValue(),
>> -              getShiftAmountTy(N0.getOperand(0).getOperand(0).getValueType()));
>> -      SDValue SRA = DAG.getNode(ISD::SRA, SDLoc(N), LargeVT,
>> -                                N0.getOperand(0).getOperand(0), Amt);
>> -      return DAG.getNode(ISD::TRUNCATE, SDLoc(N), VT, SRA);
>> +      N1C) {
>> +    SDValue N0Op0 = N0.getOperand(0);
>> +    if (ConstantSDNode *LargeShift = isConstOrConstSplat(N0Op0.getOperand(1))) {
>> +      unsigned LargeShiftVal = LargeShift->getZExtValue();
>> +      EVT LargeVT = N0Op0.getValueType();
>> +
>> +      if (LargeVT.getScalarSizeInBits() - OpSizeInBits == LargeShiftVal) {
>> +        SDValue Amt =
>> +          DAG.getConstant(LargeShiftVal + N1C->getZExtValue(),
>> +                          getShiftAmountTy(N0Op0.getOperand(0).getValueType()));
>> +        SDValue SRA = DAG.getNode(ISD::SRA, SDLoc(N), LargeVT,
>> +                                  N0Op0.getOperand(0), Amt);
>> +        return DAG.getNode(ISD::TRUNCATE, SDLoc(N), VT, SRA);
>> +      }
>>    }
>>  }
>> 
>> @@ -4210,7 +4232,7 @@ SDValue DAGCombiner::visitSRA(SDNode *N)
>>    return DAG.getNode(ISD::SRL, SDLoc(N), VT, N0, N1);
>> 
>>  if (N1C) {
>> -    SDValue NewSRA = visitShiftByConstant(N, N1C->getZExtValue());
>> +    SDValue NewSRA = visitShiftByConstant(N, N1C);
>>    if (NewSRA.getNode())
>>      return NewSRA;
>>  }
>> @@ -4230,6 +4252,8 @@ SDValue DAGCombiner::visitSRL(SDNode *N)
>>  if (VT.isVector()) {
>>    SDValue FoldedVOp = SimplifyVBinOp(N);
>>    if (FoldedVOp.getNode()) return FoldedVOp;
>> +
>> +    N1C = isConstOrConstSplat(N1);
>>  }
>> 
>>  // fold (srl c1, c2) -> c1 >>u c2
>> @@ -4250,14 +4274,15 @@ SDValue DAGCombiner::visitSRL(SDNode *N)
>>    return DAG.getConstant(0, VT);
>> 
>>  // fold (srl (srl x, c1), c2) -> 0 or (srl x, (add c1, c2))
>> -  if (N1C && N0.getOpcode() == ISD::SRL &&
>> -      N0.getOperand(1).getOpcode() == ISD::Constant) {
>> -    uint64_t c1 = cast<ConstantSDNode>(N0.getOperand(1))->getZExtValue();
>> -    uint64_t c2 = N1C->getZExtValue();
>> -    if (c1 + c2 >= OpSizeInBits)
>> -      return DAG.getConstant(0, VT);
>> -    return DAG.getNode(ISD::SRL, SDLoc(N), VT, N0.getOperand(0),
>> -                       DAG.getConstant(c1 + c2, N1.getValueType()));
>> +  if (N1C && N0.getOpcode() == ISD::SRL) {
>> +    if (ConstantSDNode *N01C = isConstOrConstSplat(N0.getOperand(1))) {
>> +      uint64_t c1 = N01C->getZExtValue();
>> +      uint64_t c2 = N1C->getZExtValue();
>> +      if (c1 + c2 >= OpSizeInBits)
>> +        return DAG.getConstant(0, VT);
>> +      return DAG.getNode(ISD::SRL, SDLoc(N), VT, N0.getOperand(0),
>> +                         DAG.getConstant(c1 + c2, N1.getValueType()));
>> +    }
>>  }
>> 
>>  // fold (srl (trunc (srl x, c1)), c2) -> 0 or (trunc (srl x, (add c1, c2)))
>> @@ -4282,18 +4307,21 @@ SDValue DAGCombiner::visitSRL(SDNode *N)
>>  }
>> 
>>  // fold (srl (shl x, c), c) -> (and x, cst2)
>> -  if (N1C && N0.getOpcode() == ISD::SHL && N0.getOperand(1) == N1 &&
>> -      N0.getValueSizeInBits() <= 64) {
>> -    uint64_t ShAmt = N1C->getZExtValue()+64-N0.getValueSizeInBits();
>> -    return DAG.getNode(ISD::AND, SDLoc(N), VT, N0.getOperand(0),
>> -                       DAG.getConstant(~0ULL >> ShAmt, VT));
>> +  if (N1C && N0.getOpcode() == ISD::SHL && N0.getOperand(1) == N1) {
>> +    unsigned BitSize = N0.getScalarValueSizeInBits();
>> +    if (BitSize <= 64) {
>> +      uint64_t ShAmt = N1C->getZExtValue() + 64 - BitSize;
>> +      return DAG.getNode(ISD::AND, SDLoc(N), VT, N0.getOperand(0),
>> +                         DAG.getConstant(~0ULL >> ShAmt, VT));
>> +    }
>>  }
>> 
>>  // fold (srl (anyextend x), c) -> (and (anyextend (srl x, c)), mask)
>>  if (N1C && N0.getOpcode() == ISD::ANY_EXTEND) {
>>    // Shifting in all undef bits?
>>    EVT SmallVT = N0.getOperand(0).getValueType();
>> -    if (N1C->getZExtValue() >= SmallVT.getSizeInBits())
>> +    unsigned BitSize = SmallVT.getScalarSizeInBits();
>> +    if (N1C->getZExtValue() >= BitSize)
>>      return DAG.getUNDEF(VT);
>> 
>>    if (!LegalTypes || TLI.isTypeDesirableForOp(ISD::SRL, SmallVT)) {
>> @@ -4302,7 +4330,7 @@ SDValue DAGCombiner::visitSRL(SDNode *N)
>>                                       N0.getOperand(0),
>>                          DAG.getConstant(ShiftAmt, getShiftAmountTy(SmallVT)));
>>      AddToWorkList(SmallShift.getNode());
>> -      APInt Mask = APInt::getAllOnesValue(VT.getSizeInBits()).lshr(ShiftAmt);
>> +      APInt Mask = APInt::getAllOnesValue(OpSizeInBits).lshr(ShiftAmt);
>>      return DAG.getNode(ISD::AND, SDLoc(N), VT,
>>                         DAG.getNode(ISD::ANY_EXTEND, SDLoc(N), VT, SmallShift),
>>                         DAG.getConstant(Mask, VT));
>> @@ -4311,14 +4339,14 @@ SDValue DAGCombiner::visitSRL(SDNode *N)
>> 
>>  // fold (srl (sra X, Y), 31) -> (srl X, 31).  This srl only looks at the sign
>>  // bit, which is unmodified by sra.
>> -  if (N1C && N1C->getZExtValue() + 1 == VT.getSizeInBits()) {
>> +  if (N1C && N1C->getZExtValue() + 1 == OpSizeInBits) {
>>    if (N0.getOpcode() == ISD::SRA)
>>      return DAG.getNode(ISD::SRL, SDLoc(N), VT, N0.getOperand(0), N1);
>>  }
>> 
>>  // fold (srl (ctlz x), "5") -> x  iff x has one bit set (the low bit).
>>  if (N1C && N0.getOpcode() == ISD::CTLZ &&
>> -      N1C->getAPIntValue() == Log2_32(VT.getSizeInBits())) {
>> +      N1C->getAPIntValue() == Log2_32(OpSizeInBits)) {
>>    APInt KnownZero, KnownOne;
>>    DAG.ComputeMaskedBits(N0.getOperand(0), KnownZero, KnownOne);
>> 
>> @@ -4365,7 +4393,7 @@ SDValue DAGCombiner::visitSRL(SDNode *N)
>>    return SDValue(N, 0);
>> 
>>  if (N1C) {
>> -    SDValue NewSRL = visitShiftByConstant(N, N1C->getZExtValue());
>> +    SDValue NewSRL = visitShiftByConstant(N, N1C);
>>    if (NewSRL.getNode())
>>      return NewSRL;
>>  }
>> 
>> Modified: llvm/trunk/lib/CodeGen/SelectionDAG/SelectionDAG.cpp
>> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/CodeGen/SelectionDAG/SelectionDAG.cpp?rev=204071&r1=204070&r2=204071&view=diff
>> ==============================================================================
>> --- llvm/trunk/lib/CodeGen/SelectionDAG/SelectionDAG.cpp (original)
>> +++ llvm/trunk/lib/CodeGen/SelectionDAG/SelectionDAG.cpp Mon Mar 17 13:58:01 2014
>> @@ -6573,6 +6573,19 @@ bool BuildVectorSDNode::isConstantSplat(
>>  return true;
>> }
>> 
>> +ConstantSDNode *BuildVectorSDNode::isConstantSplat() const {
>> +  SDValue Op0 = getOperand(0);
>> +  for (unsigned i = 0, e = getNumOperands(); i != e; ++i) {
>> +    SDValue Opi = getOperand(i);
>> +    unsigned Opc = Opi.getOpcode();
>> +    if ((Opc != ISD::UNDEF && Opc != ISD::Constant && Opc != ISD::ConstantFP) ||
>> +        Opi != Op0)
>> +      return nullptr;
>> +  }
>> +
>> +  return cast<ConstantSDNode>(Op0);
>> +}
>> +
>> bool BuildVectorSDNode::isConstant() const {
>>  for (unsigned i = 0, e = getNumOperands(); i != e; ++i) {
>>    unsigned Opc = getOperand(i).getOpcode();
>> 
>> Modified: llvm/trunk/test/CodeGen/AArch64/neon-shl-ashr-lshr.ll
>> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/AArch64/neon-shl-ashr-lshr.ll?rev=204071&r1=204070&r2=204071&view=diff
>> ==============================================================================
>> --- llvm/trunk/test/CodeGen/AArch64/neon-shl-ashr-lshr.ll (original)
>> +++ llvm/trunk/test/CodeGen/AArch64/neon-shl-ashr-lshr.ll Mon Mar 17 13:58:01 2014
>> @@ -186,14 +186,14 @@ define <2 x i64> @ashr.v2i64(<2 x i64> %
>> 
>> define <1 x i64> @shl.v1i64.0(<1 x i64> %a) {
>> ; CHECK-LABEL: shl.v1i64.0:
>> -; CHECK: shl d{{[0-9]+}}, d{{[0-9]+}}, #0
>> +; CHECK-NOT: shl d{{[0-9]+}}, d{{[0-9]+}}, #0
>>  %c = shl <1 x i64> %a, zeroinitializer
>>  ret <1 x i64> %c
>> }
>> 
>> define <2 x i32> @shl.v2i32.0(<2 x i32> %a) {
>> ; CHECK-LABEL: shl.v2i32.0:
>> -; CHECK: shl v{{[0-9]+}}.2s, v{{[0-9]+}}.2s, #0
>> +; CHECK-NOT: shl v{{[0-9]+}}.2s, v{{[0-9]+}}.2s, #0
>>  %c = shl <2 x i32> %a, zeroinitializer
>>  ret <2 x i32> %c
>> }
>> @@ -285,7 +285,7 @@ define <1 x i16> @shl.v1i16.imm(<1 x i16
>> 
>> define <1 x i32> @shl.v1i32.imm(<1 x i32> %a) {
>> ; CHECK-LABEL: shl.v1i32.imm:
>> -; CHECK: shl v{{[0-9]+}}.2s, v{{[0-9]+}}.2s, #0
>> +; CHECK-NOT: shl v{{[0-9]+}}.2s, v{{[0-9]+}}.2s, #0
>>  %c = shl <1 x i32> %a, zeroinitializer
>>  ret <1 x i32> %c
>> }
>> 
>> Modified: llvm/trunk/test/CodeGen/X86/avx2-vector-shifts.ll
>> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/X86/avx2-vector-shifts.ll?rev=204071&r1=204070&r2=204071&view=diff
>> ==============================================================================
>> --- llvm/trunk/test/CodeGen/X86/avx2-vector-shifts.ll (original)
>> +++ llvm/trunk/test/CodeGen/X86/avx2-vector-shifts.ll Mon Mar 17 13:58:01 2014
>> @@ -9,7 +9,7 @@ entry:
>> }
>> 
>> ; CHECK-LABEL: test_sllw_1:
>> -; CHECK: vpsllw  $0, %ymm0, %ymm0
>> +; CHECK-NOT: vpsllw  $0, %ymm0, %ymm0
>> ; CHECK: ret
>> 
>> define <16 x i16> @test_sllw_2(<16 x i16> %InVec) {
>> @@ -39,7 +39,7 @@ entry:
>> }
>> 
>> ; CHECK-LABEL: test_slld_1:
>> -; CHECK: vpslld  $0, %ymm0, %ymm0
>> +; CHECK-NOT: vpslld  $0, %ymm0, %ymm0
>> ; CHECK: ret
>> 
>> define <8 x i32> @test_slld_2(<8 x i32> %InVec) {
>> @@ -69,7 +69,7 @@ entry:
>> }
>> 
>> ; CHECK-LABEL: test_sllq_1:
>> -; CHECK: vpsllq  $0, %ymm0, %ymm0
>> +; CHECK-NOT: vpsllq  $0, %ymm0, %ymm0
>> ; CHECK: ret
>> 
>> define <4 x i64> @test_sllq_2(<4 x i64> %InVec) {
>> @@ -101,7 +101,7 @@ entry:
>> }
>> 
>> ; CHECK-LABEL: test_sraw_1:
>> -; CHECK: vpsraw  $0, %ymm0, %ymm0
>> +; CHECK-NOT: vpsraw  $0, %ymm0, %ymm0
>> ; CHECK: ret
>> 
>> define <16 x i16> @test_sraw_2(<16 x i16> %InVec) {
>> @@ -131,7 +131,7 @@ entry:
>> }
>> 
>> ; CHECK-LABEL: test_srad_1:
>> -; CHECK: vpsrad  $0, %ymm0, %ymm0
>> +; CHECK-NOT: vpsrad  $0, %ymm0, %ymm0
>> ; CHECK: ret
>> 
>> define <8 x i32> @test_srad_2(<8 x i32> %InVec) {
>> @@ -163,7 +163,7 @@ entry:
>> }
>> 
>> ; CHECK-LABEL: test_srlw_1:
>> -; CHECK: vpsrlw  $0, %ymm0, %ymm0
>> +; CHECK-NOT: vpsrlw  $0, %ymm0, %ymm0
>> ; CHECK: ret
>> 
>> define <16 x i16> @test_srlw_2(<16 x i16> %InVec) {
>> @@ -193,7 +193,7 @@ entry:
>> }
>> 
>> ; CHECK-LABEL: test_srld_1:
>> -; CHECK: vpsrld  $0, %ymm0, %ymm0
>> +; CHECK-NOT: vpsrld  $0, %ymm0, %ymm0
>> ; CHECK: ret
>> 
>> define <8 x i32> @test_srld_2(<8 x i32> %InVec) {
>> @@ -223,7 +223,7 @@ entry:
>> }
>> 
>> ; CHECK-LABEL: test_srlq_1:
>> -; CHECK: vpsrlq  $0, %ymm0, %ymm0
>> +; CHECK-NOT: vpsrlq  $0, %ymm0, %ymm0
>> ; CHECK: ret
>> 
>> define <4 x i64> @test_srlq_2(<4 x i64> %InVec) {
>> @@ -245,3 +245,14 @@ entry:
>> ; CHECK-LABEL: test_srlq_3:
>> ; CHECK: vpsrlq $63, %ymm0, %ymm0
>> ; CHECK: ret
>> +
>> +; CHECK-LABEL: @srl_trunc_and_v4i64
>> +; CHECK: vpand
>> +; CHECK-NEXT: vpsrlvd
>> +; CHECK: ret
>> +define <4 x i32> @srl_trunc_and_v4i64(<4 x i32> %x, <4 x i64> %y) nounwind {
>> +  %and = and <4 x i64> %y, <i64 8, i64 8, i64 8, i64 8>
>> +  %trunc = trunc <4 x i64> %and to <4 x i32>
>> +  %sra = lshr <4 x i32> %x, %trunc
>> +  ret <4 x i32> %sra
>> +}
>> 
>> Modified: llvm/trunk/test/CodeGen/X86/sse2-vector-shifts.ll
>> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/X86/sse2-vector-shifts.ll?rev=204071&r1=204070&r2=204071&view=diff
>> ==============================================================================
>> --- llvm/trunk/test/CodeGen/X86/sse2-vector-shifts.ll (original)
>> +++ llvm/trunk/test/CodeGen/X86/sse2-vector-shifts.ll Mon Mar 17 13:58:01 2014
>> @@ -9,8 +9,8 @@ entry:
>> }
>> 
>> ; CHECK-LABEL: test_sllw_1:
>> -; CHECK: psllw   $0, %xmm0
>> -; CHECK-NEXT: ret
>> +; CHECK-NOT: psllw   $0, %xmm0
>> +; CHECK: ret
>> 
>> define <8 x i16> @test_sllw_2(<8 x i16> %InVec) {
>> entry:
>> @@ -39,8 +39,8 @@ entry:
>> }
>> 
>> ; CHECK-LABEL: test_slld_1:
>> -; CHECK: pslld   $0, %xmm0
>> -; CHECK-NEXT: ret
>> +; CHECK-NOT: pslld   $0, %xmm0
>> +; CHECK: ret
>> 
>> define <4 x i32> @test_slld_2(<4 x i32> %InVec) {
>> entry:
>> @@ -69,8 +69,8 @@ entry:
>> }
>> 
>> ; CHECK-LABEL: test_sllq_1:
>> -; CHECK: psllq   $0, %xmm0
>> -; CHECK-NEXT: ret
>> +; CHECK-NOT: psllq   $0, %xmm0
>> +; CHECK: ret
>> 
>> define <2 x i64> @test_sllq_2(<2 x i64> %InVec) {
>> entry:
>> @@ -101,8 +101,8 @@ entry:
>> }
>> 
>> ; CHECK-LABEL: test_sraw_1:
>> -; CHECK: psraw   $0, %xmm0
>> -; CHECK-NEXT: ret
>> +; CHECK-NOT: psraw   $0, %xmm0
>> +; CHECK: ret
>> 
>> define <8 x i16> @test_sraw_2(<8 x i16> %InVec) {
>> entry:
>> @@ -131,8 +131,8 @@ entry:
>> }
>> 
>> ; CHECK-LABEL: test_srad_1:
>> -; CHECK: psrad   $0, %xmm0
>> -; CHECK-NEXT: ret
>> +; CHECK-NOT: psrad   $0, %xmm0
>> +; CHECK: ret
>> 
>> define <4 x i32> @test_srad_2(<4 x i32> %InVec) {
>> entry:
>> @@ -163,8 +163,8 @@ entry:
>> }
>> 
>> ; CHECK-LABEL: test_srlw_1:
>> -; CHECK: psrlw   $0, %xmm0
>> -; CHECK-NEXT: ret
>> +; CHECK-NOT: psrlw   $0, %xmm0
>> +; CHECK: ret
>> 
>> define <8 x i16> @test_srlw_2(<8 x i16> %InVec) {
>> entry:
>> @@ -193,8 +193,8 @@ entry:
>> }
>> 
>> ; CHECK-LABEL: test_srld_1:
>> -; CHECK: psrld   $0, %xmm0
>> -; CHECK-NEXT: ret
>> +; CHECK-NOT: psrld   $0, %xmm0
>> +; CHECK: ret
>> 
>> define <4 x i32> @test_srld_2(<4 x i32> %InVec) {
>> entry:
>> @@ -223,8 +223,8 @@ entry:
>> }
>> 
>> ; CHECK-LABEL: test_srlq_1:
>> -; CHECK: psrlq   $0, %xmm0
>> -; CHECK-NEXT: ret
>> +; CHECK-NOT: psrlq   $0, %xmm0
>> +; CHECK: ret
>> 
>> define <2 x i64> @test_srlq_2(<2 x i64> %InVec) {
>> entry:
>> @@ -245,3 +245,123 @@ entry:
>> ; CHECK-LABEL: test_srlq_3:
>> ; CHECK: psrlq $63, %xmm0
>> ; CHECK-NEXT: ret
>> +
>> +
>> +; CHECK-LABEL: sra_sra_v4i32:
>> +; CHECK: psrad $6, %xmm0
>> +; CHECK-NEXT: retq
>> +define <4 x i32> @sra_sra_v4i32(<4 x i32> %x) nounwind {
>> +  %sra0 = ashr <4 x i32> %x, <i32 2, i32 2, i32 2, i32 2>
>> +  %sra1 = ashr <4 x i32> %sra0, <i32 4, i32 4, i32 4, i32 4>
>> +  ret <4 x i32> %sra1
>> +}
>> +
>> +; CHECK-LABEL: @srl_srl_v4i32
>> +; CHECK: psrld $6, %xmm0
>> +; CHECK-NEXT: ret
>> +define <4 x i32> @srl_srl_v4i32(<4 x i32> %x) nounwind {
>> +  %srl0 = lshr <4 x i32> %x, <i32 2, i32 2, i32 2, i32 2>
>> +  %srl1 = lshr <4 x i32> %srl0, <i32 4, i32 4, i32 4, i32 4>
>> +  ret <4 x i32> %srl1
>> +}
>> +
>> +; CHECK-LABEL: @srl_shl_v4i32
>> +; CHECK: andps
>> +; CHECK-NEXT: retq
>> +define <4 x i32> @srl_shl_v4i32(<4 x i32> %x) nounwind {
>> +  %srl0 = shl <4 x i32> %x, <i32 4, i32 4, i32 4, i32 4>
>> +  %srl1 = lshr <4 x i32> %srl0, <i32 4, i32 4, i32 4, i32 4>
>> +  ret <4 x i32> %srl1
>> +}
>> +
>> +; CHECK-LABEL: @srl_sra_31_v4i32
>> +; CHECK: psrld $31, %xmm0
>> +; CHECK-NEXT: ret
>> +define <4 x i32> @srl_sra_31_v4i32(<4 x i32> %x, <4 x i32> %y) nounwind {
>> +  %sra = ashr <4 x i32> %x, %y
>> +  %srl1 = lshr <4 x i32> %sra, <i32 31, i32 31, i32 31, i32 31>
>> +  ret <4 x i32> %srl1
>> +}
>> +
>> +; CHECK-LABEL: @shl_shl_v4i32
>> +; CHECK: pslld $6, %xmm0
>> +; CHECK-NEXT: ret
>> +define <4 x i32> @shl_shl_v4i32(<4 x i32> %x) nounwind {
>> +  %shl0 = shl <4 x i32> %x, <i32 2, i32 2, i32 2, i32 2>
>> +  %shl1 = shl <4 x i32> %shl0, <i32 4, i32 4, i32 4, i32 4>
>> +  ret <4 x i32> %shl1
>> +}
>> +
>> +; CHECK-LABEL: @shl_sra_v4i32
>> +; CHECK: andps
>> +; CHECK-NEXT: ret
>> +define <4 x i32> @shl_sra_v4i32(<4 x i32> %x) nounwind {
>> +  %shl0 = ashr <4 x i32> %x, <i32 4, i32 4, i32 4, i32 4>
>> +  %shl1 = shl <4 x i32> %shl0, <i32 4, i32 4, i32 4, i32 4>
>> +  ret <4 x i32> %shl1
>> +}
>> +
>> +; CHECK-LABEL: @shl_srl_v4i32
>> +; CHECK: pslld $3, %xmm0
>> +; CHECK-NEXT: pand
>> +; CHECK-NEXT: ret
>> +define <4 x i32> @shl_srl_v4i32(<4 x i32> %x) nounwind {
>> +  %shl0 = lshr <4 x i32> %x, <i32 2, i32 2, i32 2, i32 2>
>> +  %shl1 = shl <4 x i32> %shl0, <i32 5, i32 5, i32 5, i32 5>
>> +  ret <4 x i32> %shl1
>> +}
>> +
>> +; CHECK-LABEL: @shl_zext_srl_v4i32
>> +; CHECK: andps
>> +; CHECK-NEXT: ret
>> +define <4 x i32> @shl_zext_srl_v4i32(<4 x i16> %x) nounwind {
>> +  %srl = lshr <4 x i16> %x, <i16 2, i16 2, i16 2, i16 2>
>> +  %zext = zext <4 x i16> %srl to <4 x i32>
>> +  %shl = shl <4 x i32> %zext, <i32 2, i32 2, i32 2, i32 2>
>> +  ret <4 x i32> %shl
>> +}
>> +
>> +; CHECK: @sra_trunc_srl_v4i32
>> +; CHECK: psrad $19, %xmm0
>> +; CHECK-NEXT: retq
>> +define <4 x i16> @sra_trunc_srl_v4i32(<4 x i32> %x) nounwind {
>> +  %srl = lshr <4 x i32> %x, <i32 16, i32 16, i32 16, i32 16>
>> +  %trunc = trunc <4 x i32> %srl to <4 x i16>
>> +  %sra = ashr <4 x i16> %trunc, <i16 3, i16 3, i16 3, i16 3>
>> +  ret <4 x i16> %sra
>> +}
>> +
>> +; CHECK-LABEL: @shl_zext_shl_v4i32
>> +; CHECK: pand
>> +; CHECK-NEXT: pslld $19, %xmm0
>> +; CHECK-NEXT: ret
>> +define <4 x i32> @shl_zext_shl_v4i32(<4 x i16> %x) nounwind {
>> +  %shl0 = shl <4 x i16> %x, <i16 2, i16 2, i16 2, i16 2>
>> +  %ext = zext <4 x i16> %shl0 to <4 x i32>
>> +  %shl1 = shl <4 x i32> %ext, <i32 17, i32 17, i32 17, i32 17>
>> +  ret <4 x i32> %shl1
>> +}
>> +
>> +; CHECK-LABEL: @sra_v4i32
>> +; CHECK: psrad $3, %xmm0
>> +; CHECK-NEXT: ret
>> +define <4 x i32> @sra_v4i32(<4 x i32> %x) nounwind {
>> +  %sra = ashr <4 x i32> %x, <i32 3, i32 3, i32 3, i32 3>
>> +  ret <4 x i32> %sra
>> +}
>> +
>> +; CHECK-LABEL: @srl_v4i32
>> +; CHECK: psrld $3, %xmm0
>> +; CHECK-NEXT: ret
>> +define <4 x i32> @srl_v4i32(<4 x i32> %x) nounwind {
>> +  %sra = lshr <4 x i32> %x, <i32 3, i32 3, i32 3, i32 3>
>> +  ret <4 x i32> %sra
>> +}
>> +
>> +; CHECK-LABEL: @shl_v4i32
>> +; CHECK: pslld $3, %xmm0
>> +; CHECK-NEXT: ret
>> +define <4 x i32> @shl_v4i32(<4 x i32> %x) nounwind {
>> +  %sra = shl <4 x i32> %x, <i32 3, i32 3, i32 3, i32 3>
>> +  ret <4 x i32> %sra
>> +}
>> 
>> 
>> _______________________________________________
>> llvm-commits mailing list
>> llvm-commits at cs.uiuc.edu
>> http://lists.cs.uiuc.edu/mailman/listinfo/llvm-commits
> 
> 
> _______________________________________________
> llvm-commits mailing list
> llvm-commits at cs.uiuc.edu
> http://lists.cs.uiuc.edu/mailman/listinfo/llvm-commits





More information about the llvm-commits mailing list