[llvm] 09c575e - [InstCombine] add/move tests for shl with binop; NFC

Sanjay Patel via llvm-commits llvm-commits at lists.llvm.org
Tue Sep 28 11:46:35 PDT 2021


Author: Sanjay Patel
Date: 2021-09-28T14:46:27-04:00
New Revision: 09c575e728e2f24f76ea7f562e61fccaa225d72d

URL: https://github.com/llvm/llvm-project/commit/09c575e728e2f24f76ea7f562e61fccaa225d72d
DIFF: https://github.com/llvm/llvm-project/commit/09c575e728e2f24f76ea7f562e61fccaa225d72d.diff

LOG: [InstCombine] add/move tests for shl with binop; NFC

Added: 
    llvm/test/Transforms/InstCombine/shl-bo.ll

Modified: 
    llvm/test/Transforms/InstCombine/shift.ll

Removed: 
    llvm/test/Transforms/InstCombine/pr19420.ll


################################################################################
diff  --git a/llvm/test/Transforms/InstCombine/pr19420.ll b/llvm/test/Transforms/InstCombine/pr19420.ll
deleted file mode 100644
index aeeed2fc10e06..0000000000000
--- a/llvm/test/Transforms/InstCombine/pr19420.ll
+++ /dev/null
@@ -1,202 +0,0 @@
-; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
-; RUN: opt -S -instcombine < %s | FileCheck %s
-
-define <4 x i32> @test_FoldShiftByConstant_CreateSHL(<4 x i32> %in) {
-; CHECK-LABEL: @test_FoldShiftByConstant_CreateSHL(
-; CHECK-NEXT:    [[VSHL_N:%.*]] = mul <4 x i32> [[IN:%.*]], <i32 0, i32 -32, i32 0, i32 -32>
-; CHECK-NEXT:    ret <4 x i32> [[VSHL_N]]
-;
-  %mul.i = mul <4 x i32> %in, <i32 0, i32 -1, i32 0, i32 -1>
-  %vshl_n = shl <4 x i32> %mul.i, <i32 5, i32 5, i32 5, i32 5>
-  ret <4 x i32> %vshl_n
-}
-
-define <8 x i16> @test_FoldShiftByConstant_CreateSHL2(<8 x i16> %in) {
-; CHECK-LABEL: @test_FoldShiftByConstant_CreateSHL2(
-; CHECK-NEXT:    [[VSHL_N:%.*]] = mul <8 x i16> [[IN:%.*]], <i16 0, i16 -32, i16 0, i16 -32, i16 0, i16 -32, i16 0, i16 -32>
-; CHECK-NEXT:    ret <8 x i16> [[VSHL_N]]
-;
-  %mul.i = mul <8 x i16> %in, <i16 0, i16 -1, i16 0, i16 -1, i16 0, i16 -1, i16 0, i16 -1>
-  %vshl_n = shl <8 x i16> %mul.i, <i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5>
-  ret <8 x i16> %vshl_n
-}
-
-define <16 x i8> @test_FoldShiftByConstant_CreateAnd(<16 x i8> %in0) {
-; CHECK-LABEL: @test_FoldShiftByConstant_CreateAnd(
-; CHECK-NEXT:    [[TMP1:%.*]] = mul <16 x i8> [[IN0:%.*]], <i8 33, i8 33, i8 33, i8 33, i8 33, i8 33, i8 33, i8 33, i8 33, i8 33, i8 33, i8 33, i8 33, i8 33, i8 33, i8 33>
-; CHECK-NEXT:    [[VSHL_N:%.*]] = and <16 x i8> [[TMP1]], <i8 -32, i8 -32, i8 -32, i8 -32, i8 -32, i8 -32, i8 -32, i8 -32, i8 -32, i8 -32, i8 -32, i8 -32, i8 -32, i8 -32, i8 -32, i8 -32>
-; CHECK-NEXT:    ret <16 x i8> [[VSHL_N]]
-;
-  %vsra_n = ashr <16 x i8> %in0, <i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5>
-  %tmp = add <16 x i8> %in0, %vsra_n
-  %vshl_n = shl <16 x i8> %tmp, <i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5>
-  ret <16 x i8> %vshl_n
-}
-
-define i32 @lshr_add_shl(i32 %x, i32 %y) {
-; CHECK-LABEL: @lshr_add_shl(
-; CHECK-NEXT:    [[B1:%.*]] = shl i32 [[Y:%.*]], 4
-; CHECK-NEXT:    [[A2:%.*]] = add i32 [[B1]], [[X:%.*]]
-; CHECK-NEXT:    [[C:%.*]] = and i32 [[A2]], -16
-; CHECK-NEXT:    ret i32 [[C]]
-;
-  %a = lshr i32 %x, 4
-  %b = add i32 %a, %y
-  %c = shl i32 %b, 4
-  ret i32 %c
-}
-
-define <2 x i32> @lshr_add_shl_v2i32(<2 x i32> %x, <2 x i32> %y) {
-; CHECK-LABEL: @lshr_add_shl_v2i32(
-; CHECK-NEXT:    [[B1:%.*]] = shl <2 x i32> [[Y:%.*]], <i32 5, i32 5>
-; CHECK-NEXT:    [[A2:%.*]] = add <2 x i32> [[B1]], [[X:%.*]]
-; CHECK-NEXT:    [[C:%.*]] = and <2 x i32> [[A2]], <i32 -32, i32 -32>
-; CHECK-NEXT:    ret <2 x i32> [[C]]
-;
-  %a = lshr <2 x i32> %x, <i32 5, i32 5>
-  %b = add <2 x i32> %a, %y
-  %c = shl <2 x i32> %b, <i32 5, i32 5>
-  ret <2 x i32> %c
-}
-
-define <2 x i32> @lshr_add_shl_v2i32_undef(<2 x i32> %x, <2 x i32> %y) {
-; CHECK-LABEL: @lshr_add_shl_v2i32_undef(
-; CHECK-NEXT:    [[A:%.*]] = lshr <2 x i32> [[X:%.*]], <i32 5, i32 undef>
-; CHECK-NEXT:    [[B:%.*]] = add <2 x i32> [[A]], [[Y:%.*]]
-; CHECK-NEXT:    [[C:%.*]] = shl <2 x i32> [[B]], <i32 undef, i32 5>
-; CHECK-NEXT:    ret <2 x i32> [[C]]
-;
-  %a = lshr <2 x i32> %x, <i32 5, i32 undef>
-  %b = add <2 x i32> %a, %y
-  %c = shl <2 x i32> %b, <i32 undef, i32 5>
-  ret <2 x i32> %c
-}
-
-define <2 x i32> @lshr_add_shl_v2i32_nonuniform(<2 x i32> %x, <2 x i32> %y) {
-; CHECK-LABEL: @lshr_add_shl_v2i32_nonuniform(
-; CHECK-NEXT:    [[A:%.*]] = lshr <2 x i32> [[X:%.*]], <i32 5, i32 6>
-; CHECK-NEXT:    [[B:%.*]] = add <2 x i32> [[A]], [[Y:%.*]]
-; CHECK-NEXT:    [[C:%.*]] = shl <2 x i32> [[B]], <i32 5, i32 6>
-; CHECK-NEXT:    ret <2 x i32> [[C]]
-;
-  %a = lshr <2 x i32> %x, <i32 5, i32 6>
-  %b = add <2 x i32> %a, %y
-  %c = shl <2 x i32> %b, <i32 5, i32 6>
-  ret <2 x i32> %c
-}
-
-define i32 @lshr_add_and_shl(i32 %x, i32 %y)  {
-; CHECK-LABEL: @lshr_add_and_shl(
-; CHECK-NEXT:    [[TMP1:%.*]] = shl i32 [[Y:%.*]], 5
-; CHECK-NEXT:    [[X_MASK:%.*]] = and i32 [[X:%.*]], 4064
-; CHECK-NEXT:    [[TMP2:%.*]] = add i32 [[X_MASK]], [[TMP1]]
-; CHECK-NEXT:    ret i32 [[TMP2]]
-;
-  %1 = lshr i32 %x, 5
-  %2 = and i32 %1, 127
-  %3 = add i32 %y, %2
-  %4 = shl i32 %3, 5
-  ret i32 %4
-}
-
-define <2 x i32> @lshr_add_and_shl_v2i32(<2 x i32> %x, <2 x i32> %y)  {
-; CHECK-LABEL: @lshr_add_and_shl_v2i32(
-; CHECK-NEXT:    [[TMP1:%.*]] = shl <2 x i32> [[Y:%.*]], <i32 5, i32 5>
-; CHECK-NEXT:    [[X_MASK:%.*]] = and <2 x i32> [[X:%.*]], <i32 4064, i32 4064>
-; CHECK-NEXT:    [[TMP2:%.*]] = add <2 x i32> [[X_MASK]], [[TMP1]]
-; CHECK-NEXT:    ret <2 x i32> [[TMP2]]
-;
-  %1 = lshr <2 x i32> %x, <i32 5, i32 5>
-  %2 = and <2 x i32> %1, <i32 127, i32 127>
-  %3 = add <2 x i32> %y, %2
-  %4 = shl <2 x i32> %3, <i32 5, i32 5>
-  ret <2 x i32> %4
-}
-
-define <2 x i32> @lshr_add_and_shl_v2i32_undef(<2 x i32> %x, <2 x i32> %y)  {
-; CHECK-LABEL: @lshr_add_and_shl_v2i32_undef(
-; CHECK-NEXT:    [[TMP1:%.*]] = lshr <2 x i32> [[X:%.*]], <i32 undef, i32 5>
-; CHECK-NEXT:    [[TMP2:%.*]] = and <2 x i32> [[TMP1]], <i32 127, i32 127>
-; CHECK-NEXT:    [[TMP3:%.*]] = add <2 x i32> [[TMP2]], [[Y:%.*]]
-; CHECK-NEXT:    [[TMP4:%.*]] = shl <2 x i32> [[TMP3]], <i32 5, i32 undef>
-; CHECK-NEXT:    ret <2 x i32> [[TMP4]]
-;
-  %1 = lshr <2 x i32> %x, <i32 undef, i32 5>
-  %2 = and <2 x i32> %1, <i32 127, i32 127>
-  %3 = add <2 x i32> %y, %2
-  %4 = shl <2 x i32> %3, <i32 5, i32 undef>
-  ret <2 x i32> %4
-}
-
-define <2 x i32> @lshr_add_and_shl_v2i32_nonuniform(<2 x i32> %x, <2 x i32> %y)  {
-; CHECK-LABEL: @lshr_add_and_shl_v2i32_nonuniform(
-; CHECK-NEXT:    [[TMP1:%.*]] = lshr <2 x i32> [[X:%.*]], <i32 5, i32 6>
-; CHECK-NEXT:    [[TMP2:%.*]] = and <2 x i32> [[TMP1]], <i32 127, i32 255>
-; CHECK-NEXT:    [[TMP3:%.*]] = add <2 x i32> [[TMP2]], [[Y:%.*]]
-; CHECK-NEXT:    [[TMP4:%.*]] = shl <2 x i32> [[TMP3]], <i32 5, i32 6>
-; CHECK-NEXT:    ret <2 x i32> [[TMP4]]
-;
-  %1 = lshr <2 x i32> %x, <i32 5, i32 6>
-  %2 = and <2 x i32> %1, <i32 127, i32 255>
-  %3 = add <2 x i32> %y, %2
-  %4 = shl <2 x i32> %3, <i32 5, i32 6>
-  ret <2 x i32> %4
-}
-
-define i32 @shl_add_and_lshr(i32 %x, i32 %y) {
-; CHECK-LABEL: @shl_add_and_lshr(
-; CHECK-NEXT:    [[C1:%.*]] = shl i32 [[Y:%.*]], 4
-; CHECK-NEXT:    [[X_MASK:%.*]] = and i32 [[X:%.*]], 128
-; CHECK-NEXT:    [[D:%.*]] = add i32 [[X_MASK]], [[C1]]
-; CHECK-NEXT:    ret i32 [[D]]
-;
-  %a = lshr i32 %x, 4
-  %b = and i32 %a, 8
-  %c = add i32 %b, %y
-  %d = shl i32 %c, 4
-  ret i32 %d
-}
-
-define <2 x i32> @shl_add_and_lshr_v2i32(<2 x i32> %x, <2 x i32> %y) {
-; CHECK-LABEL: @shl_add_and_lshr_v2i32(
-; CHECK-NEXT:    [[C1:%.*]] = shl <2 x i32> [[Y:%.*]], <i32 4, i32 4>
-; CHECK-NEXT:    [[X_MASK:%.*]] = and <2 x i32> [[X:%.*]], <i32 128, i32 128>
-; CHECK-NEXT:    [[D:%.*]] = add <2 x i32> [[X_MASK]], [[C1]]
-; CHECK-NEXT:    ret <2 x i32> [[D]]
-;
-  %a = lshr <2 x i32> %x, <i32 4, i32 4>
-  %b = and <2 x i32> %a, <i32 8, i32 8>
-  %c = add <2 x i32> %b, %y
-  %d = shl <2 x i32> %c, <i32 4, i32 4>
-  ret <2 x i32> %d
-}
-
-define <2 x i32> @shl_add_and_lshr_v2i32_undef(<2 x i32> %x, <2 x i32> %y) {
-; CHECK-LABEL: @shl_add_and_lshr_v2i32_undef(
-; CHECK-NEXT:    [[A:%.*]] = lshr <2 x i32> [[X:%.*]], <i32 4, i32 undef>
-; CHECK-NEXT:    [[B:%.*]] = and <2 x i32> [[A]], <i32 8, i32 undef>
-; CHECK-NEXT:    [[C:%.*]] = add <2 x i32> [[B]], [[Y:%.*]]
-; CHECK-NEXT:    [[D:%.*]] = shl <2 x i32> [[C]], <i32 4, i32 undef>
-; CHECK-NEXT:    ret <2 x i32> [[D]]
-;
-  %a = lshr <2 x i32> %x, <i32 4, i32 undef>
-  %b = and <2 x i32> %a, <i32 8, i32 undef>
-  %c = add <2 x i32> %b, %y
-  %d = shl <2 x i32> %c, <i32 4, i32 undef>
-  ret <2 x i32> %d
-}
-
-define <2 x i32> @shl_add_and_lshr_v2i32_nonuniform(<2 x i32> %x, <2 x i32> %y) {
-; CHECK-LABEL: @shl_add_and_lshr_v2i32_nonuniform(
-; CHECK-NEXT:    [[A:%.*]] = lshr <2 x i32> [[X:%.*]], <i32 4, i32 5>
-; CHECK-NEXT:    [[B:%.*]] = and <2 x i32> [[A]], <i32 8, i32 9>
-; CHECK-NEXT:    [[C:%.*]] = add <2 x i32> [[B]], [[Y:%.*]]
-; CHECK-NEXT:    [[D:%.*]] = shl <2 x i32> [[C]], <i32 4, i32 5>
-; CHECK-NEXT:    ret <2 x i32> [[D]]
-;
-  %a = lshr <2 x i32> %x, <i32 4, i32 5>
-  %b = and <2 x i32> %a, <i32 8, i32 9>
-  %c = add <2 x i32> %b, %y
-  %d = shl <2 x i32> %c, <i32 4, i32 5>
-  ret <2 x i32> %d
-}

diff  --git a/llvm/test/Transforms/InstCombine/shift.ll b/llvm/test/Transforms/InstCombine/shift.ll
index f644ed9bb86ec..11d9be073b025 100644
--- a/llvm/test/Transforms/InstCombine/shift.ll
+++ b/llvm/test/Transforms/InstCombine/shift.ll
@@ -345,34 +345,6 @@ define i8 @test24(i8 %X) {
   ret i8 %Q
 }
 
-define i32 @test25(i32 %i.2, i32 %AA) {
-; CHECK-LABEL: @test25(
-; CHECK-NEXT:    [[I_3:%.*]] = and i32 [[I_2:%.*]], -131072
-; CHECK-NEXT:    [[X2:%.*]] = add i32 [[I_3]], [[AA:%.*]]
-; CHECK-NEXT:    [[I_6:%.*]] = and i32 [[X2]], -131072
-; CHECK-NEXT:    ret i32 [[I_6]]
-;
-  %x = lshr i32 %AA, 17
-  %i.3 = lshr i32 %i.2, 17
-  %i.5 = add i32 %i.3, %x
-  %i.6 = shl i32 %i.5, 17
-  ret i32 %i.6
-}
-
-define <2 x i32> @test25_vector(<2 x i32> %i.2, <2 x i32> %AA) {
-; CHECK-LABEL: @test25_vector(
-; CHECK-NEXT:    [[I_3:%.*]] = and <2 x i32> [[I_2:%.*]], <i32 -131072, i32 -131072>
-; CHECK-NEXT:    [[X2:%.*]] = add <2 x i32> [[I_3]], [[AA:%.*]]
-; CHECK-NEXT:    [[I_6:%.*]] = and <2 x i32> [[X2]], <i32 -131072, i32 -131072>
-; CHECK-NEXT:    ret <2 x i32> [[I_6]]
-;
-  %x = lshr <2 x i32> %AA, <i32 17, i32 17>
-  %i.3 = lshr <2 x i32> %i.2, <i32 17, i32 17>
-  %i.5 = add <2 x i32> %i.3, %x
-  %i.6 = shl <2 x i32> %i.5, <i32 17, i32 17>
-  ret <2 x i32> %i.6
-}
-
 ;; handle casts between shifts.
 define i32 @test26(i32 %A) {
 ; CHECK-LABEL: @test26(

diff  --git a/llvm/test/Transforms/InstCombine/shl-bo.ll b/llvm/test/Transforms/InstCombine/shl-bo.ll
new file mode 100644
index 0000000000000..f581ca2cb628c
--- /dev/null
+++ b/llvm/test/Transforms/InstCombine/shl-bo.ll
@@ -0,0 +1,485 @@
+; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
+; RUN: opt < %s -instcombine -S | FileCheck %s
+
+define i8 @lshr_add(i8 %a, i8 %y) {
+; CHECK-LABEL: @lshr_add(
+; CHECK-NEXT:    [[X:%.*]] = srem i8 [[A:%.*]], 42
+; CHECK-NEXT:    [[B1:%.*]] = shl i8 [[X]], 5
+; CHECK-NEXT:    [[R2:%.*]] = add i8 [[B1]], [[Y:%.*]]
+; CHECK-NEXT:    [[L:%.*]] = and i8 [[R2]], -32
+; CHECK-NEXT:    ret i8 [[L]]
+;
+  %x = srem i8 %a, 42 ; thwart complexity-based canonicalization
+  %r = lshr i8 %y, 5
+  %b = add i8 %r, %x
+  %l = shl i8 %b, 5
+  ret i8 %l
+}
+
+define <2 x i8> @lshr_add_commute_splat(<2 x i8> %a, <2 x i8> %y) {
+; CHECK-LABEL: @lshr_add_commute_splat(
+; CHECK-NEXT:    [[X:%.*]] = srem <2 x i8> [[A:%.*]], <i8 42, i8 42>
+; CHECK-NEXT:    [[B1:%.*]] = shl <2 x i8> [[X]], <i8 5, i8 5>
+; CHECK-NEXT:    [[R2:%.*]] = add <2 x i8> [[B1]], [[Y:%.*]]
+; CHECK-NEXT:    [[L:%.*]] = and <2 x i8> [[R2]], <i8 -32, i8 -32>
+; CHECK-NEXT:    ret <2 x i8> [[L]]
+;
+  %x = srem <2 x i8> %a, <i8 42, i8 42> ; thwart complexity-based canonicalization
+  %r = lshr <2 x i8> %y, <i8 5, i8 5>
+  %b = add <2 x i8> %x, %r
+  %l = shl <2 x i8> %b, <i8 5, i8 5>
+  ret <2 x i8> %l
+}
+
+define i8 @lshr_sub(i8 %a, i8 %y) {
+; CHECK-LABEL: @lshr_sub(
+; CHECK-NEXT:    [[X:%.*]] = srem i8 [[A:%.*]], 42
+; CHECK-NEXT:    [[R:%.*]] = lshr i8 [[Y:%.*]], 3
+; CHECK-NEXT:    [[B:%.*]] = sub nsw i8 [[X]], [[R]]
+; CHECK-NEXT:    [[L:%.*]] = shl i8 [[B]], 3
+; CHECK-NEXT:    ret i8 [[L]]
+;
+  %x = srem i8 %a, 42 ; thwart complexity-based canonicalization
+  %r = lshr i8 %y, 3
+  %b = sub i8 %x, %r
+  %l = shl i8 %b, 3
+  ret i8 %l
+}
+
+define <2 x i8> @lshr_sub_commute_splat(<2 x i8> %a, <2 x i8> %y) {
+; CHECK-LABEL: @lshr_sub_commute_splat(
+; CHECK-NEXT:    [[X:%.*]] = srem <2 x i8> [[A:%.*]], <i8 42, i8 42>
+; CHECK-NEXT:    [[B1_NEG:%.*]] = mul <2 x i8> [[X]], <i8 -8, i8 -8>
+; CHECK-NEXT:    [[R2:%.*]] = add <2 x i8> [[B1_NEG]], [[Y:%.*]]
+; CHECK-NEXT:    [[L:%.*]] = and <2 x i8> [[R2]], <i8 -8, i8 -8>
+; CHECK-NEXT:    ret <2 x i8> [[L]]
+;
+  %x = srem <2 x i8> %a, <i8 42, i8 42> ; thwart complexity-based canonicalization
+  %r = lshr <2 x i8> %y, <i8 3, i8 3>
+  %b = sub <2 x i8> %r, %x
+  %l = shl <2 x i8> %b, <i8 3, i8 3>
+  ret <2 x i8> %l
+}
+
+define i8 @lshr_and(i8 %a, i8 %y) {
+; CHECK-LABEL: @lshr_and(
+; CHECK-NEXT:    [[X:%.*]] = srem i8 [[A:%.*]], 42
+; CHECK-NEXT:    [[B1:%.*]] = shl i8 [[X]], 6
+; CHECK-NEXT:    [[R2:%.*]] = and i8 [[B1]], [[Y:%.*]]
+; CHECK-NEXT:    ret i8 [[R2]]
+;
+  %x = srem i8 %a, 42 ; thwart complexity-based canonicalization
+  %r = lshr i8 %y, 6
+  %b = and i8 %r, %x
+  %l = shl i8 %b, 6
+  ret i8 %l
+}
+
+define <2 x i8> @lshr_and_commute_splat(<2 x i8> %a, <2 x i8> %y) {
+; CHECK-LABEL: @lshr_and_commute_splat(
+; CHECK-NEXT:    [[X:%.*]] = srem <2 x i8> [[A:%.*]], <i8 42, i8 42>
+; CHECK-NEXT:    [[B1:%.*]] = shl <2 x i8> [[X]], <i8 6, i8 6>
+; CHECK-NEXT:    [[R2:%.*]] = and <2 x i8> [[B1]], [[Y:%.*]]
+; CHECK-NEXT:    ret <2 x i8> [[R2]]
+;
+  %x = srem <2 x i8> %a, <i8 42, i8 42> ; thwart complexity-based canonicalization
+  %r = lshr <2 x i8> %y, <i8 6, i8 6>
+  %b = and <2 x i8> %x, %r
+  %l = shl <2 x i8> %b, <i8 6, i8 6>
+  ret <2 x i8> %l
+}
+
+define i8 @lshr_or(i8 %a, i8 %y) {
+; CHECK-LABEL: @lshr_or(
+; CHECK-NEXT:    [[X:%.*]] = srem i8 [[A:%.*]], 42
+; CHECK-NEXT:    [[B1:%.*]] = shl i8 [[X]], 4
+; CHECK-NEXT:    [[Y_MASKED:%.*]] = and i8 [[Y:%.*]], -16
+; CHECK-NEXT:    [[L:%.*]] = or i8 [[B1]], [[Y_MASKED]]
+; CHECK-NEXT:    ret i8 [[L]]
+;
+  %x = srem i8 %a, 42 ; thwart complexity-based canonicalization
+  %r = lshr i8 %y, 4
+  %b = or i8 %x, %r
+  %l = shl i8 %b, 4
+  ret i8 %l
+}
+
+define <2 x i8> @lshr_or_commute_splat(<2 x i8> %a, <2 x i8> %y) {
+; CHECK-LABEL: @lshr_or_commute_splat(
+; CHECK-NEXT:    [[X:%.*]] = srem <2 x i8> [[A:%.*]], <i8 42, i8 42>
+; CHECK-NEXT:    [[B1:%.*]] = shl <2 x i8> [[X]], <i8 4, i8 4>
+; CHECK-NEXT:    [[Y_MASKED:%.*]] = and <2 x i8> [[Y:%.*]], <i8 -16, i8 -16>
+; CHECK-NEXT:    [[L:%.*]] = or <2 x i8> [[B1]], [[Y_MASKED]]
+; CHECK-NEXT:    ret <2 x i8> [[L]]
+;
+  %x = srem <2 x i8> %a, <i8 42, i8 42> ; thwart complexity-based canonicalization
+  %r = lshr <2 x i8> %y, <i8 4, i8 4>
+  %b = or <2 x i8> %r, %x
+  %l = shl <2 x i8> %b, <i8 4, i8 4>
+  ret <2 x i8> %l
+}
+
+define i8 @lshr_xor(i8 %a, i8 %y) {
+; CHECK-LABEL: @lshr_xor(
+; CHECK-NEXT:    [[X:%.*]] = srem i8 [[A:%.*]], 42
+; CHECK-NEXT:    [[B1:%.*]] = shl i8 [[X]], 3
+; CHECK-NEXT:    [[Y_MASKED:%.*]] = and i8 [[Y:%.*]], -8
+; CHECK-NEXT:    [[L:%.*]] = xor i8 [[B1]], [[Y_MASKED]]
+; CHECK-NEXT:    ret i8 [[L]]
+;
+  %x = srem i8 %a, 42 ; thwart complexity-based canonicalization
+  %r = lshr i8 %y, 3
+  %b = xor i8 %r, %x
+  %l = shl i8 %b, 3
+  ret i8 %l
+}
+
+define <2 x i8> @lshr_xor_commute_splat(<2 x i8> %a, <2 x i8> %y) {
+; CHECK-LABEL: @lshr_xor_commute_splat(
+; CHECK-NEXT:    [[X:%.*]] = srem <2 x i8> [[A:%.*]], <i8 42, i8 42>
+; CHECK-NEXT:    [[B1:%.*]] = shl <2 x i8> [[X]], <i8 3, i8 3>
+; CHECK-NEXT:    [[Y_MASKED:%.*]] = and <2 x i8> [[Y:%.*]], <i8 -8, i8 -8>
+; CHECK-NEXT:    [[L:%.*]] = xor <2 x i8> [[B1]], [[Y_MASKED]]
+; CHECK-NEXT:    ret <2 x i8> [[L]]
+;
+  %x = srem <2 x i8> %a, <i8 42, i8 42> ; thwart complexity-based canonicalization
+  %r = lshr <2 x i8> %y, <i8 3, i8 3>
+  %b = xor <2 x i8> %x, %r
+  %l = shl <2 x i8> %b, <i8 3, i8 3>
+  ret <2 x i8> %l
+}
+
+define i8 @lshr_and_add(i8 %a, i8 %y)  {
+; CHECK-LABEL: @lshr_and_add(
+; CHECK-NEXT:    [[X:%.*]] = srem i8 [[A:%.*]], 42
+; CHECK-NEXT:    [[B1:%.*]] = shl i8 [[X]], 3
+; CHECK-NEXT:    [[Y_MASK:%.*]] = and i8 [[Y:%.*]], 96
+; CHECK-NEXT:    [[L:%.*]] = add i8 [[B1]], [[Y_MASK]]
+; CHECK-NEXT:    ret i8 [[L]]
+;
+  %x = srem i8 %a, 42 ; thwart complexity-based canonicalization
+  %r = lshr i8 %y, 3
+  %m = and i8 %r, 12
+  %b = add i8 %x, %m
+  %l = shl i8 %b, 3
+  ret i8 %l
+}
+
+define <2 x i8> @lshr_and_add_commute_splat(<2 x i8> %a, <2 x i8> %y)  {
+; CHECK-LABEL: @lshr_and_add_commute_splat(
+; CHECK-NEXT:    [[X:%.*]] = srem <2 x i8> [[A:%.*]], <i8 42, i8 42>
+; CHECK-NEXT:    [[B1:%.*]] = shl <2 x i8> [[X]], <i8 3, i8 3>
+; CHECK-NEXT:    [[Y_MASK:%.*]] = and <2 x i8> [[Y:%.*]], <i8 96, i8 96>
+; CHECK-NEXT:    [[L:%.*]] = add <2 x i8> [[Y_MASK]], [[B1]]
+; CHECK-NEXT:    ret <2 x i8> [[L]]
+;
+  %x = srem <2 x i8> %a, <i8 42, i8 42> ; thwart complexity-based canonicalization
+  %r = lshr <2 x i8> %y, <i8 3, i8 3>
+  %m = and <2 x i8> %r, <i8 12, i8 12>
+  %b = add <2 x i8> %m, %x
+  %l = shl <2 x i8> %b, <i8 3, i8 3>
+  ret <2 x i8> %l
+}
+
+define i8 @lshr_and_sub(i8 %a, i8 %y)  {
+; CHECK-LABEL: @lshr_and_sub(
+; CHECK-NEXT:    [[X:%.*]] = srem i8 [[A:%.*]], 42
+; CHECK-NEXT:    [[R:%.*]] = lshr i8 [[Y:%.*]], 2
+; CHECK-NEXT:    [[M:%.*]] = and i8 [[R]], 13
+; CHECK-NEXT:    [[B:%.*]] = sub nsw i8 [[X]], [[M]]
+; CHECK-NEXT:    [[L:%.*]] = shl i8 [[B]], 2
+; CHECK-NEXT:    ret i8 [[L]]
+;
+  %x = srem i8 %a, 42 ; thwart complexity-based canonicalization
+  %r = lshr i8 %y, 2
+  %m = and i8 %r, 13
+  %b = sub i8 %x, %m
+  %l = shl i8 %b, 2
+  ret i8 %l
+}
+
+define <2 x i8> @lshr_and_sub_commute_splat(<2 x i8> %a, <2 x i8> %y)  {
+; CHECK-LABEL: @lshr_and_sub_commute_splat(
+; CHECK-NEXT:    [[X:%.*]] = srem <2 x i8> [[A:%.*]], <i8 42, i8 42>
+; CHECK-NEXT:    [[B1_NEG:%.*]] = mul <2 x i8> [[X]], <i8 -4, i8 -4>
+; CHECK-NEXT:    [[Y_MASK:%.*]] = and <2 x i8> [[Y:%.*]], <i8 52, i8 52>
+; CHECK-NEXT:    [[L:%.*]] = add <2 x i8> [[B1_NEG]], [[Y_MASK]]
+; CHECK-NEXT:    ret <2 x i8> [[L]]
+;
+  %x = srem <2 x i8> %a, <i8 42, i8 42> ; thwart complexity-based canonicalization
+  %r = lshr <2 x i8> %y, <i8 2, i8 2>
+  %m = and <2 x i8> %r, <i8 13, i8 13>
+  %b = sub <2 x i8> %m, %x
+  %l = shl <2 x i8> %b, <i8 2, i8 2>
+  ret <2 x i8> %l
+}
+
+define i8 @lshr_and_and(i8 %a, i8 %y)  {
+; CHECK-LABEL: @lshr_and_and(
+; CHECK-NEXT:    [[X:%.*]] = srem i8 [[A:%.*]], 42
+; CHECK-NEXT:    [[B1:%.*]] = shl i8 [[X]], 2
+; CHECK-NEXT:    [[Y_MASK:%.*]] = and i8 [[Y:%.*]], 52
+; CHECK-NEXT:    [[L:%.*]] = and i8 [[Y_MASK]], [[B1]]
+; CHECK-NEXT:    ret i8 [[L]]
+;
+  %x = srem i8 %a, 42 ; thwart complexity-based canonicalization
+  %r = lshr i8 %y, 2
+  %m = and i8 %r, 13
+  %b = and i8 %m, %x
+  %l = shl i8 %b, 2
+  ret i8 %l
+}
+
+define <2 x i8> @lshr_and_and_commute_splat(<2 x i8> %a, <2 x i8> %y)  {
+; CHECK-LABEL: @lshr_and_and_commute_splat(
+; CHECK-NEXT:    [[X:%.*]] = srem <2 x i8> [[A:%.*]], <i8 42, i8 42>
+; CHECK-NEXT:    [[B1:%.*]] = shl <2 x i8> [[X]], <i8 2, i8 2>
+; CHECK-NEXT:    [[Y_MASK:%.*]] = and <2 x i8> [[Y:%.*]], <i8 52, i8 52>
+; CHECK-NEXT:    [[L:%.*]] = and <2 x i8> [[B1]], [[Y_MASK]]
+; CHECK-NEXT:    ret <2 x i8> [[L]]
+;
+  %x = srem <2 x i8> %a, <i8 42, i8 42> ; thwart complexity-based canonicalization
+  %r = lshr <2 x i8> %y, <i8 2, i8 2>
+  %m = and <2 x i8> %r, <i8 13, i8 13>
+  %b = and <2 x i8> %x, %m
+  %l = shl <2 x i8> %b, <i8 2, i8 2>
+  ret <2 x i8> %l
+}
+
+define i8 @lshr_and_or(i8 %a, i8 %y)  {
+; CHECK-LABEL: @lshr_and_or(
+; CHECK-NEXT:    [[X:%.*]] = srem i8 [[A:%.*]], 42
+; CHECK-NEXT:    [[B1:%.*]] = shl i8 [[X]], 2
+; CHECK-NEXT:    [[Y_MASK:%.*]] = and i8 [[Y:%.*]], 52
+; CHECK-NEXT:    [[L:%.*]] = or i8 [[B1]], [[Y_MASK]]
+; CHECK-NEXT:    ret i8 [[L]]
+;
+  %x = srem i8 %a, 42 ; thwart complexity-based canonicalization
+  %r = lshr i8 %y, 2
+  %m = and i8 %r, 13
+  %b = or i8 %x, %m
+  %l = shl i8 %b, 2
+  ret i8 %l
+}
+
+define <2 x i8> @lshr_and_or_commute_splat(<2 x i8> %a, <2 x i8> %y)  {
+; CHECK-LABEL: @lshr_and_or_commute_splat(
+; CHECK-NEXT:    [[X:%.*]] = srem <2 x i8> [[A:%.*]], <i8 42, i8 42>
+; CHECK-NEXT:    [[B1:%.*]] = shl <2 x i8> [[X]], <i8 2, i8 2>
+; CHECK-NEXT:    [[Y_MASK:%.*]] = and <2 x i8> [[Y:%.*]], <i8 52, i8 52>
+; CHECK-NEXT:    [[L:%.*]] = or <2 x i8> [[Y_MASK]], [[B1]]
+; CHECK-NEXT:    ret <2 x i8> [[L]]
+;
+  %x = srem <2 x i8> %a, <i8 42, i8 42> ; thwart complexity-based canonicalization
+  %r = lshr <2 x i8> %y, <i8 2, i8 2>
+  %m = and <2 x i8> %r, <i8 13, i8 13>
+  %b = or <2 x i8> %m, %x
+  %l = shl <2 x i8> %b, <i8 2, i8 2>
+  ret <2 x i8> %l
+}
+
+define i8 @lshr_and_xor(i8 %a, i8 %y)  {
+; CHECK-LABEL: @lshr_and_xor(
+; CHECK-NEXT:    [[X:%.*]] = srem i8 [[A:%.*]], 42
+; CHECK-NEXT:    [[B1:%.*]] = shl i8 [[X]], 2
+; CHECK-NEXT:    [[Y_MASK:%.*]] = and i8 [[Y:%.*]], 52
+; CHECK-NEXT:    [[L:%.*]] = xor i8 [[Y_MASK]], [[B1]]
+; CHECK-NEXT:    ret i8 [[L]]
+;
+  %x = srem i8 %a, 42 ; thwart complexity-based canonicalization
+  %r = lshr i8 %y, 2
+  %m = and i8 %r, 13
+  %b = xor i8 %m, %x
+  %l = shl i8 %b, 2
+  ret i8 %l
+}
+
+define <2 x i8> @lshr_and_xor_commute_splat(<2 x i8> %a, <2 x i8> %y)  {
+; CHECK-LABEL: @lshr_and_xor_commute_splat(
+; CHECK-NEXT:    [[X:%.*]] = srem <2 x i8> [[A:%.*]], <i8 42, i8 42>
+; CHECK-NEXT:    [[B1:%.*]] = shl <2 x i8> [[X]], <i8 2, i8 2>
+; CHECK-NEXT:    [[Y_MASK:%.*]] = and <2 x i8> [[Y:%.*]], <i8 52, i8 52>
+; CHECK-NEXT:    [[L:%.*]] = xor <2 x i8> [[B1]], [[Y_MASK]]
+; CHECK-NEXT:    ret <2 x i8> [[L]]
+;
+  %x = srem <2 x i8> %a, <i8 42, i8 42> ; thwart complexity-based canonicalization
+  %r = lshr <2 x i8> %y, <i8 2, i8 2>
+  %m = and <2 x i8> %r, <i8 13, i8 13>
+  %b = xor <2 x i8> %x, %m
+  %l = shl <2 x i8> %b, <i8 2, i8 2>
+  ret <2 x i8> %l
+}
+
+define <2 x i8> @lshr_add_shl_v2i8_undef(<2 x i8> %x, <2 x i8> %y) {
+; CHECK-LABEL: @lshr_add_shl_v2i8_undef(
+; CHECK-NEXT:    [[A:%.*]] = lshr <2 x i8> [[Y:%.*]], <i8 5, i8 undef>
+; CHECK-NEXT:    [[B:%.*]] = add <2 x i8> [[A]], [[X:%.*]]
+; CHECK-NEXT:    [[C:%.*]] = shl <2 x i8> [[B]], <i8 undef, i8 5>
+; CHECK-NEXT:    ret <2 x i8> [[C]]
+;
+  %a = lshr <2 x i8> %y, <i8 5, i8 undef>
+  %b = add <2 x i8> %a, %x
+  %c = shl <2 x i8> %b, <i8 undef, i8 5>
+  ret <2 x i8> %c
+}
+
+define <2 x i8> @lshr_add_shl_v2i8_nonuniform(<2 x i8> %x, <2 x i8> %y) {
+; CHECK-LABEL: @lshr_add_shl_v2i8_nonuniform(
+; CHECK-NEXT:    [[A:%.*]] = lshr <2 x i8> [[Y:%.*]], <i8 5, i8 6>
+; CHECK-NEXT:    [[B:%.*]] = add <2 x i8> [[A]], [[X:%.*]]
+; CHECK-NEXT:    [[C:%.*]] = shl <2 x i8> [[B]], <i8 5, i8 6>
+; CHECK-NEXT:    ret <2 x i8> [[C]]
+;
+  %a = lshr <2 x i8> %y, <i8 5, i8 6>
+  %b = add <2 x i8> %a, %x
+  %c = shl <2 x i8> %b, <i8 5, i8 6>
+  ret <2 x i8> %c
+}
+
+define i32 @lshr_add_and_shl(i32 %x, i32 %y)  {
+; CHECK-LABEL: @lshr_add_and_shl(
+; CHECK-NEXT:    [[TMP1:%.*]] = shl i32 [[Y:%.*]], 5
+; CHECK-NEXT:    [[X_MASK:%.*]] = and i32 [[X:%.*]], 4064
+; CHECK-NEXT:    [[TMP2:%.*]] = add i32 [[X_MASK]], [[TMP1]]
+; CHECK-NEXT:    ret i32 [[TMP2]]
+;
+  %1 = lshr i32 %x, 5
+  %2 = and i32 %1, 127
+  %3 = add i32 %y, %2
+  %4 = shl i32 %3, 5
+  ret i32 %4
+}
+
+define <2 x i32> @lshr_add_and_shl_v2i32(<2 x i32> %x, <2 x i32> %y)  {
+; CHECK-LABEL: @lshr_add_and_shl_v2i32(
+; CHECK-NEXT:    [[TMP1:%.*]] = shl <2 x i32> [[Y:%.*]], <i32 5, i32 5>
+; CHECK-NEXT:    [[X_MASK:%.*]] = and <2 x i32> [[X:%.*]], <i32 4064, i32 4064>
+; CHECK-NEXT:    [[TMP2:%.*]] = add <2 x i32> [[X_MASK]], [[TMP1]]
+; CHECK-NEXT:    ret <2 x i32> [[TMP2]]
+;
+  %1 = lshr <2 x i32> %x, <i32 5, i32 5>
+  %2 = and <2 x i32> %1, <i32 127, i32 127>
+  %3 = add <2 x i32> %y, %2
+  %4 = shl <2 x i32> %3, <i32 5, i32 5>
+  ret <2 x i32> %4
+}
+
+define <2 x i32> @lshr_add_and_shl_v2i32_undef(<2 x i32> %x, <2 x i32> %y)  {
+; CHECK-LABEL: @lshr_add_and_shl_v2i32_undef(
+; CHECK-NEXT:    [[TMP1:%.*]] = lshr <2 x i32> [[X:%.*]], <i32 undef, i32 5>
+; CHECK-NEXT:    [[TMP2:%.*]] = and <2 x i32> [[TMP1]], <i32 127, i32 127>
+; CHECK-NEXT:    [[TMP3:%.*]] = add <2 x i32> [[TMP2]], [[Y:%.*]]
+; CHECK-NEXT:    [[TMP4:%.*]] = shl <2 x i32> [[TMP3]], <i32 5, i32 undef>
+; CHECK-NEXT:    ret <2 x i32> [[TMP4]]
+;
+  %1 = lshr <2 x i32> %x, <i32 undef, i32 5>
+  %2 = and <2 x i32> %1, <i32 127, i32 127>
+  %3 = add <2 x i32> %y, %2
+  %4 = shl <2 x i32> %3, <i32 5, i32 undef>
+  ret <2 x i32> %4
+}
+
+define <2 x i32> @lshr_add_and_shl_v2i32_nonuniform(<2 x i32> %x, <2 x i32> %y)  {
+; CHECK-LABEL: @lshr_add_and_shl_v2i32_nonuniform(
+; CHECK-NEXT:    [[TMP1:%.*]] = lshr <2 x i32> [[X:%.*]], <i32 5, i32 6>
+; CHECK-NEXT:    [[TMP2:%.*]] = and <2 x i32> [[TMP1]], <i32 127, i32 255>
+; CHECK-NEXT:    [[TMP3:%.*]] = add <2 x i32> [[TMP2]], [[Y:%.*]]
+; CHECK-NEXT:    [[TMP4:%.*]] = shl <2 x i32> [[TMP3]], <i32 5, i32 6>
+; CHECK-NEXT:    ret <2 x i32> [[TMP4]]
+;
+  %1 = lshr <2 x i32> %x, <i32 5, i32 6>
+  %2 = and <2 x i32> %1, <i32 127, i32 255>
+  %3 = add <2 x i32> %y, %2
+  %4 = shl <2 x i32> %3, <i32 5, i32 6>
+  ret <2 x i32> %4
+}
+
+define i32 @shl_add_and_lshr(i32 %x, i32 %y) {
+; CHECK-LABEL: @shl_add_and_lshr(
+; CHECK-NEXT:    [[C1:%.*]] = shl i32 [[Y:%.*]], 4
+; CHECK-NEXT:    [[X_MASK:%.*]] = and i32 [[X:%.*]], 128
+; CHECK-NEXT:    [[D:%.*]] = add i32 [[X_MASK]], [[C1]]
+; CHECK-NEXT:    ret i32 [[D]]
+;
+  %a = lshr i32 %x, 4
+  %b = and i32 %a, 8
+  %c = add i32 %b, %y
+  %d = shl i32 %c, 4
+  ret i32 %d
+}
+
+define <2 x i32> @shl_add_and_lshr_v2i32(<2 x i32> %x, <2 x i32> %y) {
+; CHECK-LABEL: @shl_add_and_lshr_v2i32(
+; CHECK-NEXT:    [[C1:%.*]] = shl <2 x i32> [[Y:%.*]], <i32 4, i32 4>
+; CHECK-NEXT:    [[X_MASK:%.*]] = and <2 x i32> [[X:%.*]], <i32 128, i32 128>
+; CHECK-NEXT:    [[D:%.*]] = add <2 x i32> [[X_MASK]], [[C1]]
+; CHECK-NEXT:    ret <2 x i32> [[D]]
+;
+  %a = lshr <2 x i32> %x, <i32 4, i32 4>
+  %b = and <2 x i32> %a, <i32 8, i32 8>
+  %c = add <2 x i32> %b, %y
+  %d = shl <2 x i32> %c, <i32 4, i32 4>
+  ret <2 x i32> %d
+}
+
+define <2 x i32> @shl_add_and_lshr_v2i32_undef(<2 x i32> %x, <2 x i32> %y) {
+; CHECK-LABEL: @shl_add_and_lshr_v2i32_undef(
+; CHECK-NEXT:    [[A:%.*]] = lshr <2 x i32> [[X:%.*]], <i32 4, i32 undef>
+; CHECK-NEXT:    [[B:%.*]] = and <2 x i32> [[A]], <i32 8, i32 undef>
+; CHECK-NEXT:    [[C:%.*]] = add <2 x i32> [[B]], [[Y:%.*]]
+; CHECK-NEXT:    [[D:%.*]] = shl <2 x i32> [[C]], <i32 4, i32 undef>
+; CHECK-NEXT:    ret <2 x i32> [[D]]
+;
+  %a = lshr <2 x i32> %x, <i32 4, i32 undef>
+  %b = and <2 x i32> %a, <i32 8, i32 undef>
+  %c = add <2 x i32> %b, %y
+  %d = shl <2 x i32> %c, <i32 4, i32 undef>
+  ret <2 x i32> %d
+}
+
+define <2 x i32> @shl_add_and_lshr_v2i32_nonuniform(<2 x i32> %x, <2 x i32> %y) {
+; CHECK-LABEL: @shl_add_and_lshr_v2i32_nonuniform(
+; CHECK-NEXT:    [[A:%.*]] = lshr <2 x i32> [[X:%.*]], <i32 4, i32 5>
+; CHECK-NEXT:    [[B:%.*]] = and <2 x i32> [[A]], <i32 8, i32 9>
+; CHECK-NEXT:    [[C:%.*]] = add <2 x i32> [[B]], [[Y:%.*]]
+; CHECK-NEXT:    [[D:%.*]] = shl <2 x i32> [[C]], <i32 4, i32 5>
+; CHECK-NEXT:    ret <2 x i32> [[D]]
+;
+  %a = lshr <2 x i32> %x, <i32 4, i32 5>
+  %b = and <2 x i32> %a, <i32 8, i32 9>
+  %c = add <2 x i32> %b, %y
+  %d = shl <2 x i32> %c, <i32 4, i32 5>
+  ret <2 x i32> %d
+}
+
+define <4 x i32> @test_FoldShiftByConstant_CreateSHL(<4 x i32> %in) {
+; CHECK-LABEL: @test_FoldShiftByConstant_CreateSHL(
+; CHECK-NEXT:    [[VSHL_N:%.*]] = mul <4 x i32> [[IN:%.*]], <i32 0, i32 -32, i32 0, i32 -32>
+; CHECK-NEXT:    ret <4 x i32> [[VSHL_N]]
+;
+  %mul.i = mul <4 x i32> %in, <i32 0, i32 -1, i32 0, i32 -1>
+  %vshl_n = shl <4 x i32> %mul.i, <i32 5, i32 5, i32 5, i32 5>
+  ret <4 x i32> %vshl_n
+}
+
+define <8 x i16> @test_FoldShiftByConstant_CreateSHL2(<8 x i16> %in) {
+; CHECK-LABEL: @test_FoldShiftByConstant_CreateSHL2(
+; CHECK-NEXT:    [[VSHL_N:%.*]] = mul <8 x i16> [[IN:%.*]], <i16 0, i16 -32, i16 0, i16 -32, i16 0, i16 -32, i16 0, i16 -32>
+; CHECK-NEXT:    ret <8 x i16> [[VSHL_N]]
+;
+  %mul.i = mul <8 x i16> %in, <i16 0, i16 -1, i16 0, i16 -1, i16 0, i16 -1, i16 0, i16 -1>
+  %vshl_n = shl <8 x i16> %mul.i, <i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5>
+  ret <8 x i16> %vshl_n
+}
+
+define <16 x i8> @test_FoldShiftByConstant_CreateAnd(<16 x i8> %in0) {
+; CHECK-LABEL: @test_FoldShiftByConstant_CreateAnd(
+; CHECK-NEXT:    [[TMP1:%.*]] = mul <16 x i8> [[IN0:%.*]], <i8 33, i8 33, i8 33, i8 33, i8 33, i8 33, i8 33, i8 33, i8 33, i8 33, i8 33, i8 33, i8 33, i8 33, i8 33, i8 33>
+; CHECK-NEXT:    [[VSHL_N:%.*]] = and <16 x i8> [[TMP1]], <i8 -32, i8 -32, i8 -32, i8 -32, i8 -32, i8 -32, i8 -32, i8 -32, i8 -32, i8 -32, i8 -32, i8 -32, i8 -32, i8 -32, i8 -32, i8 -32>
+; CHECK-NEXT:    ret <16 x i8> [[VSHL_N]]
+;
+  %vsra_n = ashr <16 x i8> %in0, <i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5>
+  %tmp = add <16 x i8> %in0, %vsra_n
+  %vshl_n = shl <16 x i8> %tmp, <i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5>
+  ret <16 x i8> %vshl_n
+}


        


More information about the llvm-commits mailing list