[llvm] 0f14b2e - Revert "[BPI] Improve static heuristics for integer comparisons"
Dávid Bolvanský via llvm-commits
llvm-commits at lists.llvm.org
Mon Aug 17 11:45:10 PDT 2020
Author: Dávid Bolvanský
Date: 2020-08-17T20:44:33+02:00
New Revision: 0f14b2e6cbb54c84ed3b00b0db521f5ce2d1e3f2
URL: https://github.com/llvm/llvm-project/commit/0f14b2e6cbb54c84ed3b00b0db521f5ce2d1e3f2
DIFF: https://github.com/llvm/llvm-project/commit/0f14b2e6cbb54c84ed3b00b0db521f5ce2d1e3f2.diff
LOG: Revert "[BPI] Improve static heuristics for integer comparisons"
This reverts commit 50c743fa713002fe4e0c76d23043e6c1f9e9fe6f. Patch will be split to smaller ones.
Added:
llvm/test/Analysis/BranchProbabilityInfo/zero_heuristics.ll
Modified:
compiler-rt/test/profile/Linux/counter_promo_for.c
compiler-rt/test/profile/Linux/counter_promo_while.c
llvm/include/llvm/Analysis/BranchProbabilityInfo.h
llvm/lib/Analysis/BranchProbabilityInfo.cpp
llvm/test/Analysis/BranchProbabilityInfo/loop.ll
llvm/test/CodeGen/AArch64/aarch64-gep-opt.ll
llvm/test/CodeGen/AArch64/branch-relax-alignment.ll
llvm/test/CodeGen/AArch64/combine-comparisons-by-cse.ll
llvm/test/CodeGen/AArch64/cond-br-tuning.ll
llvm/test/CodeGen/AArch64/fast-isel-cmp-branch.ll
llvm/test/CodeGen/ARM/2011-12-14-machine-sink.ll
llvm/test/CodeGen/ARM/cmpxchg-weak.ll
llvm/test/CodeGen/ARM/lsr-unfolded-offset.ll
llvm/test/CodeGen/ARM/machine-cse-cmp.ll
llvm/test/CodeGen/Hexagon/newvaluejump2.ll
llvm/test/CodeGen/Mips/brcongt.ll
llvm/test/CodeGen/Mips/brconlt.ll
llvm/test/CodeGen/Mips/brconne.ll
llvm/test/CodeGen/Mips/compactbranches/no-beqzc-bnezc.ll
llvm/test/CodeGen/Mips/lcb2.ll
llvm/test/CodeGen/Mips/lcb5.ll
llvm/test/CodeGen/Mips/longbranch/compact-branches-long-branch.ll
llvm/test/CodeGen/Mips/seleq.ll
llvm/test/CodeGen/Mips/selle.ll
llvm/test/CodeGen/PowerPC/brcond.ll
llvm/test/CodeGen/PowerPC/memCmpUsedInZeroEqualityComparison.ll
llvm/test/CodeGen/PowerPC/redundant-copy-after-tail-dup.ll
llvm/test/CodeGen/RISCV/branch.ll
llvm/test/CodeGen/RISCV/rv64m-w-insts-legalization.ll
llvm/test/CodeGen/SystemZ/int-cmp-37.ll
llvm/test/CodeGen/SystemZ/int-cmp-40.ll
llvm/test/CodeGen/Thumb2/LowOverheadLoops/mve-float-loops.ll
llvm/test/CodeGen/Thumb2/LowOverheadLoops/varying-outer-2d-reduction.ll
llvm/test/CodeGen/Thumb2/mve-postinc-distribute.ll
llvm/test/CodeGen/Thumb2/mve-postinc-lsr.ll
llvm/test/CodeGen/Thumb2/mve-satmul-loops.ll
llvm/test/CodeGen/Thumb2/thumb2-branch.ll
llvm/test/CodeGen/X86/3addr-16bit.ll
llvm/test/CodeGen/X86/absolute-cmp.ll
llvm/test/CodeGen/X86/atomic-flags.ll
llvm/test/CodeGen/X86/atomic-unordered.ll
llvm/test/CodeGen/X86/bmi.ll
llvm/test/CodeGen/X86/bt.ll
llvm/test/CodeGen/X86/conditional-tailcall-pgso.ll
llvm/test/CodeGen/X86/conditional-tailcall.ll
llvm/test/CodeGen/X86/fast-isel-cmp-branch2.ll
llvm/test/CodeGen/X86/funnel-shift.ll
llvm/test/CodeGen/X86/indirect-branch-tracking-eh2.ll
llvm/test/CodeGen/X86/jump_sign.ll
llvm/test/CodeGen/X86/lsr-negative-stride.ll
llvm/test/CodeGen/X86/machine-cse.ll
llvm/test/CodeGen/X86/memcmp-more-load-pairs.ll
llvm/test/CodeGen/X86/memcmp-optsize.ll
llvm/test/CodeGen/X86/memcmp-pgso.ll
llvm/test/CodeGen/X86/memcmp.ll
llvm/test/CodeGen/X86/neg_cmp.ll
llvm/test/CodeGen/X86/nobt.ll
llvm/test/CodeGen/X86/pr29170.ll
llvm/test/CodeGen/X86/wide-integer-cmp.ll
llvm/test/DebugInfo/COFF/fpo-shrink-wrap.ll
llvm/test/Transforms/PGOProfile/counter_promo_mexits.ll
llvm/test/Transforms/PGOProfile/landingpad.ll
Removed:
llvm/test/Analysis/BranchProbabilityInfo/integer_heuristics.ll
################################################################################
diff --git a/compiler-rt/test/profile/Linux/counter_promo_for.c b/compiler-rt/test/profile/Linux/counter_promo_for.c
index 464c97cb7dd3..0efebdc95271 100644
--- a/compiler-rt/test/profile/Linux/counter_promo_for.c
+++ b/compiler-rt/test/profile/Linux/counter_promo_for.c
@@ -22,23 +22,23 @@ __attribute__((noinline)) void foo(int n, int N) {
// PROMO: load{{.*}}@__profc_foo{{.*}} 3){{.*}}
// PROMO-NEXT: add
// PROMO-NEXT: store{{.*}}@__profc_foo{{.*}} 3){{.*}}
-// PROMO: load{{.*}}@__profc_foo{{.*}} 1){{.*}}
-// PROMO-NEXT: add
-// PROMO-NEXT: store{{.*}}@__profc_foo{{.*}} 1){{.*}}
-// PROMO-NEXT: load{{.*}}@__profc_foo{{.*}} 0){{.*}}
+// PROMO: load{{.*}}@__profc_foo{{.*}} 0){{.*}}
// PROMO-NEXT: add
// PROMO-NEXT: store{{.*}}@__profc_foo{{.*}} 0){{.*}}
+// PROMO-NEXT: load{{.*}}@__profc_foo{{.*}} 1){{.*}}
+// PROMO-NEXT: add
+// PROMO-NEXT: store{{.*}}@__profc_foo{{.*}} 1){{.*}}
// PROMO: load{{.*}}@__profc_foo{{.*}} 2){{.*}}
// PROMO-NEXT: add
// PROMO-NEXT: store{{.*}}@__profc_foo{{.*}} 2){{.*}}
//
// NOPROMO-LABEL: @foo
-// NOPROMO: load{{.*}}@__profc_foo{{.*}} 1){{.*}}
-// NOPROMO-NEXT: add
-// NOPROMO-NEXT: store{{.*}}@__profc_foo{{.*}} 1){{.*}}
// NOPROMO: load{{.*}}@__profc_foo{{.*}} 0){{.*}}
// NOPROMO-NEXT: add
// NOPROMO-NEXT: store{{.*}}@__profc_foo{{.*}} 0){{.*}}
+// NOPROMO: load{{.*}}@__profc_foo{{.*}} 1){{.*}}
+// NOPROMO-NEXT: add
+// NOPROMO-NEXT: store{{.*}}@__profc_foo{{.*}} 1){{.*}}
// NOPROMO: load{{.*}}@__profc_foo{{.*}} 2){{.*}}
// NOPROMO-NEXT: add
// NOPROMO-NEXT: store{{.*}}@__profc_foo{{.*}} 2){{.*}}
diff --git a/compiler-rt/test/profile/Linux/counter_promo_while.c b/compiler-rt/test/profile/Linux/counter_promo_while.c
index fdd77982c2bf..183ef8543aff 100644
--- a/compiler-rt/test/profile/Linux/counter_promo_while.c
+++ b/compiler-rt/test/profile/Linux/counter_promo_while.c
@@ -17,23 +17,23 @@ int g;
__attribute__((noinline)) void bar(int i) { g += i; }
__attribute__((noinline)) void foo(int n, int N) {
// PROMO-LABEL: @foo
-// PROMO: load{{.*}}@__profc_foo{{.*}} 1){{.*}}
-// PROMO-NEXT: add
-// PROMO-NEXT: store{{.*}}@__profc_foo{{.*}} 1){{.*}}
-// PROMO-NEXT: load{{.*}}@__profc_foo{{.*}} 0){{.*}}
+// PROMO: load{{.*}}@__profc_foo{{.*}} 0){{.*}}
// PROMO-NEXT: add
// PROMO-NEXT: store{{.*}}@__profc_foo{{.*}} 0){{.*}}
+// PROMO-NEXT: load{{.*}}@__profc_foo{{.*}} 1){{.*}}
+// PROMO-NEXT: add
+// PROMO-NEXT: store{{.*}}@__profc_foo{{.*}} 1){{.*}}
// PROMO-NEXT: load{{.*}}@__profc_foo{{.*}} 2){{.*}}
// PROMO-NEXT: add
// PROMO-NEXT: store{{.*}}@__profc_foo{{.*}} 2){{.*}}
//
// NOPROMO-LABEL: @foo
-// NOPROMO: load{{.*}}@__profc_foo{{.*}} 1){{.*}}
-// NOPROMO-NEXT: add
-// NOPROMO-NEXT: store{{.*}}@__profc_foo{{.*}} 1){{.*}}
// NOPROMO: load{{.*}}@__profc_foo{{.*}} 0){{.*}}
// NOPROMO-NEXT: add
// NOPROMO-NEXT: store{{.*}}@__profc_foo{{.*}} 0){{.*}}
+// NOPROMO: load{{.*}}@__profc_foo{{.*}} 1){{.*}}
+// NOPROMO-NEXT: add
+// NOPROMO-NEXT: store{{.*}}@__profc_foo{{.*}} 1){{.*}}
// NOPROMO: load{{.*}}@__profc_foo{{.*}} 2){{.*}}
// NOPROMO-NEXT: add
// NOPROMO-NEXT: store{{.*}}@__profc_foo{{.*}} 2){{.*}}
diff --git a/llvm/include/llvm/Analysis/BranchProbabilityInfo.h b/llvm/include/llvm/Analysis/BranchProbabilityInfo.h
index 2c736fe9c1f0..447f14501cb6 100644
--- a/llvm/include/llvm/Analysis/BranchProbabilityInfo.h
+++ b/llvm/include/llvm/Analysis/BranchProbabilityInfo.h
@@ -315,7 +315,7 @@ class BranchProbabilityInfo {
bool calcColdCallHeuristics(const BasicBlock *BB);
bool calcPointerHeuristics(const BasicBlock *BB);
bool calcLoopBranchHeuristics(const BasicBlock *BB, const LoopInfo &LI);
- bool calcIntegerHeuristics(const BasicBlock *BB, const TargetLibraryInfo *TLI);
+ bool calcZeroHeuristics(const BasicBlock *BB, const TargetLibraryInfo *TLI);
bool calcFloatingPointHeuristics(const BasicBlock *BB);
bool calcInvokeHeuristics(const BasicBlock *BB);
};
diff --git a/llvm/lib/Analysis/BranchProbabilityInfo.cpp b/llvm/lib/Analysis/BranchProbabilityInfo.cpp
index 6df90e66b84b..eae2c4ea9da8 100644
--- a/llvm/lib/Analysis/BranchProbabilityInfo.cpp
+++ b/llvm/lib/Analysis/BranchProbabilityInfo.cpp
@@ -122,8 +122,8 @@ static const uint32_t CC_NONTAKEN_WEIGHT = 64;
static const uint32_t PH_TAKEN_WEIGHT = 20;
static const uint32_t PH_NONTAKEN_WEIGHT = 12;
-static const uint32_t INTH_TAKEN_WEIGHT = 20;
-static const uint32_t INTH_NONTAKEN_WEIGHT = 12;
+static const uint32_t ZH_TAKEN_WEIGHT = 20;
+static const uint32_t ZH_NONTAKEN_WEIGHT = 12;
static const uint32_t FPH_TAKEN_WEIGHT = 20;
static const uint32_t FPH_NONTAKEN_WEIGHT = 12;
@@ -856,7 +856,7 @@ bool BranchProbabilityInfo::calcLoopBranchHeuristics(const BasicBlock *BB,
return true;
}
-bool BranchProbabilityInfo::calcIntegerHeuristics(const BasicBlock *BB,
+bool BranchProbabilityInfo::calcZeroHeuristics(const BasicBlock *BB,
const TargetLibraryInfo *TLI) {
const BranchInst *BI = dyn_cast<BranchInst>(BB->getTerminator());
if (!BI || !BI->isConditional())
@@ -873,21 +873,10 @@ bool BranchProbabilityInfo::calcIntegerHeuristics(const BasicBlock *BB,
return dyn_cast<ConstantInt>(V);
};
- BranchProbability TakenProb(INTH_TAKEN_WEIGHT,
- INTH_TAKEN_WEIGHT + INTH_NONTAKEN_WEIGHT);
- BranchProbability UntakenProb(INTH_NONTAKEN_WEIGHT,
- INTH_TAKEN_WEIGHT + INTH_NONTAKEN_WEIGHT);
Value *RHS = CI->getOperand(1);
ConstantInt *CV = GetConstantInt(RHS);
- if (!CV) {
- // X == Y -> Unlikely
- // Otherwise -> Likely
- if (CI->isTrueWhenEqual())
- std::swap(TakenProb, UntakenProb);
- setEdgeProbability(
- BB, SmallVector<BranchProbability, 2>({TakenProb, UntakenProb}));
- return true;
- }
+ if (!CV)
+ return false;
// If the LHS is the result of AND'ing a value with a single bit bitmask,
// we don't have information about probabilities.
@@ -975,6 +964,10 @@ bool BranchProbabilityInfo::calcIntegerHeuristics(const BasicBlock *BB,
return false;
}
+ BranchProbability TakenProb(ZH_TAKEN_WEIGHT,
+ ZH_TAKEN_WEIGHT + ZH_NONTAKEN_WEIGHT);
+ BranchProbability UntakenProb(ZH_NONTAKEN_WEIGHT,
+ ZH_TAKEN_WEIGHT + ZH_NONTAKEN_WEIGHT);
if (!isProb)
std::swap(TakenProb, UntakenProb);
@@ -1228,7 +1221,7 @@ void BranchProbabilityInfo::calculate(const Function &F, const LoopInfo &LI,
continue;
if (calcPointerHeuristics(BB))
continue;
- if (calcIntegerHeuristics(BB, TLI))
+ if (calcZeroHeuristics(BB, TLI))
continue;
if (calcFloatingPointHeuristics(BB))
continue;
diff --git a/llvm/test/Analysis/BranchProbabilityInfo/loop.ll b/llvm/test/Analysis/BranchProbabilityInfo/loop.ll
index fde631a273c5..63377e3ba955 100644
--- a/llvm/test/Analysis/BranchProbabilityInfo/loop.ll
+++ b/llvm/test/Analysis/BranchProbabilityInfo/loop.ll
@@ -263,8 +263,8 @@ for.body:
%0 = load i32, i32* %c, align 4
%cmp1 = icmp eq i32 %0, %i.011
br i1 %cmp1, label %for.inc5, label %if.end
-; CHECK: edge for.body -> for.inc5 probability is 0x30000000 / 0x80000000 = 37.50%
-; CHECK: edge for.body -> if.end probability is 0x50000000 / 0x80000000 = 62.50%
+; CHECK: edge for.body -> for.inc5 probability is 0x40000000 / 0x80000000 = 50.00%
+; CHECK: edge for.body -> if.end probability is 0x40000000 / 0x80000000 = 50.00%
if.end:
call void @g1()
@@ -324,22 +324,22 @@ for.body3:
%0 = load i32, i32* %c, align 4
%cmp4 = icmp eq i32 %0, %j.017
br i1 %cmp4, label %for.inc, label %if.end
-; CHECK: edge for.body3 -> for.inc probability is 0x30000000 / 0x80000000 = 37.50%
-; CHECK: edge for.body3 -> if.end probability is 0x50000000 / 0x80000000 = 62.50%
+; CHECK: edge for.body3 -> for.inc probability is 0x40000000 / 0x80000000 = 50.00%
+; CHECK: edge for.body3 -> if.end probability is 0x40000000 / 0x80000000 = 50.00%
if.end:
%1 = load i32, i32* %arrayidx5, align 4
%cmp6 = icmp eq i32 %1, %j.017
br i1 %cmp6, label %for.inc, label %if.end8
-; CHECK: edge if.end -> for.inc probability is 0x30000000 / 0x80000000 = 37.50%
-; CHECK: edge if.end -> if.end8 probability is 0x50000000 / 0x80000000 = 62.50%
+; CHECK: edge if.end -> for.inc probability is 0x40000000 / 0x80000000 = 50.00%
+; CHECK: edge if.end -> if.end8 probability is 0x40000000 / 0x80000000 = 50.00%
if.end8:
%2 = load i32, i32* %arrayidx9, align 4
%cmp10 = icmp eq i32 %2, %j.017
br i1 %cmp10, label %for.inc, label %if.end12
-; CHECK: edge if.end8 -> for.inc probability is 0x30000000 / 0x80000000 = 37.50%
-; CHECK: edge if.end8 -> if.end12 probability is 0x50000000 / 0x80000000 = 62.50%
+; CHECK: edge if.end8 -> for.inc probability is 0x40000000 / 0x80000000 = 50.00%
+; CHECK: edge if.end8 -> if.end12 probability is 0x40000000 / 0x80000000 = 50.00%
if.end12:
call void @g2()
diff --git a/llvm/test/Analysis/BranchProbabilityInfo/integer_heuristics.ll b/llvm/test/Analysis/BranchProbabilityInfo/zero_heuristics.ll
similarity index 67%
rename from llvm/test/Analysis/BranchProbabilityInfo/integer_heuristics.ll
rename to llvm/test/Analysis/BranchProbabilityInfo/zero_heuristics.ll
index c1d894a712fc..c6e1cb8c265a 100644
--- a/llvm/test/Analysis/BranchProbabilityInfo/integer_heuristics.ll
+++ b/llvm/test/Analysis/BranchProbabilityInfo/zero_heuristics.ll
@@ -101,54 +101,3 @@ for.inc:
exit:
ret void
}
-
-declare void @foo()
-
-; CHECK-LABEL: foo1
-define i32 @foo1(i32 %x, i32 %y, i8 signext %z, i8 signext %w) {
-entry:
- %c = icmp eq i32 %x, %y
- br i1 %c, label %then, label %else
-; CHECK: edge entry -> then probability is 0x30000000 / 0x80000000 = 37.50%
-; CHECK: edge entry -> else probability is 0x50000000 / 0x80000000 = 62.50%
-then:
- tail call void @foo()
- br label %else
-; CHECK: edge then -> else probability is 0x80000000 / 0x80000000 = 100.00% [HOT edge]
-else:
- %v = phi i8 [ %z, %then ], [ %w, %entry ]
- %r = sext i8 %v to i32
- ret i32 %r
-}
-
-; CHECK-LABEL: foo2
-define i32 @foo2(i32 %x, i32 %y, i8 signext %z, i8 signext %w) {
-entry:
- %c = icmp ne i32 %x, %y
- br i1 %c, label %then, label %else
-; CHECK: edge entry -> then probability is 0x50000000 / 0x80000000 = 62.50%
-; CHECK: edge entry -> else probability is 0x30000000 / 0x80000000 = 37.50%
-then:
- br label %else
-; CHECK: edge then -> else probability is 0x80000000 / 0x80000000 = 100.00% [HOT edge]
-else:
- %v = phi i8 [ %z, %then ], [ %w, %entry ]
- %r = sext i8 %v to i32
- ret i32 %r
-}
-
-; CHECK-LABEL: foo3
-define i32 @foo3(i32 %x, i32 %y, i8 signext %z, i8 signext %w) {
-entry:
- %c = icmp ult i32 %x, %y
- br i1 %c, label %then, label %else
-; CHECK: edge entry -> then probability is 0x50000000 / 0x80000000 = 62.50%
-; CHECK: edge entry -> else probability is 0x30000000 / 0x80000000 = 37.50%
-then:
- br label %else
-; CHECK: edge then -> else probability is 0x80000000 / 0x80000000 = 100.00% [HOT edge]
-else:
- %v = phi i8 [ %z, %then ], [ %w, %entry ]
- %r = sext i8 %v to i32
- ret i32 %r
-}
diff --git a/llvm/test/CodeGen/AArch64/aarch64-gep-opt.ll b/llvm/test/CodeGen/AArch64/aarch64-gep-opt.ll
index b74873c61748..df9534ffde09 100644
--- a/llvm/test/CodeGen/AArch64/aarch64-gep-opt.ll
+++ b/llvm/test/CodeGen/AArch64/aarch64-gep-opt.ll
@@ -88,6 +88,7 @@ exit:
; CHECK-LABEL: test_GEP_across_BB:
; CHECK: ldr {{w[0-9]+}}, [{{x[0-9]+}}, #528]
; CHECK: ldr {{w[0-9]+}}, [{{x[0-9]+}}, #532]
+; CHECK-NOT: add
; CHECK: str {{w[0-9]+}}, [{{x[0-9]+}}, #532]
; CHECK: str {{w[0-9]+}}, [{{x[0-9]+}}, #528]
diff --git a/llvm/test/CodeGen/AArch64/branch-relax-alignment.ll b/llvm/test/CodeGen/AArch64/branch-relax-alignment.ll
index 308917be0015..7135dff7f573 100644
--- a/llvm/test/CodeGen/AArch64/branch-relax-alignment.ll
+++ b/llvm/test/CodeGen/AArch64/branch-relax-alignment.ll
@@ -1,26 +1,19 @@
-; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc -mtriple=aarch64-apple-darwin -aarch64-bcc-offset-bits=4 -align-all-nofallthru-blocks=4 < %s | FileCheck %s
; Long branch is assumed because the block has a higher alignment
; requirement than the function.
-define i32 @invert_bcc_block_align_higher_func(i32 %x, i32 %y) align 4 #0 {
; CHECK-LABEL: invert_bcc_block_align_higher_func:
-; CHECK: ; %bb.0:
-; CHECK-NEXT: cmp w0, w1
-; CHECK-NEXT: b.ne LBB0_1
-; CHECK-NEXT: b LBB0_2
-; CHECK-NEXT: LBB0_1: ; %bb2
-; CHECK-NEXT: mov w8, #9
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: str w8, [x8]
-; CHECK-NEXT: ret
-; CHECK-NEXT: .p2align 4
-; CHECK-NEXT: LBB0_2: ; %bb1
-; CHECK-NEXT: mov w8, #42
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: str w8, [x8]
-; CHECK-NEXT: ret
+; CHECK: b.eq [[JUMP_BB1:LBB[0-9]+_[0-9]+]]
+; CHECK-NEXT: b [[JUMP_BB2:LBB[0-9]+_[0-9]+]]
+
+; CHECK: [[JUMP_BB1]]:
+; CHECK: ret
+; CHECK: .p2align 4
+
+; CHECK: [[JUMP_BB2]]:
+; CHECK: ret
+define i32 @invert_bcc_block_align_higher_func(i32 %x, i32 %y) align 4 #0 {
%1 = icmp eq i32 %x, %y
br i1 %1, label %bb1, label %bb2
diff --git a/llvm/test/CodeGen/AArch64/combine-comparisons-by-cse.ll b/llvm/test/CodeGen/AArch64/combine-comparisons-by-cse.ll
index f8aab08da1cd..e7c6e3b5ef7b 100644
--- a/llvm/test/CodeGen/AArch64/combine-comparisons-by-cse.ll
+++ b/llvm/test/CodeGen/AArch64/combine-comparisons-by-cse.ll
@@ -1,4 +1,3 @@
-; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc -debugify-and-strip-all-safe < %s -mtriple=aarch64-linux-gnu | FileCheck %s
; marked as external to prevent possible optimizations
@@ -9,34 +8,12 @@
; (a > 10 && b == c) || (a >= 10 && b == d)
define i32 @combine_gt_ge_10() #0 {
-; CHECK-LABEL: combine_gt_ge_10:
-; CHECK: // %bb.0: // %entry
-; CHECK-NEXT: adrp x8, a
-; CHECK-NEXT: ldr w8, [x8, :lo12:a]
-; CHECK-NEXT: cmp w8, #10 // =10
-; CHECK-NEXT: adrp x8, b
-; CHECK-NEXT: b.le .LBB0_2
-; CHECK-NEXT: // %bb.1: // %land.lhs.true
-; CHECK-NEXT: adrp x10, c
-; CHECK-NEXT: ldr w9, [x8, :lo12:b]
-; CHECK-NEXT: ldr w10, [x10, :lo12:c]
-; CHECK-NEXT: cmp w9, w10
-; CHECK-NEXT: b.ne .LBB0_3
-; CHECK-NEXT: b .LBB0_5
-; CHECK-NEXT: .LBB0_2: // %lor.lhs.false
-; CHECK-NEXT: b.lt .LBB0_4
-; CHECK-NEXT: .LBB0_3: // %land.lhs.true3
-; CHECK-NEXT: adrp x9, d
-; CHECK-NEXT: ldr w8, [x8, :lo12:b]
-; CHECK-NEXT: ldr w9, [x9, :lo12:d]
-; CHECK-NEXT: cmp w8, w9
-; CHECK-NEXT: b.eq .LBB0_5
-; CHECK-NEXT: .LBB0_4: // %if.end
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
-; CHECK-NEXT: .LBB0_5:
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
+; CHECK-LABEL: combine_gt_ge_10
+; CHECK: cmp
+; CHECK: b.le
+; CHECK: ret
+; CHECK-NOT: cmp
+; CHECK: b.lt
entry:
%0 = load i32, i32* @a, align 4
%cmp = icmp sgt i32 %0, 10
@@ -68,35 +45,12 @@ return: ; preds = %if.end, %land.lhs.t
; (a > 5 && b == c) || (a < 5 && b == d)
define i32 @combine_gt_lt_5() #0 {
-; CHECK-LABEL: combine_gt_lt_5:
-; CHECK: // %bb.0: // %entry
-; CHECK-NEXT: adrp x8, a
-; CHECK-NEXT: ldr w8, [x8, :lo12:a]
-; CHECK-NEXT: cmp w8, #5 // =5
-; CHECK-NEXT: b.le .LBB1_2
-; CHECK-NEXT: // %bb.1: // %land.lhs.true
-; CHECK-NEXT: adrp x8, b
-; CHECK-NEXT: adrp x9, c
-; CHECK-NEXT: ldr w8, [x8, :lo12:b]
-; CHECK-NEXT: ldr w9, [x9, :lo12:c]
-; CHECK-NEXT: cmp w8, w9
-; CHECK-NEXT: b.ne .LBB1_4
-; CHECK-NEXT: b .LBB1_5
-; CHECK-NEXT: .LBB1_2: // %lor.lhs.false
-; CHECK-NEXT: b.ge .LBB1_4
-; CHECK-NEXT: // %bb.3: // %land.lhs.true3
-; CHECK-NEXT: adrp x8, b
-; CHECK-NEXT: adrp x9, d
-; CHECK-NEXT: ldr w8, [x8, :lo12:b]
-; CHECK-NEXT: ldr w9, [x9, :lo12:d]
-; CHECK-NEXT: cmp w8, w9
-; CHECK-NEXT: b.eq .LBB1_5
-; CHECK-NEXT: .LBB1_4: // %if.end
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
-; CHECK-NEXT: .LBB1_5:
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
+; CHECK-LABEL: combine_gt_lt_5
+; CHECK: cmp
+; CHECK: b.le
+; CHECK: ret
+; CHECK-NOT: cmp
+; CHECK: b.ge
entry:
%0 = load i32, i32* @a, align 4
%cmp = icmp sgt i32 %0, 5
@@ -128,34 +82,12 @@ return: ; preds = %if.end, %land.lhs.t
; (a < 5 && b == c) || (a <= 5 && b == d)
define i32 @combine_lt_ge_5() #0 {
-; CHECK-LABEL: combine_lt_ge_5:
-; CHECK: // %bb.0: // %entry
-; CHECK-NEXT: adrp x8, a
-; CHECK-NEXT: ldr w8, [x8, :lo12:a]
-; CHECK-NEXT: cmp w8, #5 // =5
-; CHECK-NEXT: adrp x8, b
-; CHECK-NEXT: b.ge .LBB2_2
-; CHECK-NEXT: // %bb.1: // %land.lhs.true
-; CHECK-NEXT: adrp x10, c
-; CHECK-NEXT: ldr w9, [x8, :lo12:b]
-; CHECK-NEXT: ldr w10, [x10, :lo12:c]
-; CHECK-NEXT: cmp w9, w10
-; CHECK-NEXT: b.ne .LBB2_3
-; CHECK-NEXT: b .LBB2_5
-; CHECK-NEXT: .LBB2_2: // %lor.lhs.false
-; CHECK-NEXT: b.gt .LBB2_4
-; CHECK-NEXT: .LBB2_3: // %land.lhs.true3
-; CHECK-NEXT: adrp x9, d
-; CHECK-NEXT: ldr w8, [x8, :lo12:b]
-; CHECK-NEXT: ldr w9, [x9, :lo12:d]
-; CHECK-NEXT: cmp w8, w9
-; CHECK-NEXT: b.eq .LBB2_5
-; CHECK-NEXT: .LBB2_4: // %if.end
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
-; CHECK-NEXT: .LBB2_5:
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
+; CHECK-LABEL: combine_lt_ge_5
+; CHECK: cmp
+; CHECK: b.ge
+; CHECK: ret
+; CHECK-NOT: cmp
+; CHECK: b.gt
entry:
%0 = load i32, i32* @a, align 4
%cmp = icmp slt i32 %0, 5
@@ -187,35 +119,12 @@ return: ; preds = %if.end, %land.lhs.t
; (a < 5 && b == c) || (a > 5 && b == d)
define i32 @combine_lt_gt_5() #0 {
-; CHECK-LABEL: combine_lt_gt_5:
-; CHECK: // %bb.0: // %entry
-; CHECK-NEXT: adrp x8, a
-; CHECK-NEXT: ldr w8, [x8, :lo12:a]
-; CHECK-NEXT: cmp w8, #5 // =5
-; CHECK-NEXT: b.ge .LBB3_2
-; CHECK-NEXT: // %bb.1: // %land.lhs.true
-; CHECK-NEXT: adrp x8, b
-; CHECK-NEXT: adrp x9, c
-; CHECK-NEXT: ldr w8, [x8, :lo12:b]
-; CHECK-NEXT: ldr w9, [x9, :lo12:c]
-; CHECK-NEXT: cmp w8, w9
-; CHECK-NEXT: b.ne .LBB3_4
-; CHECK-NEXT: b .LBB3_5
-; CHECK-NEXT: .LBB3_2: // %lor.lhs.false
-; CHECK-NEXT: b.le .LBB3_4
-; CHECK-NEXT: // %bb.3: // %land.lhs.true3
-; CHECK-NEXT: adrp x8, b
-; CHECK-NEXT: adrp x9, d
-; CHECK-NEXT: ldr w8, [x8, :lo12:b]
-; CHECK-NEXT: ldr w9, [x9, :lo12:d]
-; CHECK-NEXT: cmp w8, w9
-; CHECK-NEXT: b.eq .LBB3_5
-; CHECK-NEXT: .LBB3_4: // %if.end
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
-; CHECK-NEXT: .LBB3_5:
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
+; CHECK-LABEL: combine_lt_gt_5
+; CHECK: cmp
+; CHECK: b.ge
+; CHECK: ret
+; CHECK-NOT: cmp
+; CHECK: b.le
entry:
%0 = load i32, i32* @a, align 4
%cmp = icmp slt i32 %0, 5
@@ -247,35 +156,12 @@ return: ; preds = %if.end, %land.lhs.t
; (a > -5 && b == c) || (a < -5 && b == d)
define i32 @combine_gt_lt_n5() #0 {
-; CHECK-LABEL: combine_gt_lt_n5:
-; CHECK: // %bb.0: // %entry
-; CHECK-NEXT: adrp x8, a
-; CHECK-NEXT: ldr w8, [x8, :lo12:a]
-; CHECK-NEXT: cmn w8, #5 // =5
-; CHECK-NEXT: b.le .LBB4_2
-; CHECK-NEXT: // %bb.1: // %land.lhs.true
-; CHECK-NEXT: adrp x8, b
-; CHECK-NEXT: adrp x9, c
-; CHECK-NEXT: ldr w8, [x8, :lo12:b]
-; CHECK-NEXT: ldr w9, [x9, :lo12:c]
-; CHECK-NEXT: cmp w8, w9
-; CHECK-NEXT: b.ne .LBB4_4
-; CHECK-NEXT: b .LBB4_5
-; CHECK-NEXT: .LBB4_2: // %lor.lhs.false
-; CHECK-NEXT: b.ge .LBB4_4
-; CHECK-NEXT: // %bb.3: // %land.lhs.true3
-; CHECK-NEXT: adrp x8, b
-; CHECK-NEXT: adrp x9, d
-; CHECK-NEXT: ldr w8, [x8, :lo12:b]
-; CHECK-NEXT: ldr w9, [x9, :lo12:d]
-; CHECK-NEXT: cmp w8, w9
-; CHECK-NEXT: b.eq .LBB4_5
-; CHECK-NEXT: .LBB4_4: // %if.end
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
-; CHECK-NEXT: .LBB4_5:
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
+; CHECK-LABEL: combine_gt_lt_n5
+; CHECK: cmn
+; CHECK: b.le
+; CHECK: ret
+; CHECK-NOT: cmn
+; CHECK: b.ge
entry:
%0 = load i32, i32* @a, align 4
%cmp = icmp sgt i32 %0, -5
@@ -307,35 +193,12 @@ return: ; preds = %if.end, %land.lhs.t
; (a < -5 && b == c) || (a > -5 && b == d)
define i32 @combine_lt_gt_n5() #0 {
-; CHECK-LABEL: combine_lt_gt_n5:
-; CHECK: // %bb.0: // %entry
-; CHECK-NEXT: adrp x8, a
-; CHECK-NEXT: ldr w8, [x8, :lo12:a]
-; CHECK-NEXT: cmn w8, #5 // =5
-; CHECK-NEXT: b.ge .LBB5_2
-; CHECK-NEXT: // %bb.1: // %land.lhs.true
-; CHECK-NEXT: adrp x8, b
-; CHECK-NEXT: adrp x9, c
-; CHECK-NEXT: ldr w8, [x8, :lo12:b]
-; CHECK-NEXT: ldr w9, [x9, :lo12:c]
-; CHECK-NEXT: cmp w8, w9
-; CHECK-NEXT: b.ne .LBB5_4
-; CHECK-NEXT: b .LBB5_5
-; CHECK-NEXT: .LBB5_2: // %lor.lhs.false
-; CHECK-NEXT: b.le .LBB5_4
-; CHECK-NEXT: // %bb.3: // %land.lhs.true3
-; CHECK-NEXT: adrp x8, b
-; CHECK-NEXT: adrp x9, d
-; CHECK-NEXT: ldr w8, [x8, :lo12:b]
-; CHECK-NEXT: ldr w9, [x9, :lo12:d]
-; CHECK-NEXT: cmp w8, w9
-; CHECK-NEXT: b.eq .LBB5_5
-; CHECK-NEXT: .LBB5_4: // %if.end
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
-; CHECK-NEXT: .LBB5_5:
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
+; CHECK-LABEL: combine_lt_gt_n5
+; CHECK: cmn
+; CHECK: b.ge
+; CHECK: ret
+; CHECK-NOT: cmn
+; CHECK: b.le
entry:
%0 = load i32, i32* @a, align 4
%cmp = icmp slt i32 %0, -5
@@ -373,38 +236,6 @@ declare %struct.Struct* @Update(%struct.Struct*) #1
; no checks for this case, it just should be processed without errors
define void @combine_non_adjacent_cmp_br(%struct.Struct* nocapture readonly %hdCall) #0 {
-; CHECK-LABEL: combine_non_adjacent_cmp_br:
-; CHECK: // %bb.0: // %entry
-; CHECK-NEXT: str x30, [sp, #-48]! // 8-byte Folded Spill
-; CHECK-NEXT: stp x22, x21, [sp, #16] // 16-byte Folded Spill
-; CHECK-NEXT: stp x20, x19, [sp, #32] // 16-byte Folded Spill
-; CHECK-NEXT: .cfi_def_cfa_offset 48
-; CHECK-NEXT: .cfi_offset w19, -8
-; CHECK-NEXT: .cfi_offset w20, -16
-; CHECK-NEXT: .cfi_offset w21, -24
-; CHECK-NEXT: .cfi_offset w22, -32
-; CHECK-NEXT: .cfi_offset w30, -48
-; CHECK-NEXT: ldr x19, [x0]
-; CHECK-NEXT: mov w20, #24
-; CHECK-NEXT: adrp x22, glob
-; CHECK-NEXT: add x21, x19, #2 // =2
-; CHECK-NEXT: .LBB6_1: // %land.rhs
-; CHECK-NEXT: // =>This Inner Loop Header: Depth=1
-; CHECK-NEXT: ldr x8, [x20]
-; CHECK-NEXT: cmp x8, #1 // =1
-; CHECK-NEXT: b.lt .LBB6_3
-; CHECK-NEXT: // %bb.2: // %while.body
-; CHECK-NEXT: // in Loop: Header=BB6_1 Depth=1
-; CHECK-NEXT: ldr x0, [x22, :lo12:glob]
-; CHECK-NEXT: bl Update
-; CHECK-NEXT: sub x21, x21, #2 // =2
-; CHECK-NEXT: cmp x19, x21
-; CHECK-NEXT: b.lt .LBB6_1
-; CHECK-NEXT: .LBB6_3: // %while.end
-; CHECK-NEXT: ldp x20, x19, [sp, #32] // 16-byte Folded Reload
-; CHECK-NEXT: ldp x22, x21, [sp, #16] // 16-byte Folded Reload
-; CHECK-NEXT: ldr x30, [sp], #48 // 8-byte Folded Reload
-; CHECK-NEXT: ret
entry:
%size = getelementptr inbounds %struct.Struct, %struct.Struct* %hdCall, i64 0, i32 0
%0 = load i64, i64* %size, align 8
@@ -431,49 +262,11 @@ while.end:
declare void @do_something() #1
define i32 @do_nothing_if_resultant_opcodes_would_
diff er() #0 {
-; CHECK-LABEL: do_nothing_if_resultant_opcodes_would_
diff er:
-; CHECK: // %bb.0: // %entry
-; CHECK-NEXT: str x30, [sp, #-32]! // 8-byte Folded Spill
-; CHECK-NEXT: stp x20, x19, [sp, #16] // 16-byte Folded Spill
-; CHECK-NEXT: .cfi_def_cfa_offset 32
-; CHECK-NEXT: .cfi_offset w19, -8
-; CHECK-NEXT: .cfi_offset w20, -16
-; CHECK-NEXT: .cfi_offset w30, -32
-; CHECK-NEXT: adrp x19, a
-; CHECK-NEXT: ldr w8, [x19, :lo12:a]
-; CHECK-NEXT: cmn w8, #2 // =2
-; CHECK-NEXT: b.le .LBB7_2
-; CHECK-NEXT: // %bb.1: // %while.end
-; CHECK-NEXT: cmp w8, #1 // =1
-; CHECK-NEXT: b.le .LBB7_5
-; CHECK-NEXT: b .LBB7_6
-; CHECK-NEXT: .LBB7_2: // %while.body.preheader
-; CHECK-NEXT: sub w20, w8, #1 // =1
-; CHECK-NEXT: .LBB7_3: // %while.body
-; CHECK-NEXT: // =>This Inner Loop Header: Depth=1
-; CHECK-NEXT: bl do_something
-; CHECK-NEXT: adds w20, w20, #1 // =1
-; CHECK-NEXT: b.mi .LBB7_3
-; CHECK-NEXT: // %bb.4: // %while.cond.while.end_crit_edge
-; CHECK-NEXT: ldr w8, [x19, :lo12:a]
-; CHECK-NEXT: cmp w8, #1 // =1
-; CHECK-NEXT: b.gt .LBB7_6
-; CHECK-NEXT: .LBB7_5: // %land.lhs.true
-; CHECK-NEXT: adrp x8, b
-; CHECK-NEXT: adrp x9, d
-; CHECK-NEXT: ldr w8, [x8, :lo12:b]
-; CHECK-NEXT: ldr w9, [x9, :lo12:d]
-; CHECK-NEXT: cmp w8, w9
-; CHECK-NEXT: b.eq .LBB7_7
-; CHECK-NEXT: .LBB7_6: // %if.end
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: b .LBB7_8
-; CHECK-NEXT: .LBB7_7:
-; CHECK-NEXT: mov w0, #123
-; CHECK-NEXT: .LBB7_8: // %return
-; CHECK-NEXT: ldp x20, x19, [sp, #16] // 16-byte Folded Reload
-; CHECK-NEXT: ldr x30, [sp], #32 // 8-byte Folded Reload
-; CHECK-NEXT: ret
+; CHECK-LABEL: do_nothing_if_resultant_opcodes_would_
diff er
+; CHECK: cmn
+; CHECK: b.gt
+; CHECK: cmp
+; CHECK: b.gt
entry:
%0 = load i32, i32* @a, align 4
%cmp4 = icmp slt i32 %0, -1
@@ -513,43 +306,11 @@ return: ; preds = %if.end, %land.lhs.t
}
define i32 @do_nothing_if_compares_can_not_be_adjusted_to_each_other() #0 {
-; CHECK-LABEL: do_nothing_if_compares_can_not_be_adjusted_to_each_other:
-; CHECK: // %bb.0: // %entry
-; CHECK-NEXT: stp x30, x19, [sp, #-16]! // 16-byte Folded Spill
-; CHECK-NEXT: .cfi_def_cfa_offset 16
-; CHECK-NEXT: .cfi_offset w19, -8
-; CHECK-NEXT: .cfi_offset w30, -16
-; CHECK-NEXT: adrp x8, a
-; CHECK-NEXT: ldr w8, [x8, :lo12:a]
-; CHECK-NEXT: cmp w8, #0 // =0
-; CHECK-NEXT: b.gt .LBB8_3
-; CHECK-NEXT: // %bb.1: // %while.body.preheader
-; CHECK-NEXT: sub w19, w8, #1 // =1
-; CHECK-NEXT: .LBB8_2: // %while.body
-; CHECK-NEXT: // =>This Inner Loop Header: Depth=1
-; CHECK-NEXT: bl do_something
-; CHECK-NEXT: adds w19, w19, #1 // =1
-; CHECK-NEXT: b.mi .LBB8_2
-; CHECK-NEXT: .LBB8_3: // %while.end
-; CHECK-NEXT: adrp x8, c
-; CHECK-NEXT: ldr w8, [x8, :lo12:c]
-; CHECK-NEXT: cmn w8, #2 // =2
-; CHECK-NEXT: b.lt .LBB8_5
-; CHECK-NEXT: // %bb.4: // %land.lhs.true
-; CHECK-NEXT: adrp x8, b
-; CHECK-NEXT: adrp x9, d
-; CHECK-NEXT: ldr w8, [x8, :lo12:b]
-; CHECK-NEXT: ldr w9, [x9, :lo12:d]
-; CHECK-NEXT: cmp w8, w9
-; CHECK-NEXT: b.eq .LBB8_6
-; CHECK-NEXT: .LBB8_5: // %if.end
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ldp x30, x19, [sp], #16 // 16-byte Folded Reload
-; CHECK-NEXT: ret
-; CHECK-NEXT: .LBB8_6:
-; CHECK-NEXT: mov w0, #123
-; CHECK-NEXT: ldp x30, x19, [sp], #16 // 16-byte Folded Reload
-; CHECK-NEXT: ret
+; CHECK-LABEL: do_nothing_if_compares_can_not_be_adjusted_to_each_other
+; CHECK: cmp
+; CHECK: b.gt
+; CHECK: cmn
+; CHECK: b.lt
entry:
%0 = load i32, i32* @a, align 4
%cmp4 = icmp slt i32 %0, 1
@@ -595,44 +356,19 @@ return: ; preds = %if.end, %land.lhs.t
; b.gt .LBB0_5
define i32 @fcmpri(i32 %argc, i8** nocapture readonly %argv) {
+
; CHECK-LABEL: fcmpri:
-; CHECK: // %bb.0: // %entry
-; CHECK-NEXT: str d8, [sp, #-32]! // 8-byte Folded Spill
-; CHECK-NEXT: stp x30, x19, [sp, #16] // 16-byte Folded Spill
-; CHECK-NEXT: .cfi_def_cfa_offset 32
-; CHECK-NEXT: .cfi_offset w19, -8
-; CHECK-NEXT: .cfi_offset w30, -16
-; CHECK-NEXT: .cfi_offset b8, -32
-; CHECK-NEXT: cmp w0, #2 // =2
-; CHECK-NEXT: b.lt .LBB9_3
-; CHECK-NEXT: // %bb.1: // %land.lhs.true
-; CHECK-NEXT: ldr x8, [x1, #8]
-; CHECK-NEXT: cbz x8, .LBB9_3
-; CHECK-NEXT: // %bb.2:
-; CHECK-NEXT: mov w0, #3
-; CHECK-NEXT: b .LBB9_4
-; CHECK-NEXT: .LBB9_3: // %if.end
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: bl zoo
-; CHECK-NEXT: mov w19, w0
-; CHECK-NEXT: mov w0, #-1
-; CHECK-NEXT: bl yoo
-; CHECK-NEXT: cmp w19, #0 // =0
-; CHECK-NEXT: cinc w0, w19, gt
-; CHECK-NEXT: mov w1, #2
-; CHECK-NEXT: mov v8.16b, v0.16b
-; CHECK-NEXT: bl xoo
-; CHECK-NEXT: fmov d0, #-1.00000000
-; CHECK-NEXT: fadd d0, d8, d0
-; CHECK-NEXT: fcmp d8, #0.0
-; CHECK-NEXT: fcsel d0, d8, d0, gt
-; CHECK-NEXT: fmov d1, #-2.00000000
-; CHECK-NEXT: bl woo
-; CHECK-NEXT: mov w0, #4
-; CHECK-NEXT: .LBB9_4: // %return
-; CHECK-NEXT: ldp x30, x19, [sp, #16] // 16-byte Folded Reload
-; CHECK-NEXT: ldr d8, [sp], #32 // 8-byte Folded Reload
-; CHECK-NEXT: ret
+; CHECK: cmp w0, #2
+; CHECK: b.lt .LBB9_3
+; CHECK-NOT: cmp w0, #1
+; CHECK-NOT: b.le .LBB9_3
+
+; CHECK-LABEL-DAG: .LBB9_3
+; CHECK: cmp w19, #0
+; CHECK: fcmp d8, #0.0
+; CHECK-NOT: cmp w19, #1
+; CHECK-NOT: b.ge .LBB9_5
+
entry:
%cmp = icmp sgt i32 %argc, 1
br i1 %cmp, label %land.lhs.true, label %if.end
@@ -669,27 +405,10 @@ return: ; preds = %land.lhs.true, %con
define void @cmp_shifted(i32 %in, i32 %lhs, i32 %rhs) {
; CHECK-LABEL: cmp_shifted:
-; CHECK: // %bb.0:
-; CHECK-NEXT: str x30, [sp, #-16]! // 8-byte Folded Spill
-; CHECK-NEXT: .cfi_def_cfa_offset 16
-; CHECK-NEXT: .cfi_offset w30, -16
-; CHECK-NEXT: cmp w0, #2, lsl #12 // =8192
-; CHECK-NEXT: b.lt .LBB10_2
-; CHECK-NEXT: // %bb.1: // %true
-; CHECK-NEXT: mov w0, #128
-; CHECK-NEXT: b .LBB10_5
-; CHECK-NEXT: .LBB10_2: // %false
-; CHECK-NEXT: cmp w0, #1 // =1
-; CHECK-NEXT: b.lt .LBB10_4
-; CHECK-NEXT: // %bb.3: // %truer
-; CHECK-NEXT: mov w0, #42
-; CHECK-NEXT: b .LBB10_5
-; CHECK-NEXT: .LBB10_4: // %falser
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: .LBB10_5: // %true
-; CHECK-NEXT: bl zoo
-; CHECK-NEXT: ldr x30, [sp], #16 // 8-byte Folded Reload
-; CHECK-NEXT: ret
+; CHECK: cmp w0, #2, lsl #12
+; [...]
+; CHECK: cmp w0, #1
+
%tst_low = icmp sgt i32 %in, 8191
br i1 %tst_low, label %true, label %false
@@ -711,38 +430,10 @@ falser:
}
define i32 @combine_gt_ge_sel(i64 %v, i64* %p) #0 {
-; CHECK-LABEL: combine_gt_ge_sel:
-; CHECK: // %bb.0: // %entry
-; CHECK-NEXT: adrp x8, a
-; CHECK-NEXT: ldr w8, [x8, :lo12:a]
-; CHECK-NEXT: cmp w8, #0 // =0
-; CHECK-NEXT: csel x9, x0, xzr, gt
-; CHECK-NEXT: str x9, [x1]
-; CHECK-NEXT: b.le .LBB11_2
-; CHECK-NEXT: // %bb.1: // %lor.lhs.false
-; CHECK-NEXT: cmp w8, #2 // =2
-; CHECK-NEXT: b.ge .LBB11_3
-; CHECK-NEXT: b .LBB11_4
-; CHECK-NEXT: .LBB11_2: // %land.lhs.true
-; CHECK-NEXT: adrp x8, b
-; CHECK-NEXT: adrp x9, c
-; CHECK-NEXT: ldr w8, [x8, :lo12:b]
-; CHECK-NEXT: ldr w9, [x9, :lo12:c]
-; CHECK-NEXT: cmp w8, w9
-; CHECK-NEXT: b.eq .LBB11_5
-; CHECK-NEXT: .LBB11_3: // %land.lhs.true3
-; CHECK-NEXT: adrp x8, b
-; CHECK-NEXT: adrp x9, d
-; CHECK-NEXT: ldr w8, [x8, :lo12:b]
-; CHECK-NEXT: ldr w9, [x9, :lo12:d]
-; CHECK-NEXT: cmp w8, w9
-; CHECK-NEXT: b.eq .LBB11_5
-; CHECK-NEXT: .LBB11_4: // %if.end
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
-; CHECK-NEXT: .LBB11_5:
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
+; CHECK-LABEL: combine_gt_ge_sel
+; CHECK: ldr [[reg1:w[0-9]*]],
+; CHECK: cmp [[reg1]], #0
+; CHECK: csel {{.*}}, gt
entry:
%0 = load i32, i32* @a, align 4
%cmp = icmp sgt i32 %0, 0
diff --git a/llvm/test/CodeGen/AArch64/cond-br-tuning.ll b/llvm/test/CodeGen/AArch64/cond-br-tuning.ll
index e98d4110fd27..e0b6a2f05037 100644
--- a/llvm/test/CodeGen/AArch64/cond-br-tuning.ll
+++ b/llvm/test/CodeGen/AArch64/cond-br-tuning.ll
@@ -1,22 +1,14 @@
-; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc -debugify-and-strip-all-safe < %s -O3 -mtriple=aarch64-eabi -verify-machineinstrs | FileCheck %s
target datalayout = "e-m:e-i8:8:32-i16:16:32-i64:64-i128:128-n32:64-S128"
target triple = "aarch64-linaro-linux-gnueabi"
; CMN is an alias of ADDS.
-define void @test_add_cbz(i32 %a, i32 %b, i32* %ptr) {
; CHECK-LABEL: test_add_cbz:
-; CHECK: // %bb.0:
-; CHECK-NEXT: cmn w0, w1
-; CHECK-NEXT: b.eq .LBB0_2
-; CHECK-NEXT: // %bb.1: // %L1
-; CHECK-NEXT: str wzr, [x2]
-; CHECK-NEXT: ret
-; CHECK-NEXT: .LBB0_2: // %L2
-; CHECK-NEXT: mov w8, #1
-; CHECK-NEXT: str w8, [x2]
-; CHECK-NEXT: ret
+; CHECK: cmn w0, w1
+; CHECK: b.eq
+; CHECK: ret
+define void @test_add_cbz(i32 %a, i32 %b, i32* %ptr) {
%c = add nsw i32 %a, %b
%d = icmp ne i32 %c, 0
br i1 %d, label %L1, label %L2
@@ -28,17 +20,11 @@ L2:
ret void
}
-define void @test_add_cbz_multiple_use(i32 %a, i32 %b, i32* %ptr) {
; CHECK-LABEL: test_add_cbz_multiple_use:
-; CHECK: // %bb.0:
-; CHECK-NEXT: adds w8, w0, w1
-; CHECK-NEXT: b.eq .LBB1_2
-; CHECK-NEXT: // %bb.1: // %L1
-; CHECK-NEXT: str wzr, [x2]
-; CHECK-NEXT: ret
-; CHECK-NEXT: .LBB1_2: // %L2
-; CHECK-NEXT: str w8, [x2]
-; CHECK-NEXT: ret
+; CHECK: adds
+; CHECK: b.eq
+; CHECK: ret
+define void @test_add_cbz_multiple_use(i32 %a, i32 %b, i32* %ptr) {
%c = add nsw i32 %a, %b
%d = icmp ne i32 %c, 0
br i1 %d, label %L1, label %L2
@@ -50,18 +36,10 @@ L2:
ret void
}
-define void @test_add_cbz_64(i64 %a, i64 %b, i64* %ptr) {
; CHECK-LABEL: test_add_cbz_64:
-; CHECK: // %bb.0:
-; CHECK-NEXT: cmn x0, x1
-; CHECK-NEXT: b.eq .LBB2_2
-; CHECK-NEXT: // %bb.1: // %L1
-; CHECK-NEXT: str xzr, [x2]
-; CHECK-NEXT: ret
-; CHECK-NEXT: .LBB2_2: // %L2
-; CHECK-NEXT: mov w8, #1
-; CHECK-NEXT: str x8, [x2]
-; CHECK-NEXT: ret
+; CHECK: cmn x0, x1
+; CHECK: b.eq
+define void @test_add_cbz_64(i64 %a, i64 %b, i64* %ptr) {
%c = add nsw i64 %a, %b
%d = icmp ne i64 %c, 0
br i1 %d, label %L1, label %L2
@@ -73,18 +51,10 @@ L2:
ret void
}
-define void @test_and_cbz(i32 %a, i32* %ptr) {
; CHECK-LABEL: test_and_cbz:
-; CHECK: // %bb.0:
-; CHECK-NEXT: tst w0, #0x6
-; CHECK-NEXT: b.eq .LBB3_2
-; CHECK-NEXT: // %bb.1: // %L1
-; CHECK-NEXT: str wzr, [x1]
-; CHECK-NEXT: ret
-; CHECK-NEXT: .LBB3_2: // %L2
-; CHECK-NEXT: mov w8, #1
-; CHECK-NEXT: str w8, [x1]
-; CHECK-NEXT: ret
+; CHECK: tst w0, #0x6
+; CHECK: b.eq
+define void @test_and_cbz(i32 %a, i32* %ptr) {
%c = and i32 %a, 6
%d = icmp ne i32 %c, 0
br i1 %d, label %L1, label %L2
@@ -96,18 +66,10 @@ L2:
ret void
}
-define void @test_bic_cbnz(i32 %a, i32 %b, i32* %ptr) {
; CHECK-LABEL: test_bic_cbnz:
-; CHECK: // %bb.0:
-; CHECK-NEXT: bics wzr, w1, w0
-; CHECK-NEXT: b.eq .LBB4_2
-; CHECK-NEXT: // %bb.1: // %L2
-; CHECK-NEXT: mov w8, #1
-; CHECK-NEXT: str w8, [x2]
-; CHECK-NEXT: ret
-; CHECK-NEXT: .LBB4_2: // %L1
-; CHECK-NEXT: str wzr, [x2]
-; CHECK-NEXT: ret
+; CHECK: bics wzr, w1, w0
+; CHECK: b.ne
+define void @test_bic_cbnz(i32 %a, i32 %b, i32* %ptr) {
%c = and i32 %a, %b
%d = icmp eq i32 %c, %b
br i1 %d, label %L1, label %L2
@@ -119,15 +81,11 @@ L2:
ret void
}
-define void @test_add_tbz(i32 %a, i32 %b, i32* %ptr) {
; CHECK-LABEL: test_add_tbz:
-; CHECK: // %bb.0: // %entry
-; CHECK-NEXT: adds w8, w0, w1
-; CHECK-NEXT: b.pl .LBB5_2
-; CHECK-NEXT: // %bb.1: // %L1
-; CHECK-NEXT: str w8, [x2]
-; CHECK-NEXT: .LBB5_2: // %L2
-; CHECK-NEXT: ret
+; CHECK: adds
+; CHECK: b.pl
+; CHECK: ret
+define void @test_add_tbz(i32 %a, i32 %b, i32* %ptr) {
entry:
%add = add nsw i32 %a, %b
%cmp36 = icmp sge i32 %add, 0
@@ -139,15 +97,11 @@ L2:
ret void
}
-define void @test_subs_tbz(i32 %a, i32 %b, i32* %ptr) {
; CHECK-LABEL: test_subs_tbz:
-; CHECK: // %bb.0: // %entry
-; CHECK-NEXT: subs w8, w0, w1
-; CHECK-NEXT: b.pl .LBB6_2
-; CHECK-NEXT: // %bb.1: // %L1
-; CHECK-NEXT: str w8, [x2]
-; CHECK-NEXT: .LBB6_2: // %L2
-; CHECK-NEXT: ret
+; CHECK: subs
+; CHECK: b.pl
+; CHECK: ret
+define void @test_subs_tbz(i32 %a, i32 %b, i32* %ptr) {
entry:
%sub = sub nsw i32 %a, %b
%cmp36 = icmp sge i32 %sub, 0
@@ -159,15 +113,11 @@ L2:
ret void
}
+; CHECK-LABEL: test_add_tbnz
+; CHECK: adds
+; CHECK: b.mi
+; CHECK: ret
define void @test_add_tbnz(i32 %a, i32 %b, i32* %ptr) {
-; CHECK-LABEL: test_add_tbnz:
-; CHECK: // %bb.0: // %entry
-; CHECK-NEXT: adds w8, w0, w1
-; CHECK-NEXT: b.mi .LBB7_2
-; CHECK-NEXT: // %bb.1: // %L1
-; CHECK-NEXT: str w8, [x2]
-; CHECK-NEXT: .LBB7_2: // %L2
-; CHECK-NEXT: ret
entry:
%add = add nsw i32 %a, %b
%cmp36 = icmp slt i32 %add, 0
@@ -179,15 +129,11 @@ L2:
ret void
}
+; CHECK-LABEL: test_subs_tbnz
+; CHECK: subs
+; CHECK: b.mi
+; CHECK: ret
define void @test_subs_tbnz(i32 %a, i32 %b, i32* %ptr) {
-; CHECK-LABEL: test_subs_tbnz:
-; CHECK: // %bb.0: // %entry
-; CHECK-NEXT: subs w8, w0, w1
-; CHECK-NEXT: b.mi .LBB8_2
-; CHECK-NEXT: // %bb.1: // %L1
-; CHECK-NEXT: str w8, [x2]
-; CHECK-NEXT: .LBB8_2: // %L2
-; CHECK-NEXT: ret
entry:
%sub = sub nsw i32 %a, %b
%cmp36 = icmp slt i32 %sub, 0
@@ -203,22 +149,11 @@ declare void @foo()
declare void @bar(i32)
; Don't transform since the call will clobber the NZCV bits.
-define void @test_call_clobber(i32 %unused, i32 %a) {
; CHECK-LABEL: test_call_clobber:
-; CHECK: // %bb.0: // %entry
-; CHECK-NEXT: stp x30, x19, [sp, #-16]! // 16-byte Folded Spill
-; CHECK-NEXT: .cfi_def_cfa_offset 16
-; CHECK-NEXT: .cfi_offset w19, -8
-; CHECK-NEXT: .cfi_offset w30, -16
-; CHECK-NEXT: and w19, w1, #0x6
-; CHECK-NEXT: mov w0, w19
-; CHECK-NEXT: bl bar
-; CHECK-NEXT: cbnz w19, .LBB9_2
-; CHECK-NEXT: // %bb.1: // %if.end
-; CHECK-NEXT: ldp x30, x19, [sp], #16 // 16-byte Folded Reload
-; CHECK-NEXT: ret
-; CHECK-NEXT: .LBB9_2: // %if.then
-; CHECK-NEXT: bl foo
+; CHECK: and w[[DST:[0-9]+]], w1, #0x6
+; CHECK: bl bar
+; CHECK: cbnz w[[DST]]
+define void @test_call_clobber(i32 %unused, i32 %a) {
entry:
%c = and i32 %a, 6
call void @bar(i32 %c)
diff --git a/llvm/test/CodeGen/AArch64/fast-isel-cmp-branch.ll b/llvm/test/CodeGen/AArch64/fast-isel-cmp-branch.ll
index d1c762585a9e..ce47bc42453c 100644
--- a/llvm/test/CodeGen/AArch64/fast-isel-cmp-branch.ll
+++ b/llvm/test/CodeGen/AArch64/fast-isel-cmp-branch.ll
@@ -1,18 +1,10 @@
-; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
-; RUN: llc -aarch64-enable-atomic-cfg-tidy=0 -mtriple=aarch64-apple-darwin < %s | FileCheck %s --check-prefixes=CHECK,NOFASTISEL
-; RUN: llc -fast-isel -fast-isel-abort=1 -aarch64-enable-atomic-cfg-tidy=0 -mtriple=aarch64-apple-darwin < %s | FileCheck %s --check-prefixes=CHECK,FASTISEL
+; RUN: llc -aarch64-enable-atomic-cfg-tidy=0 -mtriple=aarch64-apple-darwin < %s | FileCheck %s
+; RUN: llc -fast-isel -fast-isel-abort=1 -aarch64-enable-atomic-cfg-tidy=0 -mtriple=aarch64-apple-darwin < %s | FileCheck %s
define i32 @fcmp_oeq(float %x, float %y) {
-; CHECK-LABEL: fcmp_oeq:
-; CHECK: ; %bb.0:
-; CHECK-NEXT: fcmp s0, s1
-; CHECK-NEXT: b.ne LBB0_2
-; CHECK-NEXT: ; %bb.1: ; %bb1
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
-; CHECK-NEXT: LBB0_2: ; %bb2
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
+; CHECK-LABEL: fcmp_oeq
+; CHECK: fcmp s0, s1
+; CHECK-NEXT: b.ne {{LBB.+_2}}
%1 = fcmp oeq float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -22,16 +14,9 @@ bb1:
}
define i32 @fcmp_ogt(float %x, float %y) {
-; CHECK-LABEL: fcmp_ogt:
-; CHECK: ; %bb.0:
-; CHECK-NEXT: fcmp s0, s1
-; CHECK-NEXT: b.le LBB1_2
-; CHECK-NEXT: ; %bb.1: ; %bb1
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
-; CHECK-NEXT: LBB1_2: ; %bb2
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
+; CHECK-LABEL: fcmp_ogt
+; CHECK: fcmp s0, s1
+; CHECK-NEXT: b.le {{LBB.+_2}}
%1 = fcmp ogt float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -41,16 +26,9 @@ bb1:
}
define i32 @fcmp_oge(float %x, float %y) {
-; CHECK-LABEL: fcmp_oge:
-; CHECK: ; %bb.0:
-; CHECK-NEXT: fcmp s0, s1
-; CHECK-NEXT: b.lt LBB2_2
-; CHECK-NEXT: ; %bb.1: ; %bb1
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
-; CHECK-NEXT: LBB2_2: ; %bb2
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
+; CHECK-LABEL: fcmp_oge
+; CHECK: fcmp s0, s1
+; CHECK-NEXT: b.lt {{LBB.+_2}}
%1 = fcmp oge float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -60,16 +38,9 @@ bb1:
}
define i32 @fcmp_olt(float %x, float %y) {
-; CHECK-LABEL: fcmp_olt:
-; CHECK: ; %bb.0:
-; CHECK-NEXT: fcmp s0, s1
-; CHECK-NEXT: b.pl LBB3_2
-; CHECK-NEXT: ; %bb.1: ; %bb1
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
-; CHECK-NEXT: LBB3_2: ; %bb2
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
+; CHECK-LABEL: fcmp_olt
+; CHECK: fcmp s0, s1
+; CHECK-NEXT: b.pl {{LBB.+_2}}
%1 = fcmp olt float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -79,16 +50,9 @@ bb1:
}
define i32 @fcmp_ole(float %x, float %y) {
-; CHECK-LABEL: fcmp_ole:
-; CHECK: ; %bb.0:
-; CHECK-NEXT: fcmp s0, s1
-; CHECK-NEXT: b.hi LBB4_2
-; CHECK-NEXT: ; %bb.1: ; %bb1
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
-; CHECK-NEXT: LBB4_2: ; %bb2
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
+; CHECK-LABEL: fcmp_ole
+; CHECK: fcmp s0, s1
+; CHECK-NEXT: b.hi {{LBB.+_2}}
%1 = fcmp ole float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -98,30 +62,10 @@ bb1:
}
define i32 @fcmp_one(float %x, float %y) {
-; NOFASTISEL-LABEL: fcmp_one:
-; NOFASTISEL: ; %bb.0:
-; NOFASTISEL-NEXT: fcmp s0, s1
-; NOFASTISEL-NEXT: b.mi LBB5_1
-; NOFASTISEL-NEXT: b.gt LBB5_1
-; NOFASTISEL-NEXT: b LBB5_2
-; NOFASTISEL-NEXT: LBB5_1: ; %bb1
-; NOFASTISEL-NEXT: mov w0, wzr
-; NOFASTISEL-NEXT: ret
-; NOFASTISEL-NEXT: LBB5_2: ; %bb2
-; NOFASTISEL-NEXT: mov w0, #1
-; NOFASTISEL-NEXT: ret
-;
-; FASTISEL-LABEL: fcmp_one:
-; FASTISEL: ; %bb.0:
-; FASTISEL-NEXT: fcmp s0, s1
-; FASTISEL-NEXT: b.mi LBB5_2
-; FASTISEL-NEXT: b.gt LBB5_2
-; FASTISEL-NEXT: ; %bb.1: ; %bb2
-; FASTISEL-NEXT: mov w0, #1
-; FASTISEL-NEXT: ret
-; FASTISEL-NEXT: LBB5_2: ; %bb1
-; FASTISEL-NEXT: mov w0, wzr
-; FASTISEL-NEXT: ret
+; CHECK-LABEL: fcmp_one
+; CHECK: fcmp s0, s1
+; CHECK-NEXT: b.mi
+; CHECK-NEXT: b.gt
%1 = fcmp one float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -131,16 +75,9 @@ bb1:
}
define i32 @fcmp_ord(float %x, float %y) {
-; CHECK-LABEL: fcmp_ord:
-; CHECK: ; %bb.0:
-; CHECK-NEXT: fcmp s0, s1
-; CHECK-NEXT: b.vs LBB6_2
-; CHECK-NEXT: ; %bb.1: ; %bb1
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
-; CHECK-NEXT: LBB6_2: ; %bb2
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
+; CHECK-LABEL: fcmp_ord
+; CHECK: fcmp s0, s1
+; CHECK-NEXT: b.vs {{LBB.+_2}}
%1 = fcmp ord float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -150,16 +87,9 @@ bb1:
}
define i32 @fcmp_uno(float %x, float %y) {
-; CHECK-LABEL: fcmp_uno:
-; CHECK: ; %bb.0:
-; CHECK-NEXT: fcmp s0, s1
-; CHECK-NEXT: b.vs LBB7_2
-; CHECK-NEXT: ; %bb.1: ; %bb2
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
-; CHECK-NEXT: LBB7_2: ; %bb1
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
+; CHECK-LABEL: fcmp_uno
+; CHECK: fcmp s0, s1
+; CHECK-NEXT: b.vs {{LBB.+_2}}
%1 = fcmp uno float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -169,30 +99,10 @@ bb1:
}
define i32 @fcmp_ueq(float %x, float %y) {
-; NOFASTISEL-LABEL: fcmp_ueq:
-; NOFASTISEL: ; %bb.0:
-; NOFASTISEL-NEXT: fcmp s0, s1
-; NOFASTISEL-NEXT: b.eq LBB8_2
-; NOFASTISEL-NEXT: b.vs LBB8_2
-; NOFASTISEL-NEXT: b LBB8_1
-; NOFASTISEL-NEXT: LBB8_1: ; %bb2
-; NOFASTISEL-NEXT: mov w0, #1
-; NOFASTISEL-NEXT: ret
-; NOFASTISEL-NEXT: LBB8_2: ; %bb1
-; NOFASTISEL-NEXT: mov w0, wzr
-; NOFASTISEL-NEXT: ret
-;
-; FASTISEL-LABEL: fcmp_ueq:
-; FASTISEL: ; %bb.0:
-; FASTISEL-NEXT: fcmp s0, s1
-; FASTISEL-NEXT: b.eq LBB8_2
-; FASTISEL-NEXT: b.vs LBB8_2
-; FASTISEL-NEXT: ; %bb.1: ; %bb2
-; FASTISEL-NEXT: mov w0, #1
-; FASTISEL-NEXT: ret
-; FASTISEL-NEXT: LBB8_2: ; %bb1
-; FASTISEL-NEXT: mov w0, wzr
-; FASTISEL-NEXT: ret
+; CHECK-LABEL: fcmp_ueq
+; CHECK: fcmp s0, s1
+; CHECK-NEXT: b.eq {{LBB.+_2}}
+; CHECK-NEXT: b.vs {{LBB.+_2}}
%1 = fcmp ueq float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -202,16 +112,9 @@ bb1:
}
define i32 @fcmp_ugt(float %x, float %y) {
-; CHECK-LABEL: fcmp_ugt:
-; CHECK: ; %bb.0:
-; CHECK-NEXT: fcmp s0, s1
-; CHECK-NEXT: b.ls LBB9_2
-; CHECK-NEXT: ; %bb.1: ; %bb1
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
-; CHECK-NEXT: LBB9_2: ; %bb2
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
+; CHECK-LABEL: fcmp_ugt
+; CHECK: fcmp s0, s1
+; CHECK-NEXT: b.ls {{LBB.+_2}}
%1 = fcmp ugt float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -221,16 +124,9 @@ bb1:
}
define i32 @fcmp_uge(float %x, float %y) {
-; CHECK-LABEL: fcmp_uge:
-; CHECK: ; %bb.0:
-; CHECK-NEXT: fcmp s0, s1
-; CHECK-NEXT: b.mi LBB10_2
-; CHECK-NEXT: ; %bb.1: ; %bb1
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
-; CHECK-NEXT: LBB10_2: ; %bb2
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
+; CHECK-LABEL: fcmp_uge
+; CHECK: fcmp s0, s1
+; CHECK-NEXT: b.mi {{LBB.+_2}}
%1 = fcmp uge float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -240,16 +136,9 @@ bb1:
}
define i32 @fcmp_ult(float %x, float %y) {
-; CHECK-LABEL: fcmp_ult:
-; CHECK: ; %bb.0:
-; CHECK-NEXT: fcmp s0, s1
-; CHECK-NEXT: b.ge LBB11_2
-; CHECK-NEXT: ; %bb.1: ; %bb1
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
-; CHECK-NEXT: LBB11_2: ; %bb2
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
+; CHECK-LABEL: fcmp_ult
+; CHECK: fcmp s0, s1
+; CHECK-NEXT: b.ge {{LBB.+_2}}
%1 = fcmp ult float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -259,16 +148,9 @@ bb1:
}
define i32 @fcmp_ule(float %x, float %y) {
-; CHECK-LABEL: fcmp_ule:
-; CHECK: ; %bb.0:
-; CHECK-NEXT: fcmp s0, s1
-; CHECK-NEXT: b.gt LBB12_2
-; CHECK-NEXT: ; %bb.1: ; %bb1
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
-; CHECK-NEXT: LBB12_2: ; %bb2
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
+; CHECK-LABEL: fcmp_ule
+; CHECK: fcmp s0, s1
+; CHECK-NEXT: b.gt {{LBB.+_2}}
%1 = fcmp ule float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -278,16 +160,9 @@ bb1:
}
define i32 @fcmp_une(float %x, float %y) {
-; CHECK-LABEL: fcmp_une:
-; CHECK: ; %bb.0:
-; CHECK-NEXT: fcmp s0, s1
-; CHECK-NEXT: b.eq LBB13_2
-; CHECK-NEXT: ; %bb.1: ; %bb1
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
-; CHECK-NEXT: LBB13_2: ; %bb2
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
+; CHECK-LABEL: fcmp_une
+; CHECK: fcmp s0, s1
+; CHECK-NEXT: b.eq {{LBB.+_2}}
%1 = fcmp une float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -297,16 +172,9 @@ bb1:
}
define i32 @icmp_eq(i32 %x, i32 %y) {
-; CHECK-LABEL: icmp_eq:
-; CHECK: ; %bb.0:
-; CHECK-NEXT: cmp w0, w1
-; CHECK-NEXT: b.eq LBB14_2
-; CHECK-NEXT: ; %bb.1: ; %bb2
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
-; CHECK-NEXT: LBB14_2: ; %bb1
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
+; CHECK-LABEL: icmp_eq
+; CHECK: cmp w0, w1
+; CHECK-NEXT: b.ne {{LBB.+_2}}
%1 = icmp eq i32 %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -316,16 +184,9 @@ bb1:
}
define i32 @icmp_ne(i32 %x, i32 %y) {
-; CHECK-LABEL: icmp_ne:
-; CHECK: ; %bb.0:
-; CHECK-NEXT: cmp w0, w1
-; CHECK-NEXT: b.eq LBB15_2
-; CHECK-NEXT: ; %bb.1: ; %bb1
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
-; CHECK-NEXT: LBB15_2: ; %bb2
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
+; CHECK-LABEL: icmp_ne
+; CHECK: cmp w0, w1
+; CHECK-NEXT: b.eq {{LBB.+_2}}
%1 = icmp ne i32 %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -335,16 +196,9 @@ bb1:
}
define i32 @icmp_ugt(i32 %x, i32 %y) {
-; CHECK-LABEL: icmp_ugt:
-; CHECK: ; %bb.0:
-; CHECK-NEXT: cmp w0, w1
-; CHECK-NEXT: b.ls LBB16_2
-; CHECK-NEXT: ; %bb.1: ; %bb1
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
-; CHECK-NEXT: LBB16_2: ; %bb2
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
+; CHECK-LABEL: icmp_ugt
+; CHECK: cmp w0, w1
+; CHECK-NEXT: b.ls {{LBB.+_2}}
%1 = icmp ugt i32 %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -354,16 +208,9 @@ bb1:
}
define i32 @icmp_uge(i32 %x, i32 %y) {
-; CHECK-LABEL: icmp_uge:
-; CHECK: ; %bb.0:
-; CHECK-NEXT: cmp w0, w1
-; CHECK-NEXT: b.hs LBB17_2
-; CHECK-NEXT: ; %bb.1: ; %bb2
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
-; CHECK-NEXT: LBB17_2: ; %bb1
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
+; CHECK-LABEL: icmp_uge
+; CHECK: cmp w0, w1
+; CHECK-NEXT: b.lo {{LBB.+_2}}
%1 = icmp uge i32 %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -373,16 +220,9 @@ bb1:
}
define i32 @icmp_ult(i32 %x, i32 %y) {
-; CHECK-LABEL: icmp_ult:
-; CHECK: ; %bb.0:
-; CHECK-NEXT: cmp w0, w1
-; CHECK-NEXT: b.hs LBB18_2
-; CHECK-NEXT: ; %bb.1: ; %bb1
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
-; CHECK-NEXT: LBB18_2: ; %bb2
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
+; CHECK-LABEL: icmp_ult
+; CHECK: cmp w0, w1
+; CHECK-NEXT: b.hs {{LBB.+_2}}
%1 = icmp ult i32 %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -392,16 +232,9 @@ bb1:
}
define i32 @icmp_ule(i32 %x, i32 %y) {
-; CHECK-LABEL: icmp_ule:
-; CHECK: ; %bb.0:
-; CHECK-NEXT: cmp w0, w1
-; CHECK-NEXT: b.ls LBB19_2
-; CHECK-NEXT: ; %bb.1: ; %bb2
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
-; CHECK-NEXT: LBB19_2: ; %bb1
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
+; CHECK-LABEL: icmp_ule
+; CHECK: cmp w0, w1
+; CHECK-NEXT: b.hi {{LBB.+_2}}
%1 = icmp ule i32 %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -411,16 +244,9 @@ bb1:
}
define i32 @icmp_sgt(i32 %x, i32 %y) {
-; CHECK-LABEL: icmp_sgt:
-; CHECK: ; %bb.0:
-; CHECK-NEXT: cmp w0, w1
-; CHECK-NEXT: b.le LBB20_2
-; CHECK-NEXT: ; %bb.1: ; %bb1
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
-; CHECK-NEXT: LBB20_2: ; %bb2
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
+; CHECK-LABEL: icmp_sgt
+; CHECK: cmp w0, w1
+; CHECK-NEXT: b.le {{LBB.+_2}}
%1 = icmp sgt i32 %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -430,16 +256,9 @@ bb1:
}
define i32 @icmp_sge(i32 %x, i32 %y) {
-; CHECK-LABEL: icmp_sge:
-; CHECK: ; %bb.0:
-; CHECK-NEXT: cmp w0, w1
-; CHECK-NEXT: b.ge LBB21_2
-; CHECK-NEXT: ; %bb.1: ; %bb2
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
-; CHECK-NEXT: LBB21_2: ; %bb1
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
+; CHECK-LABEL: icmp_sge
+; CHECK: cmp w0, w1
+; CHECK-NEXT: b.lt {{LBB.+_2}}
%1 = icmp sge i32 %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -449,16 +268,9 @@ bb1:
}
define i32 @icmp_slt(i32 %x, i32 %y) {
-; CHECK-LABEL: icmp_slt:
-; CHECK: ; %bb.0:
-; CHECK-NEXT: cmp w0, w1
-; CHECK-NEXT: b.ge LBB22_2
-; CHECK-NEXT: ; %bb.1: ; %bb1
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
-; CHECK-NEXT: LBB22_2: ; %bb2
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
+; CHECK-LABEL: icmp_slt
+; CHECK: cmp w0, w1
+; CHECK-NEXT: b.ge {{LBB.+_2}}
%1 = icmp slt i32 %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -468,16 +280,9 @@ bb1:
}
define i32 @icmp_sle(i32 %x, i32 %y) {
-; CHECK-LABEL: icmp_sle:
-; CHECK: ; %bb.0:
-; CHECK-NEXT: cmp w0, w1
-; CHECK-NEXT: b.le LBB23_2
-; CHECK-NEXT: ; %bb.1: ; %bb2
-; CHECK-NEXT: mov w0, #1
-; CHECK-NEXT: ret
-; CHECK-NEXT: LBB23_2: ; %bb1
-; CHECK-NEXT: mov w0, wzr
-; CHECK-NEXT: ret
+; CHECK-LABEL: icmp_sle
+; CHECK: cmp w0, w1
+; CHECK-NEXT: b.gt {{LBB.+_2}}
%1 = icmp sle i32 %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
diff --git a/llvm/test/CodeGen/ARM/2011-12-14-machine-sink.ll b/llvm/test/CodeGen/ARM/2011-12-14-machine-sink.ll
index 5716d7dbc405..88019f450e36 100644
--- a/llvm/test/CodeGen/ARM/2011-12-14-machine-sink.ll
+++ b/llvm/test/CodeGen/ARM/2011-12-14-machine-sink.ll
@@ -1,14 +1,9 @@
-; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; REQUIRES: asserts
; RUN: llc < %s -o /dev/null -stats 2>&1 | FileCheck %s -check-prefix=STATS
; Radar 10266272
target datalayout = "e-p:32:32:32-i1:8:32-i8:8:32-i16:16:32-i32:32:32-i64:32:64-f32:32:32-f64:32:64-v64:32:64-v128:32:128-a0:0:32-n32-S32"
target triple = "thumbv7-apple-ios4.0.0"
-
-; STATS: 1 machine-cse - Number of common subexpression eliminated
-; STATS: 1 machine-sink - Number of critical edges split
-; STATS: 1 machine-sink - Number of machine instructions sunk
-
+; STATS-NOT: machine-sink
define i32 @foo(i32 %h, i32 %arg1) nounwind readonly ssp {
entry:
diff --git a/llvm/test/CodeGen/ARM/cmpxchg-weak.ll b/llvm/test/CodeGen/ARM/cmpxchg-weak.ll
index 17f2ed74ecc0..5ee07828526c 100644
--- a/llvm/test/CodeGen/ARM/cmpxchg-weak.ll
+++ b/llvm/test/CodeGen/ARM/cmpxchg-weak.ll
@@ -5,27 +5,27 @@ define void @test_cmpxchg_weak(i32 *%addr, i32 %desired, i32 %new) {
%pair = cmpxchg weak i32* %addr, i32 %desired, i32 %new seq_cst monotonic
%oldval = extractvalue { i32, i1 } %pair, 0
-; CHECK-NEXT: @ %bb.0: @ %cmpxchg.start
-; CHECK-NEXT: ldrex r3, [r0]
-; CHECK-NEXT: cmp r3, r1
-; CHECK-NEXT: beq LBB0_2
-; CHECK-NEXT: @ %bb.1: @ %cmpxchg.nostore
+; CHECK-NEXT: %bb.0:
+; CHECK-NEXT: ldrex [[LOADED:r[0-9]+]], [r0]
+; CHECK-NEXT: cmp [[LOADED]], r1
+; CHECK-NEXT: bne [[LDFAILBB:LBB[0-9]+_[0-9]+]]
+; CHECK-NEXT: %bb.1:
+; CHECK-NEXT: dmb ish
+; CHECK-NEXT: strex [[SUCCESS:r[0-9]+]], r2, [r0]
+; CHECK-NEXT: cmp [[SUCCESS]], #0
+; CHECK-NEXT: beq [[SUCCESSBB:LBB[0-9]+_[0-9]+]]
+; CHECK-NEXT: %bb.2:
+; CHECK-NEXT: str r3, [r0]
+; CHECK-NEXT: bx lr
+; CHECK-NEXT: [[LDFAILBB]]:
; CHECK-NEXT: clrex
-; CHECK-NEXT: b LBB0_3
-; CHECK-NEXT: LBB0_2: @ %cmpxchg.fencedstore
-; CHECK-NEXT: dmb ish
-; CHECK-NEXT: strex r1, r2, [r0]
-; CHECK-NEXT: cmp r1, #0
-; CHECK-NEXT: beq LBB0_4
-; CHECK-NEXT: LBB0_3: @ %cmpxchg.end
; CHECK-NEXT: str r3, [r0]
; CHECK-NEXT: bx lr
-; CHECK-NEXT: LBB0_4: @ %cmpxchg.success
+; CHECK-NEXT: [[SUCCESSBB]]:
; CHECK-NEXT: dmb ish
; CHECK-NEXT: str r3, [r0]
; CHECK-NEXT: bx lr
-
store i32 %oldval, i32* %addr
ret void
}
@@ -37,24 +37,23 @@ define i1 @test_cmpxchg_weak_to_bool(i32, i32 *%addr, i32 %desired, i32 %new) {
%pair = cmpxchg weak i32* %addr, i32 %desired, i32 %new seq_cst monotonic
%success = extractvalue { i32, i1 } %pair, 1
-; CHECK-NEXT: @ %bb.0: @ %cmpxchg.start
-; CHECK-NEXT: ldrex r0, [r1]
-; CHECK-NEXT: cmp r0, r2
-; CHECK-NEXT: beq LBB1_2
-; CHECK-NEXT: @ %bb.1: @ %cmpxchg.nostore
+; CHECK-NEXT: %bb.0:
+; CHECK-NEXT: ldrex [[LOADED:r[0-9]+]], [r1]
+; CHECK-NEXT: cmp [[LOADED]], r2
+; CHECK-NEXT: bne [[LDFAILBB:LBB[0-9]+_[0-9]+]]
+; CHECK-NEXT: %bb.1:
+; CHECK-NEXT: dmb ish
; CHECK-NEXT: mov r0, #0
-; CHECK-NEXT: clrex
-; CHECK-NEXT: bx lr
-; CHECK-NEXT: LBB1_2: @ %cmpxchg.fencedstore
-; CHECK-NEXT: dmb ish
-; CHECK-NEXT: mov r0, #0
-; CHECK-NEXT: strex r2, r3, [r1]
-; CHECK-NEXT: cmp r2, #0
+; CHECK-NEXT: strex [[SUCCESS:r[0-9]+]], r3, [r1]
+; CHECK-NEXT: cmp [[SUCCESS]], #0
; CHECK-NEXT: bxne lr
; CHECK-NEXT: mov r0, #1
; CHECK-NEXT: dmb ish
; CHECK-NEXT: bx lr
-
+; CHECK-NEXT: [[LDFAILBB]]:
+; CHECK-NEXT: mov r0, #0
+; CHECK-NEXT: clrex
+; CHECK-NEXT: bx lr
ret i1 %success
}
diff --git a/llvm/test/CodeGen/ARM/lsr-unfolded-offset.ll b/llvm/test/CodeGen/ARM/lsr-unfolded-offset.ll
index 8656c858be5a..a9a353cad575 100644
--- a/llvm/test/CodeGen/ARM/lsr-unfolded-offset.ll
+++ b/llvm/test/CodeGen/ARM/lsr-unfolded-offset.ll
@@ -1,4 +1,3 @@
-; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc -regalloc=greedy -arm-atomic-cfg-tidy=0 < %s | FileCheck %s
; LSR shouldn't introduce more induction variables than needed, increasing
@@ -8,11 +7,8 @@
; CHECK: sub sp, #{{40|36|32|28|24}}
; CHECK: %for.inc
-; CHECK: adds r6, #1
-; CHECK: adds r4, #24
-; CHECK: cmp r1, r6
-; CHECK: bne LBB0_3
-
+; CHECK-NOT: ldr
+; CHECK: add
target datalayout = "e-p:32:32:32-i1:8:32-i8:8:32-i16:16:32-i32:32:32-i64:32:32-f32:32:32-f64:32:32-v64:32:64-v128:32:128-a0:0:32-n32"
target triple = "thumbv7-apple-ios"
diff --git a/llvm/test/CodeGen/ARM/machine-cse-cmp.ll b/llvm/test/CodeGen/ARM/machine-cse-cmp.ll
index e079f9fd9657..49dbb03135f5 100644
--- a/llvm/test/CodeGen/ARM/machine-cse-cmp.ll
+++ b/llvm/test/CodeGen/ARM/machine-cse-cmp.ll
@@ -52,7 +52,7 @@ entry:
; CHECK-LABEL: f3:
; CHECK-NOT: sub
; CHECK: cmp
-; CHECK: bge
+; CHECK: blt
%0 = load i32, i32* %offset, align 4
%cmp = icmp slt i32 %0, %size
%s = sub nsw i32 %0, %size
diff --git a/llvm/test/CodeGen/Hexagon/newvaluejump2.ll b/llvm/test/CodeGen/Hexagon/newvaluejump2.ll
index f03c7b407039..99c9d1a60af7 100644
--- a/llvm/test/CodeGen/Hexagon/newvaluejump2.ll
+++ b/llvm/test/CodeGen/Hexagon/newvaluejump2.ll
@@ -1,4 +1,3 @@
-; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc -march=hexagon -mcpu=hexagonv5 -disable-hexagon-misched < %s \
; RUN: | FileCheck %s
; Check that we generate new value jump, both registers, with one
@@ -6,35 +5,8 @@
@Reg = common global i32 0, align 4
define i32 @main() nounwind {
-; CHECK-LABEL: main:
-; CHECK: // %bb.0: // %entry
-; CHECK-NEXT: {
-; CHECK-NEXT: r1 = memw(gp+#Reg)
-; CHECK-NEXT: allocframe(r29,#8):raw
-; CHECK-NEXT: }
-; CHECK-NEXT: {
-; CHECK-NEXT: r0 = memw(r29+#4)
-; CHECK-NEXT: if (!cmp.gt(r0.new,r1)) jump:nt .LBB0_1
-; CHECK-NEXT: }
-; CHECK-NEXT: // %bb.2: // %if.else
-; CHECK-NEXT: {
-; CHECK-NEXT: call baz
-; CHECK-NEXT: r1:0 = combine(#20,#10)
-; CHECK-NEXT: }
-; CHECK-NEXT: {
-; CHECK-NEXT: r0 = #0
-; CHECK-NEXT: dealloc_return
-; CHECK-NEXT: }
-; CHECK-NEXT: .LBB0_1: // %if.then
-; CHECK-NEXT: {
-; CHECK-NEXT: call bar
-; CHECK-NEXT: r1:0 = combine(#2,#1)
-; CHECK-NEXT: }
-; CHECK-NEXT: {
-; CHECK-NEXT: r0 = #0
-; CHECK-NEXT: dealloc_return
-; CHECK-NEXT: }
entry:
+; CHECK: if (cmp.gt(r{{[0-9]+}}.new,r{{[0-9]+}})) jump:{{[t|nt]}} .LBB{{[0-9]+}}_{{[0-9]+}}
%Reg2 = alloca i32, align 4
%0 = load i32, i32* %Reg2, align 4
%1 = load i32, i32* @Reg, align 4
diff --git a/llvm/test/CodeGen/Mips/brcongt.ll b/llvm/test/CodeGen/Mips/brcongt.ll
index 223245bc622d..7dffdb411211 100644
--- a/llvm/test/CodeGen/Mips/brcongt.ll
+++ b/llvm/test/CodeGen/Mips/brcongt.ll
@@ -1,4 +1,3 @@
-; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc -march=mipsel -mattr=mips16 -relocation-model=pic -O3 < %s | FileCheck %s -check-prefix=16
@i = global i32 5, align 4
@@ -7,32 +6,14 @@
@result = global i32 0, align 4
define void @test() nounwind {
-; 16-LABEL: test:
-; 16: # %bb.0: # %entry
-; 16-NEXT: lui $2, %hi(_gp_disp)
-; 16-NEXT: addiu $2, $2, %lo(_gp_disp)
-; 16-NEXT: li $2, %hi(_gp_disp)
-; 16-NEXT: addiu $3, $pc, %lo(_gp_disp)
-; 16-NEXT: sll $2, $2, 16
-; 16-NEXT: addu $2, $3, $2
-; 16-NEXT: lw $3, %got(i)($2)
-; 16-NEXT: lw $4, %got(j)($2)
-; 16-NEXT: lw $3, 0($3)
-; 16-NEXT: lw $4, 0($4)
-; 16-NEXT: slt $4, $3
-; 16-NEXT: bteqz $BB0_2 # 16 bit inst
-; 16-NEXT: # %bb.1: # %if.end
-; 16-NEXT: jrc $ra
-; 16-NEXT: $BB0_2: # %if.then
-; 16-NEXT: lw $2, %got(result)($2)
-; 16-NEXT: li $3, 1
-; 16-NEXT: sw $3, 0($2)
-; 16-NEXT: jrc $ra
entry:
%0 = load i32, i32* @i, align 4
%1 = load i32, i32* @j, align 4
%cmp = icmp sgt i32 %0, %1
br i1 %cmp, label %if.end, label %if.then
+; 16: slt ${{[0-9]+}}, ${{[0-9]+}}
+; 16: btnez $[[LABEL:[0-9A-Ba-b_]+]]
+; 16: $[[LABEL]]:
if.then: ; preds = %entry
store i32 1, i32* @result, align 4
br label %if.end
diff --git a/llvm/test/CodeGen/Mips/brconlt.ll b/llvm/test/CodeGen/Mips/brconlt.ll
index 3b4ea80d0205..65f6c347b671 100644
--- a/llvm/test/CodeGen/Mips/brconlt.ll
+++ b/llvm/test/CodeGen/Mips/brconlt.ll
@@ -1,4 +1,3 @@
-; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc -march=mipsel -mattr=mips16 -relocation-model=pic -O3 < %s | FileCheck %s -check-prefix=16
; RUN: llc -march=mips -mattr=micromips -mcpu=mips32r6 -relocation-model=pic -O3 < %s | FileCheck %s -check-prefix=MM32R6
@@ -8,52 +7,16 @@
@result = global i32 0, align 4
define void @test() nounwind {
-; 16-LABEL: test:
-; 16: # %bb.0: # %entry
-; 16-NEXT: lui $2, %hi(_gp_disp)
-; 16-NEXT: addiu $2, $2, %lo(_gp_disp)
-; 16-NEXT: li $2, %hi(_gp_disp)
-; 16-NEXT: addiu $3, $pc, %lo(_gp_disp)
-; 16-NEXT: sll $2, $2, 16
-; 16-NEXT: addu $2, $3, $2
-; 16-NEXT: lw $3, %got(i)($2)
-; 16-NEXT: lw $4, %got(j)($2)
-; 16-NEXT: lw $3, 0($3)
-; 16-NEXT: lw $4, 0($4)
-; 16-NEXT: slt $4, $3
-; 16-NEXT: bteqz $BB0_2 # 16 bit inst
-; 16-NEXT: # %bb.1: # %if.end
-; 16-NEXT: jrc $ra
-; 16-NEXT: $BB0_2: # %if.then
-; 16-NEXT: lw $2, %got(result)($2)
-; 16-NEXT: li $3, 1
-; 16-NEXT: sw $3, 0($2)
-; 16-NEXT: jrc $ra
-;
-; MM32R6-LABEL: test:
-; MM32R6: # %bb.0: # %entry
-; MM32R6-NEXT: lui $2, %hi(_gp_disp)
-; MM32R6-NEXT: addiu $2, $2, %lo(_gp_disp)
-; MM32R6-NEXT: addu $2, $2, $25
-; MM32R6-NEXT: lw $3, %got(i)($2)
-; MM32R6-NEXT: lw $4, %got(j)($2)
-; MM32R6-NEXT: lw16 $3, 0($3)
-; MM32R6-NEXT: lw16 $4, 0($4)
-; MM32R6-NEXT: slt $1, $4, $3
-; MM32R6-NEXT: beqzc $1, $BB0_2
-; MM32R6-NEXT: # %bb.1: # %if.end
-; MM32R6-NEXT: jrc $ra
-; MM32R6-NEXT: $BB0_2: # %if.then
-; MM32R6-NEXT: lw $2, %got(result)($2)
-; MM32R6-NEXT: li16 $3, 1
-; MM32R6-NEXT: sw16 $3, 0($2)
-; MM32R6-NEXT: jrc $ra
entry:
%0 = load i32, i32* @j, align 4
%1 = load i32, i32* @i, align 4
%cmp = icmp slt i32 %0, %1
br i1 %cmp, label %if.end, label %if.then
+; 16: slt ${{[0-9]+}}, ${{[0-9]+}}
+; MM32R6: slt ${{[0-9]+}}, ${{[0-9]+}}
+; 16: btnez $[[LABEL:[0-9A-Ba-b_]+]]
+; 16: $[[LABEL]]:
if.then: ; preds = %entry
store i32 1, i32* @result, align 4
diff --git a/llvm/test/CodeGen/Mips/brconne.ll b/llvm/test/CodeGen/Mips/brconne.ll
index da11e842ada7..e0cbe378fe3c 100644
--- a/llvm/test/CodeGen/Mips/brconne.ll
+++ b/llvm/test/CodeGen/Mips/brconne.ll
@@ -1,4 +1,3 @@
-; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc -march=mipsel -mattr=mips16 -relocation-model=pic -O3 < %s | FileCheck %s -check-prefix=16
@i = global i32 5, align 4
@@ -6,32 +5,15 @@
@result = global i32 0, align 4
define void @test() nounwind {
-; 16-LABEL: test:
-; 16: # %bb.0: # %entry
-; 16-NEXT: lui $2, %hi(_gp_disp)
-; 16-NEXT: addiu $2, $2, %lo(_gp_disp)
-; 16-NEXT: li $2, %hi(_gp_disp)
-; 16-NEXT: addiu $3, $pc, %lo(_gp_disp)
-; 16-NEXT: sll $2, $2, 16
-; 16-NEXT: addu $2, $3, $2
-; 16-NEXT: lw $3, %got(i)($2)
-; 16-NEXT: lw $4, %got(j)($2)
-; 16-NEXT: lw $3, 0($3)
-; 16-NEXT: lw $4, 0($4)
-; 16-NEXT: cmp $4, $3
-; 16-NEXT: bteqz $BB0_2 # 16 bit inst
-; 16-NEXT: # %bb.1: # %if.end
-; 16-NEXT: jrc $ra
-; 16-NEXT: $BB0_2: # %if.then
-; 16-NEXT: lw $2, %got(result)($2)
-; 16-NEXT: li $3, 1
-; 16-NEXT: sw $3, 0($2)
-; 16-NEXT: jrc $ra
entry:
%0 = load i32, i32* @j, align 4
%1 = load i32, i32* @i, align 4
%cmp = icmp eq i32 %0, %1
br i1 %cmp, label %if.then, label %if.end
+; 16: cmp ${{[0-9]+}}, ${{[0-9]+}}
+; 16: btnez $[[LABEL:[0-9A-Ba-b_]+]]
+; 16: lw ${{[0-9]+}}, %got(result)(${{[0-9]+}})
+; 16: $[[LABEL]]:
if.then: ; preds = %entry
store i32 1, i32* @result, align 4
diff --git a/llvm/test/CodeGen/Mips/compactbranches/no-beqzc-bnezc.ll b/llvm/test/CodeGen/Mips/compactbranches/no-beqzc-bnezc.ll
index 27194ef77f7c..5e607c2639f3 100644
--- a/llvm/test/CodeGen/Mips/compactbranches/no-beqzc-bnezc.ll
+++ b/llvm/test/CodeGen/Mips/compactbranches/no-beqzc-bnezc.ll
@@ -1,57 +1,16 @@
-; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
-; RUN: llc -march=mipsel -mcpu=mips32r6 -disable-mips-delay-filler < %s | FileCheck %s --check-prefixes=ENCODING,MIPSEL32
-; RUN: llc -march=mips -mcpu=mips32r6 -disable-mips-delay-filler < %s | FileCheck %s --check-prefixes=ENCODING,MIPS32
-; RUN: llc -march=mipsel -mcpu=mips64r6 -disable-mips-delay-filler -target-abi=n64 < %s | FileCheck %s --check-prefixes=ENCODING,MIPSEL64
-; RUN: llc -march=mips -mcpu=mips64r6 -disable-mips-delay-filler -target-abi=n64 < %s | FileCheck %s --check-prefixes=ENCODING,MIPS64
+; RUN: llc -march=mipsel -mcpu=mips32r6 -disable-mips-delay-filler < %s | FileCheck %s
+; RUN: llc -march=mips -mcpu=mips32r6 -disable-mips-delay-filler < %s -filetype=obj \
+; RUN: -o - | llvm-objdump -d - | FileCheck %s --check-prefix=ENCODING
+; RUN: llc -march=mipsel -mcpu=mips64r6 -disable-mips-delay-filler -target-abi=n64 < %s | FileCheck %s
+; RUN: llc -march=mips -mcpu=mips64r6 -disable-mips-delay-filler -target-abi=n64 < %s -filetype=obj \
+; RUN: -o - | llvm-objdump -d - | FileCheck %s --check-prefix=ENCODING
; bnezc and beqzc have restriction that $rt != 0
define i32 @f() {
; CHECK-LABEL: f:
; CHECK-NOT: bnezc $0
-; MIPSEL32-LABEL: f:
-; MIPSEL32: # %bb.0:
-; MIPSEL32-NEXT: bnez $zero, $BB0_2
-; MIPSEL32-NEXT: nop
-; MIPSEL32-NEXT: # %bb.1: # %if.then
-; MIPSEL32-NEXT: addiu $2, $zero, 1
-; MIPSEL32-NEXT: jrc $ra
-; MIPSEL32-NEXT: $BB0_2: # %if.end
-; MIPSEL32-NEXT: addiu $2, $zero, 0
-; MIPSEL32-NEXT: jrc $ra
-;
-; MIPS32-LABEL: f:
-; MIPS32: # %bb.0:
-; MIPS32-NEXT: bnez $zero, $BB0_2
-; MIPS32-NEXT: nop
-; MIPS32-NEXT: # %bb.1: # %if.then
-; MIPS32-NEXT: addiu $2, $zero, 1
-; MIPS32-NEXT: jrc $ra
-; MIPS32-NEXT: $BB0_2: # %if.end
-; MIPS32-NEXT: addiu $2, $zero, 0
-; MIPS32-NEXT: jrc $ra
-;
-; MIPSEL64-LABEL: f:
-; MIPSEL64: # %bb.0:
-; MIPSEL64-NEXT: bnez $zero, .LBB0_2
-; MIPSEL64-NEXT: nop
-; MIPSEL64-NEXT: # %bb.1: # %if.then
-; MIPSEL64-NEXT: addiu $2, $zero, 1
-; MIPSEL64-NEXT: jrc $ra
-; MIPSEL64-NEXT: .LBB0_2: # %if.end
-; MIPSEL64-NEXT: addiu $2, $zero, 0
-; MIPSEL64-NEXT: jrc $ra
-;
-; MIPS64-LABEL: f:
-; MIPS64: # %bb.0:
-; MIPS64-NEXT: bnez $zero, .LBB0_2
-; MIPS64-NEXT: nop
-; MIPS64-NEXT: # %bb.1: # %if.then
-; MIPS64-NEXT: addiu $2, $zero, 1
-; MIPS64-NEXT: jrc $ra
-; MIPS64-NEXT: .LBB0_2: # %if.end
-; MIPS64-NEXT: addiu $2, $zero, 0
-; MIPS64-NEXT: jrc $ra
+
%cmp = icmp eq i32 1, 1
br i1 %cmp, label %if.then, label %if.end
@@ -65,49 +24,7 @@ define i32 @f() {
define i32 @f1() {
; CHECK-LABEL: f1:
; CHECK-NOT: beqzc $0
-; MIPSEL32-LABEL: f1:
-; MIPSEL32: # %bb.0:
-; MIPSEL32-NEXT: b $BB1_2
-; MIPSEL32-NEXT: nop
-; MIPSEL32-NEXT: # %bb.1: # %if.end
-; MIPSEL32-NEXT: addiu $2, $zero, 0
-; MIPSEL32-NEXT: jrc $ra
-; MIPSEL32-NEXT: $BB1_2: # %if.then
-; MIPSEL32-NEXT: addiu $2, $zero, 1
-; MIPSEL32-NEXT: jrc $ra
-;
-; MIPS32-LABEL: f1:
-; MIPS32: # %bb.0:
-; MIPS32-NEXT: b $BB1_2
-; MIPS32-NEXT: nop
-; MIPS32-NEXT: # %bb.1: # %if.end
-; MIPS32-NEXT: addiu $2, $zero, 0
-; MIPS32-NEXT: jrc $ra
-; MIPS32-NEXT: $BB1_2: # %if.then
-; MIPS32-NEXT: addiu $2, $zero, 1
-; MIPS32-NEXT: jrc $ra
-;
-; MIPSEL64-LABEL: f1:
-; MIPSEL64: # %bb.0:
-; MIPSEL64-NEXT: b .LBB1_2
-; MIPSEL64-NEXT: nop
-; MIPSEL64-NEXT: # %bb.1: # %if.end
-; MIPSEL64-NEXT: addiu $2, $zero, 0
-; MIPSEL64-NEXT: jrc $ra
-; MIPSEL64-NEXT: .LBB1_2: # %if.then
-; MIPSEL64-NEXT: addiu $2, $zero, 1
-; MIPSEL64-NEXT: jrc $ra
-;
-; MIPS64-LABEL: f1:
-; MIPS64: # %bb.0:
-; MIPS64-NEXT: b .LBB1_2
-; MIPS64-NEXT: nop
-; MIPS64-NEXT: # %bb.1: # %if.end
-; MIPS64-NEXT: addiu $2, $zero, 0
-; MIPS64-NEXT: jrc $ra
-; MIPS64-NEXT: .LBB1_2: # %if.then
-; MIPS64-NEXT: addiu $2, $zero, 1
-; MIPS64-NEXT: jrc $ra
+
%cmp = icmp eq i32 0, 0
br i1 %cmp, label %if.then, label %if.end
@@ -123,49 +40,9 @@ define i32 @f1() {
; beqc and bnec have the restriction that $rs < $rt.
define i32 @f2(i32 %a, i32 %b) {
-; MIPSEL32-LABEL: f2:
-; MIPSEL32: # %bb.0:
-; MIPSEL32-NEXT: beqc $5, $4, $BB2_2
-; MIPSEL32-NEXT: # %bb.1: # %if.end
-; MIPSEL32-NEXT: addiu $2, $zero, 0
-; MIPSEL32-NEXT: jrc $ra
-; MIPSEL32-NEXT: $BB2_2: # %if.then
-; MIPSEL32-NEXT: addiu $2, $zero, 1
-; MIPSEL32-NEXT: jrc $ra
-;
-; MIPS32-LABEL: f2:
-; MIPS32: # %bb.0:
-; MIPS32-NEXT: beqc $5, $4, $BB2_2
-; MIPS32-NEXT: # %bb.1: # %if.end
-; MIPS32-NEXT: addiu $2, $zero, 0
-; MIPS32-NEXT: jrc $ra
-; MIPS32-NEXT: $BB2_2: # %if.then
-; MIPS32-NEXT: addiu $2, $zero, 1
-; MIPS32-NEXT: jrc $ra
-;
-; MIPSEL64-LABEL: f2:
-; MIPSEL64: # %bb.0:
-; MIPSEL64-NEXT: sll $1, $4, 0
-; MIPSEL64-NEXT: sll $2, $5, 0
-; MIPSEL64-NEXT: beqc $2, $1, .LBB2_2
-; MIPSEL64-NEXT: # %bb.1: # %if.end
-; MIPSEL64-NEXT: addiu $2, $zero, 0
-; MIPSEL64-NEXT: jrc $ra
-; MIPSEL64-NEXT: .LBB2_2: # %if.then
-; MIPSEL64-NEXT: addiu $2, $zero, 1
-; MIPSEL64-NEXT: jrc $ra
-;
-; MIPS64-LABEL: f2:
-; MIPS64: # %bb.0:
-; MIPS64-NEXT: sll $1, $4, 0
-; MIPS64-NEXT: sll $2, $5, 0
-; MIPS64-NEXT: beqc $2, $1, .LBB2_2
-; MIPS64-NEXT: # %bb.1: # %if.end
-; MIPS64-NEXT: addiu $2, $zero, 0
-; MIPS64-NEXT: jrc $ra
-; MIPS64-NEXT: .LBB2_2: # %if.then
-; MIPS64-NEXT: addiu $2, $zero, 1
-; MIPS64-NEXT: jrc $ra
+; ENCODING-LABEL: <f2>:
+; ENCODING-NOT: beqc $5, $4
+; ENCODING-NOT: bnec $5, $4
%cmp = icmp eq i32 %b, %a
br i1 %cmp, label %if.then, label %if.end
@@ -180,53 +57,7 @@ define i32 @f2(i32 %a, i32 %b) {
define i64 @f3() {
; CHECK-LABEL: f3:
; CHECK-NOT: bnezc $0
-; MIPSEL32-LABEL: f3:
-; MIPSEL32: # %bb.0:
-; MIPSEL32-NEXT: bnez $zero, $BB3_2
-; MIPSEL32-NEXT: nop
-; MIPSEL32-NEXT: # %bb.1: # %if.then
-; MIPSEL32-NEXT: addiu $2, $zero, 1
-; MIPSEL32-NEXT: addiu $3, $zero, 0
-; MIPSEL32-NEXT: jrc $ra
-; MIPSEL32-NEXT: $BB3_2: # %if.end
-; MIPSEL32-NEXT: addiu $2, $zero, 0
-; MIPSEL32-NEXT: addiu $3, $zero, 0
-; MIPSEL32-NEXT: jrc $ra
-;
-; MIPS32-LABEL: f3:
-; MIPS32: # %bb.0:
-; MIPS32-NEXT: bnez $zero, $BB3_2
-; MIPS32-NEXT: nop
-; MIPS32-NEXT: # %bb.1: # %if.then
-; MIPS32-NEXT: addiu $2, $zero, 0
-; MIPS32-NEXT: addiu $3, $zero, 1
-; MIPS32-NEXT: jrc $ra
-; MIPS32-NEXT: $BB3_2: # %if.end
-; MIPS32-NEXT: addiu $2, $zero, 0
-; MIPS32-NEXT: addiu $3, $zero, 0
-; MIPS32-NEXT: jrc $ra
-;
-; MIPSEL64-LABEL: f3:
-; MIPSEL64: # %bb.0:
-; MIPSEL64-NEXT: bnez $zero, .LBB3_2
-; MIPSEL64-NEXT: nop
-; MIPSEL64-NEXT: # %bb.1: # %if.then
-; MIPSEL64-NEXT: daddiu $2, $zero, 1
-; MIPSEL64-NEXT: jrc $ra
-; MIPSEL64-NEXT: .LBB3_2: # %if.end
-; MIPSEL64-NEXT: daddiu $2, $zero, 0
-; MIPSEL64-NEXT: jrc $ra
-;
-; MIPS64-LABEL: f3:
-; MIPS64: # %bb.0:
-; MIPS64-NEXT: bnez $zero, .LBB3_2
-; MIPS64-NEXT: nop
-; MIPS64-NEXT: # %bb.1: # %if.then
-; MIPS64-NEXT: daddiu $2, $zero, 1
-; MIPS64-NEXT: jrc $ra
-; MIPS64-NEXT: .LBB3_2: # %if.end
-; MIPS64-NEXT: daddiu $2, $zero, 0
-; MIPS64-NEXT: jrc $ra
+
%cmp = icmp eq i64 1, 1
br i1 %cmp, label %if.then, label %if.end
@@ -240,53 +71,7 @@ define i64 @f3() {
define i64 @f4() {
; CHECK-LABEL: f4:
; CHECK-NOT: beqzc $0
-; MIPSEL32-LABEL: f4:
-; MIPSEL32: # %bb.0:
-; MIPSEL32-NEXT: b $BB4_2
-; MIPSEL32-NEXT: nop
-; MIPSEL32-NEXT: # %bb.1: # %if.end
-; MIPSEL32-NEXT: addiu $2, $zero, 0
-; MIPSEL32-NEXT: addiu $3, $zero, 0
-; MIPSEL32-NEXT: jrc $ra
-; MIPSEL32-NEXT: $BB4_2: # %if.then
-; MIPSEL32-NEXT: addiu $2, $zero, 1
-; MIPSEL32-NEXT: addiu $3, $zero, 0
-; MIPSEL32-NEXT: jrc $ra
-;
-; MIPS32-LABEL: f4:
-; MIPS32: # %bb.0:
-; MIPS32-NEXT: b $BB4_2
-; MIPS32-NEXT: nop
-; MIPS32-NEXT: # %bb.1: # %if.end
-; MIPS32-NEXT: addiu $2, $zero, 0
-; MIPS32-NEXT: addiu $3, $zero, 0
-; MIPS32-NEXT: jrc $ra
-; MIPS32-NEXT: $BB4_2: # %if.then
-; MIPS32-NEXT: addiu $2, $zero, 0
-; MIPS32-NEXT: addiu $3, $zero, 1
-; MIPS32-NEXT: jrc $ra
-;
-; MIPSEL64-LABEL: f4:
-; MIPSEL64: # %bb.0:
-; MIPSEL64-NEXT: b .LBB4_2
-; MIPSEL64-NEXT: nop
-; MIPSEL64-NEXT: # %bb.1: # %if.end
-; MIPSEL64-NEXT: daddiu $2, $zero, 0
-; MIPSEL64-NEXT: jrc $ra
-; MIPSEL64-NEXT: .LBB4_2: # %if.then
-; MIPSEL64-NEXT: daddiu $2, $zero, 1
-; MIPSEL64-NEXT: jrc $ra
-;
-; MIPS64-LABEL: f4:
-; MIPS64: # %bb.0:
-; MIPS64-NEXT: b .LBB4_2
-; MIPS64-NEXT: nop
-; MIPS64-NEXT: # %bb.1: # %if.end
-; MIPS64-NEXT: daddiu $2, $zero, 0
-; MIPS64-NEXT: jrc $ra
-; MIPS64-NEXT: .LBB4_2: # %if.then
-; MIPS64-NEXT: daddiu $2, $zero, 1
-; MIPS64-NEXT: jrc $ra
+
%cmp = icmp eq i64 0, 0
br i1 %cmp, label %if.then, label %if.end
@@ -302,55 +87,9 @@ define i64 @f4() {
; beqc and bnec have the restriction that $rs < $rt.
define i64 @f5(i64 %a, i64 %b) {
-; MIPSEL32-LABEL: f5:
-; MIPSEL32: # %bb.0:
-; MIPSEL32-NEXT: xor $1, $7, $5
-; MIPSEL32-NEXT: xor $2, $6, $4
-; MIPSEL32-NEXT: or $1, $2, $1
-; MIPSEL32-NEXT: beqzc $1, $BB5_2
-; MIPSEL32-NEXT: # %bb.1: # %if.end
-; MIPSEL32-NEXT: addiu $2, $zero, 0
-; MIPSEL32-NEXT: addiu $3, $zero, 0
-; MIPSEL32-NEXT: jrc $ra
-; MIPSEL32-NEXT: $BB5_2: # %if.then
-; MIPSEL32-NEXT: addiu $2, $zero, 1
-; MIPSEL32-NEXT: addiu $3, $zero, 0
-; MIPSEL32-NEXT: jrc $ra
-;
-; MIPS32-LABEL: f5:
-; MIPS32: # %bb.0:
-; MIPS32-NEXT: xor $1, $6, $4
-; MIPS32-NEXT: xor $2, $7, $5
-; MIPS32-NEXT: or $1, $2, $1
-; MIPS32-NEXT: beqzc $1, $BB5_2
-; MIPS32-NEXT: # %bb.1: # %if.end
-; MIPS32-NEXT: addiu $2, $zero, 0
-; MIPS32-NEXT: addiu $3, $zero, 0
-; MIPS32-NEXT: jrc $ra
-; MIPS32-NEXT: $BB5_2: # %if.then
-; MIPS32-NEXT: addiu $2, $zero, 0
-; MIPS32-NEXT: addiu $3, $zero, 1
-; MIPS32-NEXT: jrc $ra
-;
-; MIPSEL64-LABEL: f5:
-; MIPSEL64: # %bb.0:
-; MIPSEL64-NEXT: beqc $5, $4, .LBB5_2
-; MIPSEL64-NEXT: # %bb.1: # %if.end
-; MIPSEL64-NEXT: daddiu $2, $zero, 0
-; MIPSEL64-NEXT: jrc $ra
-; MIPSEL64-NEXT: .LBB5_2: # %if.then
-; MIPSEL64-NEXT: daddiu $2, $zero, 1
-; MIPSEL64-NEXT: jrc $ra
-;
-; MIPS64-LABEL: f5:
-; MIPS64: # %bb.0:
-; MIPS64-NEXT: beqc $5, $4, .LBB5_2
-; MIPS64-NEXT: # %bb.1: # %if.end
-; MIPS64-NEXT: daddiu $2, $zero, 0
-; MIPS64-NEXT: jrc $ra
-; MIPS64-NEXT: .LBB5_2: # %if.then
-; MIPS64-NEXT: daddiu $2, $zero, 1
-; MIPS64-NEXT: jrc $ra
+; ENCODING-LABEL: <f5>:
+; ENCODING-NOT: beqc $5, $4
+; ENCODING-NOT: bnec $5, $4
%cmp = icmp eq i64 %b, %a
br i1 %cmp, label %if.then, label %if.end
@@ -365,47 +104,7 @@ define i64 @f5(i64 %a, i64 %b) {
define i32 @f6(i32 %a) {
; CHECK-LABEL: f6:
; CHECK: beqzc ${{[0-9]+}}, {{((\$)|(\.L))}}BB
-; MIPSEL32-LABEL: f6:
-; MIPSEL32: # %bb.0:
-; MIPSEL32-NEXT: beqzc $4, $BB6_2
-; MIPSEL32-NEXT: # %bb.1: # %if.end
-; MIPSEL32-NEXT: addiu $2, $zero, 0
-; MIPSEL32-NEXT: jrc $ra
-; MIPSEL32-NEXT: $BB6_2: # %if.then
-; MIPSEL32-NEXT: addiu $2, $zero, 1
-; MIPSEL32-NEXT: jrc $ra
-;
-; MIPS32-LABEL: f6:
-; MIPS32: # %bb.0:
-; MIPS32-NEXT: beqzc $4, $BB6_2
-; MIPS32-NEXT: # %bb.1: # %if.end
-; MIPS32-NEXT: addiu $2, $zero, 0
-; MIPS32-NEXT: jrc $ra
-; MIPS32-NEXT: $BB6_2: # %if.then
-; MIPS32-NEXT: addiu $2, $zero, 1
-; MIPS32-NEXT: jrc $ra
-;
-; MIPSEL64-LABEL: f6:
-; MIPSEL64: # %bb.0:
-; MIPSEL64-NEXT: sll $1, $4, 0
-; MIPSEL64-NEXT: beqzc $1, .LBB6_2
-; MIPSEL64-NEXT: # %bb.1: # %if.end
-; MIPSEL64-NEXT: addiu $2, $zero, 0
-; MIPSEL64-NEXT: jrc $ra
-; MIPSEL64-NEXT: .LBB6_2: # %if.then
-; MIPSEL64-NEXT: addiu $2, $zero, 1
-; MIPSEL64-NEXT: jrc $ra
-;
-; MIPS64-LABEL: f6:
-; MIPS64: # %bb.0:
-; MIPS64-NEXT: sll $1, $4, 0
-; MIPS64-NEXT: beqzc $1, .LBB6_2
-; MIPS64-NEXT: # %bb.1: # %if.end
-; MIPS64-NEXT: addiu $2, $zero, 0
-; MIPS64-NEXT: jrc $ra
-; MIPS64-NEXT: .LBB6_2: # %if.then
-; MIPS64-NEXT: addiu $2, $zero, 1
-; MIPS64-NEXT: jrc $ra
+
%cmp = icmp eq i32 %a, 0
br i1 %cmp, label %if.then, label %if.end
@@ -419,47 +118,7 @@ define i32 @f6(i32 %a) {
define i32 @f7(i32 %a) {
; CHECK-LABEL: f7:
; CHECK: bnezc ${{[0-9]+}}, {{((\$)|(\.L))}}BB
-; MIPSEL32-LABEL: f7:
-; MIPSEL32: # %bb.0:
-; MIPSEL32-NEXT: beqzc $4, $BB7_2
-; MIPSEL32-NEXT: # %bb.1: # %if.end
-; MIPSEL32-NEXT: addiu $2, $zero, 0
-; MIPSEL32-NEXT: jrc $ra
-; MIPSEL32-NEXT: $BB7_2: # %if.then
-; MIPSEL32-NEXT: addiu $2, $zero, 1
-; MIPSEL32-NEXT: jrc $ra
-;
-; MIPS32-LABEL: f7:
-; MIPS32: # %bb.0:
-; MIPS32-NEXT: beqzc $4, $BB7_2
-; MIPS32-NEXT: # %bb.1: # %if.end
-; MIPS32-NEXT: addiu $2, $zero, 0
-; MIPS32-NEXT: jrc $ra
-; MIPS32-NEXT: $BB7_2: # %if.then
-; MIPS32-NEXT: addiu $2, $zero, 1
-; MIPS32-NEXT: jrc $ra
-;
-; MIPSEL64-LABEL: f7:
-; MIPSEL64: # %bb.0:
-; MIPSEL64-NEXT: sll $1, $4, 0
-; MIPSEL64-NEXT: beqzc $1, .LBB7_2
-; MIPSEL64-NEXT: # %bb.1: # %if.end
-; MIPSEL64-NEXT: addiu $2, $zero, 0
-; MIPSEL64-NEXT: jrc $ra
-; MIPSEL64-NEXT: .LBB7_2: # %if.then
-; MIPSEL64-NEXT: addiu $2, $zero, 1
-; MIPSEL64-NEXT: jrc $ra
-;
-; MIPS64-LABEL: f7:
-; MIPS64: # %bb.0:
-; MIPS64-NEXT: sll $1, $4, 0
-; MIPS64-NEXT: beqzc $1, .LBB7_2
-; MIPS64-NEXT: # %bb.1: # %if.end
-; MIPS64-NEXT: addiu $2, $zero, 0
-; MIPS64-NEXT: jrc $ra
-; MIPS64-NEXT: .LBB7_2: # %if.then
-; MIPS64-NEXT: addiu $2, $zero, 1
-; MIPS64-NEXT: jrc $ra
+
%cmp = icmp eq i32 0, %a
br i1 %cmp, label %if.then, label %if.end
diff --git a/llvm/test/CodeGen/Mips/lcb2.ll b/llvm/test/CodeGen/Mips/lcb2.ll
index 6a0be713c47f..4987c606e330 100644
--- a/llvm/test/CodeGen/Mips/lcb2.ll
+++ b/llvm/test/CodeGen/Mips/lcb2.ll
@@ -1,4 +1,3 @@
-; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc -mtriple=mipsel-linux-gnu -march=mipsel -mattr=mips16 -mattr=+soft-float -mips16-hard-float -relocation-model=static -mips16-constant-islands=true < %s | FileCheck %s -check-prefix=lcb
; RUN: llc -mtriple=mipsel-linux-gnu -march=mipsel -mattr=mips16 -mattr=+soft-float -mips16-hard-float -relocation-model=static -mips16-constant-islands=true < %s | FileCheck %s -check-prefix=lcbn
@@ -9,51 +8,6 @@
; Function Attrs: nounwind optsize
define i32 @bnez() #0 {
-; lcb-LABEL: bnez:
-; lcb: # %bb.0: # %entry
-; lcb-NEXT: li $2, %hi(i)
-; lcb-NEXT: sll $2, $2, 16
-; lcb-NEXT: lw $3, %lo(i)($2)
-; lcb-NEXT: bnez $3, $BB0_2
-; lcb-NEXT: # %bb.1: # %if.then
-; lcb-NEXT: li $3, 0
-; lcb-NEXT: #APP
-; lcb-NEXT: .set push
-; lcb-NEXT: .set at
-; lcb-NEXT: .set macro
-; lcb-NEXT: .set reorder
-; lcb-EMPTY:
-; lcb-NEXT: .space 10000
-; lcb-EMPTY:
-; lcb-NEXT: .set pop
-; lcb-NEXT: #NO_APP
-; lcb-NEXT: sw $3, %lo(i)($2)
-; lcb-NEXT: $BB0_2: # %if.end
-; lcb-NEXT: li $2, 0
-; lcb-NEXT: jrc $ra
-;
-; lcbn-LABEL: bnez:
-; lcbn: # %bb.0: # %entry
-; lcbn-NEXT: li $2, %hi(i)
-; lcbn-NEXT: sll $2, $2, 16
-; lcbn-NEXT: lw $3, %lo(i)($2)
-; lcbn-NEXT: bnez $3, $BB0_2
-; lcbn-NEXT: # %bb.1: # %if.then
-; lcbn-NEXT: li $3, 0
-; lcbn-NEXT: #APP
-; lcbn-NEXT: .set push
-; lcbn-NEXT: .set at
-; lcbn-NEXT: .set macro
-; lcbn-NEXT: .set reorder
-; lcbn-EMPTY:
-; lcbn-NEXT: .space 10000
-; lcbn-EMPTY:
-; lcbn-NEXT: .set pop
-; lcbn-NEXT: #NO_APP
-; lcbn-NEXT: sw $3, %lo(i)($2)
-; lcbn-NEXT: $BB0_2: # %if.end
-; lcbn-NEXT: li $2, 0
-; lcbn-NEXT: jrc $ra
entry:
%0 = load i32, i32* @i, align 4, !tbaa !1
%cmp = icmp eq i32 %0, 0
@@ -67,90 +21,15 @@ if.then: ; preds = %entry
if.end: ; preds = %if.then, %entry
ret i32 0
}
+; lcb: .ent bnez
+; lcbn: .ent bnez
+; lcb: bnez ${{[0-9]+}}, $BB{{[0-9]+}}_{{[0-9]+}}
+; lcbn-NOT: bnez ${{[0-9]+}}, $BB{{[0-9]+}}_{{[0-9]+}} # 16 bit inst
+; lcb: .end bnez
+; lcbn: .end bnez
; Function Attrs: nounwind optsize
define i32 @beqz() #0 {
-; lcb-LABEL: beqz:
-; lcb: # %bb.0: # %entry
-; lcb-NEXT: li $2, %hi(i)
-; lcb-NEXT: sll $2, $2, 16
-; lcb-NEXT: lw $2, %lo(i)($2)
-; lcb-NEXT: beqz $2, $BB1_2
-; lcb-NEXT: # %bb.1: # %if.else
-; lcb-NEXT: li $2, %hi(j)
-; lcb-NEXT: sll $2, $2, 16
-; lcb-NEXT: li $3, 55
-; lcb-NEXT: sw $3, %lo(j)($2)
-; lcb-NEXT: #APP
-; lcb-NEXT: .set push
-; lcb-NEXT: .set at
-; lcb-NEXT: .set macro
-; lcb-NEXT: .set reorder
-; lcb-EMPTY:
-; lcb-NEXT: .space 10000
-; lcb-EMPTY:
-; lcb-NEXT: .set pop
-; lcb-NEXT: #NO_APP
-; lcb-NEXT: b $BB1_3
-; lcb-NEXT: $BB1_2: # %if.then
-; lcb-NEXT: li $2, %hi(j)
-; lcb-NEXT: sll $2, $2, 16
-; lcb-NEXT: li $3, 10
-; lcb-NEXT: sw $3, %lo(j)($2)
-; lcb-NEXT: #APP
-; lcb-NEXT: .set push
-; lcb-NEXT: .set at
-; lcb-NEXT: .set macro
-; lcb-NEXT: .set reorder
-; lcb-EMPTY:
-; lcb-NEXT: .space 10000
-; lcb-EMPTY:
-; lcb-NEXT: .set pop
-; lcb-NEXT: #NO_APP
-; lcb-NEXT: $BB1_3: # %if.end
-; lcb-NEXT: li $2, 0
-; lcb-NEXT: jrc $ra
-;
-; lcbn-LABEL: beqz:
-; lcbn: # %bb.0: # %entry
-; lcbn-NEXT: li $2, %hi(i)
-; lcbn-NEXT: sll $2, $2, 16
-; lcbn-NEXT: lw $2, %lo(i)($2)
-; lcbn-NEXT: beqz $2, $BB1_2
-; lcbn-NEXT: # %bb.1: # %if.else
-; lcbn-NEXT: li $2, %hi(j)
-; lcbn-NEXT: sll $2, $2, 16
-; lcbn-NEXT: li $3, 55
-; lcbn-NEXT: sw $3, %lo(j)($2)
-; lcbn-NEXT: #APP
-; lcbn-NEXT: .set push
-; lcbn-NEXT: .set at
-; lcbn-NEXT: .set macro
-; lcbn-NEXT: .set reorder
-; lcbn-EMPTY:
-; lcbn-NEXT: .space 10000
-; lcbn-EMPTY:
-; lcbn-NEXT: .set pop
-; lcbn-NEXT: #NO_APP
-; lcbn-NEXT: b $BB1_3
-; lcbn-NEXT: $BB1_2: # %if.then
-; lcbn-NEXT: li $2, %hi(j)
-; lcbn-NEXT: sll $2, $2, 16
-; lcbn-NEXT: li $3, 10
-; lcbn-NEXT: sw $3, %lo(j)($2)
-; lcbn-NEXT: #APP
-; lcbn-NEXT: .set push
-; lcbn-NEXT: .set at
-; lcbn-NEXT: .set macro
-; lcbn-NEXT: .set reorder
-; lcbn-EMPTY:
-; lcbn-NEXT: .space 10000
-; lcbn-EMPTY:
-; lcbn-NEXT: .set pop
-; lcbn-NEXT: #NO_APP
-; lcbn-NEXT: $BB1_3: # %if.end
-; lcbn-NEXT: li $2, 0
-; lcbn-NEXT: jrc $ra
entry:
%0 = load i32, i32* @i, align 4, !tbaa !1
%cmp = icmp eq i32 %0, 0
@@ -170,93 +49,16 @@ if.end: ; preds = %if.else, %if.then
ret i32 0
}
+; lcb: .ent beqz
+; lcbn: .ent beqz
+; lcb: beqz ${{[0-9]+}}, $BB{{[0-9]+}}_{{[0-9]+}}
+; lcbn-NOT: beqz ${{[0-9]+}}, $BB{{[0-9]+}}_{{[0-9]+}} # 16 bit inst
+; lcb: .end beqz
+; lcbn: .end beqz
+
+
; Function Attrs: nounwind optsize
define void @bteqz() #0 {
-; lcb-LABEL: bteqz:
-; lcb: # %bb.0: # %entry
-; lcb-NEXT: li $2, %hi(j)
-; lcb-NEXT: sll $2, $2, 16
-; lcb-NEXT: lw $2, %lo(j)($2)
-; lcb-NEXT: li $3, %hi(i)
-; lcb-NEXT: sll $3, $3, 16
-; lcb-NEXT: lw $3, %lo(i)($3)
-; lcb-NEXT: cmp $3, $2
-; lcb-NEXT: bteqz $BB2_2
-; lcb-NEXT: # %bb.1: # %if.else
-; lcb-NEXT: li $2, %hi(k)
-; lcb-NEXT: sll $2, $2, 16
-; lcb-NEXT: #APP
-; lcb-NEXT: .set push
-; lcb-NEXT: .set at
-; lcb-NEXT: .set macro
-; lcb-NEXT: .set reorder
-; lcb-EMPTY:
-; lcb-NEXT: .space 1000
-; lcb-EMPTY:
-; lcb-NEXT: .set pop
-; lcb-NEXT: #NO_APP
-; lcb-NEXT: li $3, 2
-; lcb-NEXT: sw $3, %lo(k)($2)
-; lcb-NEXT: jrc $ra
-; lcb-NEXT: $BB2_2: # %if.then
-; lcb-NEXT: li $2, %hi(k)
-; lcb-NEXT: sll $2, $2, 16
-; lcb-NEXT: li $3, 1
-; lcb-NEXT: sw $3, %lo(k)($2)
-; lcb-NEXT: #APP
-; lcb-NEXT: .set push
-; lcb-NEXT: .set at
-; lcb-NEXT: .set macro
-; lcb-NEXT: .set reorder
-; lcb-EMPTY:
-; lcb-NEXT: .space 1000
-; lcb-EMPTY:
-; lcb-NEXT: .set pop
-; lcb-NEXT: #NO_APP
-; lcb-NEXT: jrc $ra
-;
-; lcbn-LABEL: bteqz:
-; lcbn: # %bb.0: # %entry
-; lcbn-NEXT: li $2, %hi(j)
-; lcbn-NEXT: sll $2, $2, 16
-; lcbn-NEXT: lw $2, %lo(j)($2)
-; lcbn-NEXT: li $3, %hi(i)
-; lcbn-NEXT: sll $3, $3, 16
-; lcbn-NEXT: lw $3, %lo(i)($3)
-; lcbn-NEXT: cmp $3, $2
-; lcbn-NEXT: bteqz $BB2_2
-; lcbn-NEXT: # %bb.1: # %if.else
-; lcbn-NEXT: li $2, %hi(k)
-; lcbn-NEXT: sll $2, $2, 16
-; lcbn-NEXT: #APP
-; lcbn-NEXT: .set push
-; lcbn-NEXT: .set at
-; lcbn-NEXT: .set macro
-; lcbn-NEXT: .set reorder
-; lcbn-EMPTY:
-; lcbn-NEXT: .space 1000
-; lcbn-EMPTY:
-; lcbn-NEXT: .set pop
-; lcbn-NEXT: #NO_APP
-; lcbn-NEXT: li $3, 2
-; lcbn-NEXT: sw $3, %lo(k)($2)
-; lcbn-NEXT: jrc $ra
-; lcbn-NEXT: $BB2_2: # %if.then
-; lcbn-NEXT: li $2, %hi(k)
-; lcbn-NEXT: sll $2, $2, 16
-; lcbn-NEXT: li $3, 1
-; lcbn-NEXT: sw $3, %lo(k)($2)
-; lcbn-NEXT: #APP
-; lcbn-NEXT: .set push
-; lcbn-NEXT: .set at
-; lcbn-NEXT: .set macro
-; lcbn-NEXT: .set reorder
-; lcbn-EMPTY:
-; lcbn-NEXT: .space 1000
-; lcbn-EMPTY:
-; lcbn-NEXT: .set pop
-; lcbn-NEXT: #NO_APP
-; lcbn-NEXT: jrc $ra
entry:
%0 = load i32, i32* @i, align 4, !tbaa !1
%1 = load i32, i32* @j, align 4, !tbaa !1
@@ -277,65 +79,16 @@ if.end: ; preds = %if.else, %if.then
ret void
}
+; lcb: .ent bteqz
+; lcbn: .ent bteqz
+; lcb: btnez $BB{{[0-9]+}}_{{[0-9]+}}
+; lcbn-NOT: btnez $BB{{[0-9]+}}_{{[0-9]+}} # 16 bit inst
+; lcb: .end bteqz
+; lcbn: .end bteqz
+
+
; Function Attrs: nounwind optsize
define void @btz() #0 {
-; lcb-LABEL: btz:
-; lcb: # %bb.0: # %entry
-; lcb-NEXT: li $2, %hi(i)
-; lcb-NEXT: sll $2, $2, 16
-; lcb-NEXT: lw $4, %lo(i)($2)
-; lcb-NEXT: li $3, %hi(j)
-; lcb-NEXT: sll $3, $3, 16
-; lcb-NEXT: lw $5, %lo(j)($3)
-; lcb-NEXT: slt $5, $4
-; lcb-NEXT: bteqz $BB3_2
-; lcb-NEXT: $BB3_1: # %if.then
-; lcb-NEXT: # =>This Inner Loop Header: Depth=1
-; lcb-NEXT: #APP
-; lcb-NEXT: .set push
-; lcb-NEXT: .set at
-; lcb-NEXT: .set macro
-; lcb-NEXT: .set reorder
-; lcb-EMPTY:
-; lcb-NEXT: .space 60000
-; lcb-EMPTY:
-; lcb-NEXT: .set pop
-; lcb-NEXT: #NO_APP
-; lcb-NEXT: lw $4, %lo(i)($2)
-; lcb-NEXT: lw $5, %lo(j)($3)
-; lcb-NEXT: slt $5, $4
-; lcb-NEXT: btnez $BB3_1
-; lcb-NEXT: $BB3_2: # %if.end
-; lcb-NEXT: jrc $ra
-;
-; lcbn-LABEL: btz:
-; lcbn: # %bb.0: # %entry
-; lcbn-NEXT: li $2, %hi(i)
-; lcbn-NEXT: sll $2, $2, 16
-; lcbn-NEXT: lw $4, %lo(i)($2)
-; lcbn-NEXT: li $3, %hi(j)
-; lcbn-NEXT: sll $3, $3, 16
-; lcbn-NEXT: lw $5, %lo(j)($3)
-; lcbn-NEXT: slt $5, $4
-; lcbn-NEXT: bteqz $BB3_2
-; lcbn-NEXT: $BB3_1: # %if.then
-; lcbn-NEXT: # =>This Inner Loop Header: Depth=1
-; lcbn-NEXT: #APP
-; lcbn-NEXT: .set push
-; lcbn-NEXT: .set at
-; lcbn-NEXT: .set macro
-; lcbn-NEXT: .set reorder
-; lcbn-EMPTY:
-; lcbn-NEXT: .space 60000
-; lcbn-EMPTY:
-; lcbn-NEXT: .set pop
-; lcbn-NEXT: #NO_APP
-; lcbn-NEXT: lw $4, %lo(i)($2)
-; lcbn-NEXT: lw $5, %lo(j)($3)
-; lcbn-NEXT: slt $5, $4
-; lcbn-NEXT: btnez $BB3_1
-; lcbn-NEXT: $BB3_2: # %if.end
-; lcbn-NEXT: jrc $ra
entry:
%0 = load i32, i32* @i, align 4, !tbaa !1
%1 = load i32, i32* @j, align 4, !tbaa !1
@@ -353,6 +106,15 @@ if.end: ; preds = %if.then, %entry
ret void
}
+; lcb: .ent btz
+; lcbn: .ent btz
+; lcb: bteqz $BB{{[0-9]+}}_{{[0-9]+}}
+; lcbn-NOT: bteqz $BB{{[0-9]+}}_{{[0-9]+}} # 16 bit inst
+; lcb: btnez $BB{{[0-9]+}}_{{[0-9]+}}
+; lcbn-NOT: btnez $BB{{[0-9]+}}_{{[0-9]+}} # 16 bit inst
+; lcb: .end btz
+; lcbn: .end btz
+
attributes #0 = { nounwind optsize "less-precise-fpmad"="false" "frame-pointer"="none" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" }
attributes #1 = { nounwind }
diff --git a/llvm/test/CodeGen/Mips/lcb5.ll b/llvm/test/CodeGen/Mips/lcb5.ll
index 6b8ebfd07e6f..96e924a44f72 100644
--- a/llvm/test/CodeGen/Mips/lcb5.ll
+++ b/llvm/test/CodeGen/Mips/lcb5.ll
@@ -1,4 +1,3 @@
-; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc -mtriple=mipsel-linux-gnu -march=mipsel -mattr=mips16 -mattr=+soft-float -mips16-hard-float -relocation-model=static < %s | FileCheck %s -check-prefix=ci
@i = global i32 0, align 4
@@ -7,41 +6,6 @@
; Function Attrs: nounwind optsize
define i32 @x0() #0 {
-; ci-LABEL: x0:
-; ci: # %bb.0: # %entry
-; ci-NEXT: li $2, %hi(i)
-; ci-NEXT: sll $2, $2, 16
-; ci-NEXT: lw $3, %lo(i)($2)
-; ci-NEXT: beqz $3, $BB0_2
-; ci-NEXT: # %bb.1: # %if.else
-; ci-NEXT: li $3, 1
-; ci-NEXT: #APP
-; ci-NEXT: .set push
-; ci-NEXT: .set at
-; ci-NEXT: .set macro
-; ci-NEXT: .set reorder
-; ci-EMPTY:
-; ci-NEXT: .space 1004
-; ci-EMPTY:
-; ci-NEXT: .set pop
-; ci-NEXT: #NO_APP
-; ci-NEXT: b $BB0_3 # 16 bit inst
-; ci-NEXT: $BB0_2: # %if.then
-; ci-NEXT: li $3, 0
-; ci-NEXT: #APP
-; ci-NEXT: .set push
-; ci-NEXT: .set at
-; ci-NEXT: .set macro
-; ci-NEXT: .set reorder
-; ci-EMPTY:
-; ci-NEXT: .space 1000
-; ci-EMPTY:
-; ci-NEXT: .set pop
-; ci-NEXT: #NO_APP
-; ci-NEXT: $BB0_3: # %if.end
-; ci-NEXT: sw $3, %lo(i)($2)
-; ci-NEXT: li $2, 0
-; ci-NEXT: jrc $ra
entry:
%0 = load i32, i32* @i, align 4, !tbaa !1
%cmp = icmp eq i32 %0, 0
@@ -61,48 +25,13 @@ if.end: ; preds = %if.else, %if.then
ret i32 0
}
+; ci: .ent x0
+; ci: beqz $3, $BB0_2
+; ci: $BB0_2:
+; ci: .end x0
+
; Function Attrs: nounwind optsize
define i32 @x1() #0 {
-; ci-LABEL: x1:
-; ci: # %bb.0: # %entry
-; ci-NEXT: li $2, %hi(i)
-; ci-NEXT: sll $2, $2, 16
-; ci-NEXT: lw $3, %lo(i)($2)
-; ci-NEXT: bnez $3, $BB1_1 # 16 bit inst
-; ci-NEXT: jal $BB1_2 # branch
-; ci-NEXT: nop
-; ci-NEXT: $BB1_1: # %if.else
-; ci-NEXT: li $3, 1
-; ci-NEXT: #APP
-; ci-NEXT: .set push
-; ci-NEXT: .set at
-; ci-NEXT: .set macro
-; ci-NEXT: .set reorder
-; ci-EMPTY:
-; ci-NEXT: .space 1000004
-; ci-EMPTY:
-; ci-NEXT: .set pop
-; ci-NEXT: #NO_APP
-; ci-NEXT: jal $BB1_3 # branch
-; ci-NEXT: nop
-; ci-NEXT: .p2align 2
-; ci-NEXT: $BB1_2: # %if.then
-; ci-NEXT: li $3, 0
-; ci-NEXT: #APP
-; ci-NEXT: .set push
-; ci-NEXT: .set at
-; ci-NEXT: .set macro
-; ci-NEXT: .set reorder
-; ci-EMPTY:
-; ci-NEXT: .space 1000000
-; ci-EMPTY:
-; ci-NEXT: .set pop
-; ci-NEXT: #NO_APP
-; ci-NEXT: .p2align 2
-; ci-NEXT: $BB1_3: # %if.end
-; ci-NEXT: sw $3, %lo(i)($2)
-; ci-NEXT: li $2, 0
-; ci-NEXT: jrc $ra
entry:
%0 = load i32, i32* @i, align 4, !tbaa !1
%cmp = icmp eq i32 %0, 0
@@ -122,49 +51,15 @@ if.end: ; preds = %if.else, %if.then
ret i32 0
}
+; ci: .ent x1
+; ci: bnez $3, $BB1_1 # 16 bit inst
+; ci: jal $BB1_2 # branch
+; ci: nop
+; ci: $BB1_1:
+; ci: .end x1
; Function Attrs: nounwind optsize
define i32 @y0() #0 {
-; ci-LABEL: y0:
-; ci: # %bb.0: # %entry
-; ci-NEXT: li $2, %hi(i)
-; ci-NEXT: sll $2, $2, 16
-; ci-NEXT: lw $2, %lo(i)($2)
-; ci-NEXT: beqz $2, $BB2_2
-; ci-NEXT: # %bb.1: # %if.else
-; ci-NEXT: li $2, %hi(j)
-; ci-NEXT: sll $2, $2, 16
-; ci-NEXT: li $3, 55
-; ci-NEXT: sw $3, %lo(j)($2)
-; ci-NEXT: #APP
-; ci-NEXT: .set push
-; ci-NEXT: .set at
-; ci-NEXT: .set macro
-; ci-NEXT: .set reorder
-; ci-EMPTY:
-; ci-NEXT: .space 1004
-; ci-EMPTY:
-; ci-NEXT: .set pop
-; ci-NEXT: #NO_APP
-; ci-NEXT: b $BB2_3 # 16 bit inst
-; ci-NEXT: $BB2_2: # %if.then
-; ci-NEXT: li $2, %hi(j)
-; ci-NEXT: sll $2, $2, 16
-; ci-NEXT: li $3, 10
-; ci-NEXT: sw $3, %lo(j)($2)
-; ci-NEXT: #APP
-; ci-NEXT: .set push
-; ci-NEXT: .set at
-; ci-NEXT: .set macro
-; ci-NEXT: .set reorder
-; ci-EMPTY:
-; ci-NEXT: .space 1000
-; ci-EMPTY:
-; ci-NEXT: .set pop
-; ci-NEXT: #NO_APP
-; ci-NEXT: $BB2_3: # %if.end
-; ci-NEXT: li $2, 0
-; ci-NEXT: jrc $ra
entry:
%0 = load i32, i32* @i, align 4, !tbaa !1
%cmp = icmp eq i32 %0, 0
@@ -184,53 +79,12 @@ if.end: ; preds = %if.else, %if.then
ret i32 0
}
+; ci: .ent y0
+; ci: beqz $2, $BB2_2
+; ci: .end y0
+
; Function Attrs: nounwind optsize
define i32 @y1() #0 {
-; ci-LABEL: y1:
-; ci: # %bb.0: # %entry
-; ci-NEXT: li $2, %hi(i)
-; ci-NEXT: sll $2, $2, 16
-; ci-NEXT: lw $2, %lo(i)($2)
-; ci-NEXT: bnez $2, $BB3_1 # 16 bit inst
-; ci-NEXT: jal $BB3_2 # branch
-; ci-NEXT: nop
-; ci-NEXT: $BB3_1: # %if.else
-; ci-NEXT: li $2, %hi(j)
-; ci-NEXT: sll $2, $2, 16
-; ci-NEXT: li $3, 55
-; ci-NEXT: sw $3, %lo(j)($2)
-; ci-NEXT: #APP
-; ci-NEXT: .set push
-; ci-NEXT: .set at
-; ci-NEXT: .set macro
-; ci-NEXT: .set reorder
-; ci-EMPTY:
-; ci-NEXT: .space 1000004
-; ci-EMPTY:
-; ci-NEXT: .set pop
-; ci-NEXT: #NO_APP
-; ci-NEXT: jal $BB3_3 # branch
-; ci-NEXT: nop
-; ci-NEXT: .p2align 2
-; ci-NEXT: $BB3_2: # %if.then
-; ci-NEXT: li $2, %hi(j)
-; ci-NEXT: sll $2, $2, 16
-; ci-NEXT: li $3, 10
-; ci-NEXT: sw $3, %lo(j)($2)
-; ci-NEXT: #APP
-; ci-NEXT: .set push
-; ci-NEXT: .set at
-; ci-NEXT: .set macro
-; ci-NEXT: .set reorder
-; ci-EMPTY:
-; ci-NEXT: .space 1000000
-; ci-EMPTY:
-; ci-NEXT: .set pop
-; ci-NEXT: #NO_APP
-; ci-NEXT: .p2align 2
-; ci-NEXT: $BB3_3: # %if.end
-; ci-NEXT: li $2, 0
-; ci-NEXT: jrc $ra
entry:
%0 = load i32, i32* @i, align 4, !tbaa !1
%cmp = icmp eq i32 %0, 0
@@ -250,51 +104,15 @@ if.end: ; preds = %if.else, %if.then
ret i32 0
}
+; ci: .ent y1
+; ci: bnez $2, $BB3_1 # 16 bit inst
+; ci: jal $BB3_2 # branch
+; ci: nop
+; ci: $BB3_1:
+; ci: .end y1
; Function Attrs: nounwind optsize
define void @z0() #0 {
-; ci-LABEL: z0:
-; ci: # %bb.0: # %entry
-; ci-NEXT: li $2, %hi(j)
-; ci-NEXT: sll $2, $2, 16
-; ci-NEXT: lw $2, %lo(j)($2)
-; ci-NEXT: li $3, %hi(i)
-; ci-NEXT: sll $3, $3, 16
-; ci-NEXT: lw $3, %lo(i)($3)
-; ci-NEXT: cmp $3, $2
-; ci-NEXT: bteqz $BB4_2
-; ci-NEXT: # %bb.1: # %if.else
-; ci-NEXT: li $2, %hi(k)
-; ci-NEXT: sll $2, $2, 16
-; ci-NEXT: #APP
-; ci-NEXT: .set push
-; ci-NEXT: .set at
-; ci-NEXT: .set macro
-; ci-NEXT: .set reorder
-; ci-EMPTY:
-; ci-NEXT: .space 10004
-; ci-EMPTY:
-; ci-NEXT: .set pop
-; ci-NEXT: #NO_APP
-; ci-NEXT: li $3, 2
-; ci-NEXT: sw $3, %lo(k)($2)
-; ci-NEXT: jrc $ra
-; ci-NEXT: $BB4_2: # %if.then
-; ci-NEXT: li $2, %hi(k)
-; ci-NEXT: sll $2, $2, 16
-; ci-NEXT: li $3, 1
-; ci-NEXT: sw $3, %lo(k)($2)
-; ci-NEXT: #APP
-; ci-NEXT: .set push
-; ci-NEXT: .set at
-; ci-NEXT: .set macro
-; ci-NEXT: .set reorder
-; ci-EMPTY:
-; ci-NEXT: .space 10000
-; ci-EMPTY:
-; ci-NEXT: .set pop
-; ci-NEXT: #NO_APP
-; ci-NEXT: jrc $ra
entry:
%0 = load i32, i32* @i, align 4, !tbaa !1
%1 = load i32, i32* @j, align 4, !tbaa !1
@@ -315,54 +133,12 @@ if.end: ; preds = %if.else, %if.then
ret void
}
+; ci: .ent z0
+; ci: btnez $BB4_2
+; ci: .end z0
; Function Attrs: nounwind optsize
define void @z1() #0 {
-; ci-LABEL: z1:
-; ci: # %bb.0: # %entry
-; ci-NEXT: li $2, %hi(j)
-; ci-NEXT: sll $2, $2, 16
-; ci-NEXT: lw $2, %lo(j)($2)
-; ci-NEXT: li $3, %hi(i)
-; ci-NEXT: sll $3, $3, 16
-; ci-NEXT: lw $3, %lo(i)($3)
-; ci-NEXT: cmp $3, $2
-; ci-NEXT: btnez $BB5_1 # 16 bit inst
-; ci-NEXT: jal $BB5_2 # branch
-; ci-NEXT: nop
-; ci-NEXT: $BB5_1: # %if.else
-; ci-NEXT: li $2, %hi(k)
-; ci-NEXT: sll $2, $2, 16
-; ci-NEXT: #APP
-; ci-NEXT: .set push
-; ci-NEXT: .set at
-; ci-NEXT: .set macro
-; ci-NEXT: .set reorder
-; ci-EMPTY:
-; ci-NEXT: .space 10000004
-; ci-EMPTY:
-; ci-NEXT: .set pop
-; ci-NEXT: #NO_APP
-; ci-NEXT: li $3, 2
-; ci-NEXT: sw $3, %lo(k)($2)
-; ci-NEXT: jrc $ra
-; ci-NEXT: .p2align 2
-; ci-NEXT: $BB5_2: # %if.then
-; ci-NEXT: li $2, %hi(k)
-; ci-NEXT: sll $2, $2, 16
-; ci-NEXT: li $3, 1
-; ci-NEXT: sw $3, %lo(k)($2)
-; ci-NEXT: #APP
-; ci-NEXT: .set push
-; ci-NEXT: .set at
-; ci-NEXT: .set macro
-; ci-NEXT: .set reorder
-; ci-EMPTY:
-; ci-NEXT: .space 10000000
-; ci-EMPTY:
-; ci-NEXT: .set pop
-; ci-NEXT: #NO_APP
-; ci-NEXT: jrc $ra
entry:
%0 = load i32, i32* @i, align 4, !tbaa !1
%1 = load i32, i32* @j, align 4, !tbaa !1
@@ -383,37 +159,15 @@ if.end: ; preds = %if.else, %if.then
ret void
}
+; ci: .ent z1
+; ci: bteqz $BB5_1 # 16 bit inst
+; ci: jal $BB5_2 # branch
+; ci: nop
+; ci: $BB5_1:
+; ci: .end z1
; Function Attrs: nounwind optsize
define void @z3() #0 {
-; ci-LABEL: z3:
-; ci: # %bb.0: # %entry
-; ci-NEXT: li $2, %hi(i)
-; ci-NEXT: sll $2, $2, 16
-; ci-NEXT: lw $4, %lo(i)($2)
-; ci-NEXT: li $3, %hi(j)
-; ci-NEXT: sll $3, $3, 16
-; ci-NEXT: lw $5, %lo(j)($3)
-; ci-NEXT: slt $5, $4
-; ci-NEXT: bteqz $BB6_2
-; ci-NEXT: $BB6_1: # %if.then
-; ci-NEXT: # =>This Inner Loop Header: Depth=1
-; ci-NEXT: #APP
-; ci-NEXT: .set push
-; ci-NEXT: .set at
-; ci-NEXT: .set macro
-; ci-NEXT: .set reorder
-; ci-EMPTY:
-; ci-NEXT: .space 10000
-; ci-EMPTY:
-; ci-NEXT: .set pop
-; ci-NEXT: #NO_APP
-; ci-NEXT: lw $4, %lo(i)($2)
-; ci-NEXT: lw $5, %lo(j)($3)
-; ci-NEXT: slt $5, $4
-; ci-NEXT: btnez $BB6_1
-; ci-NEXT: $BB6_2: # %if.end
-; ci-NEXT: jrc $ra
entry:
%0 = load i32, i32* @i, align 4, !tbaa !1
%1 = load i32, i32* @j, align 4, !tbaa !1
@@ -431,42 +185,12 @@ if.end: ; preds = %if.then, %entry
ret void
}
+; ci: .ent z3
+; ci: bteqz $BB6_2
+; ci: .end z3
+
; Function Attrs: nounwind optsize
define void @z4() #0 {
-; ci-LABEL: z4:
-; ci: # %bb.0: # %entry
-; ci-NEXT: li $2, %hi(i)
-; ci-NEXT: sll $2, $2, 16
-; ci-NEXT: lw $4, %lo(i)($2)
-; ci-NEXT: li $3, %hi(j)
-; ci-NEXT: sll $3, $3, 16
-; ci-NEXT: lw $5, %lo(j)($3)
-; ci-NEXT: slt $5, $4
-; ci-NEXT: btnez $BB7_1 # 16 bit inst
-; ci-NEXT: jal $BB7_2 # branch
-; ci-NEXT: nop
-; ci-NEXT: .p2align 2
-; ci-NEXT: $BB7_1: # %if.then
-; ci-NEXT: # =>This Inner Loop Header: Depth=1
-; ci-NEXT: #APP
-; ci-NEXT: .set push
-; ci-NEXT: .set at
-; ci-NEXT: .set macro
-; ci-NEXT: .set reorder
-; ci-EMPTY:
-; ci-NEXT: .space 10000000
-; ci-EMPTY:
-; ci-NEXT: .set pop
-; ci-NEXT: #NO_APP
-; ci-NEXT: lw $4, %lo(i)($2)
-; ci-NEXT: lw $5, %lo(j)($3)
-; ci-NEXT: slt $5, $4
-; ci-NEXT: bteqz $BB7_2 # 16 bit inst
-; ci-NEXT: jal $BB7_1 # branch
-; ci-NEXT: nop
-; ci-NEXT: .p2align 2
-; ci-NEXT: $BB7_2: # %if.end
-; ci-NEXT: jrc $ra
entry:
%0 = load i32, i32* @i, align 4, !tbaa !1
%1 = load i32, i32* @j, align 4, !tbaa !1
@@ -484,6 +208,14 @@ if.end: ; preds = %if.then, %entry
ret void
}
+; ci: .ent z4
+; ci: btnez $BB7_1 # 16 bit inst
+; ci: jal $BB7_2 # branch
+; ci: nop
+; ci: .p2align 2
+; ci: $BB7_1:
+; ci: .end z4
+
attributes #0 = { nounwind optsize "less-precise-fpmad"="false" "frame-pointer"="none" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" }
attributes #1 = { nounwind }
diff --git a/llvm/test/CodeGen/Mips/longbranch/compact-branches-long-branch.ll b/llvm/test/CodeGen/Mips/longbranch/compact-branches-long-branch.ll
index dbd071f81cbf..709cd477a778 100644
--- a/llvm/test/CodeGen/Mips/longbranch/compact-branches-long-branch.ll
+++ b/llvm/test/CodeGen/Mips/longbranch/compact-branches-long-branch.ll
@@ -1,4 +1,3 @@
-; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc < %s -march=mips -mcpu=mips32r6 -force-mips-long-branch | FileCheck %s
; Check that when MIPS32R6 with the static relocation model with the usage of
@@ -10,29 +9,11 @@ declare i32 @f(i32)
declare i32 @g()
-define i32 @test1(i32 %a) {
; CHECK-LABEL: test1:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: addiu $sp, $sp, -24
-; CHECK-NEXT: .cfi_def_cfa_offset 24
-; CHECK-NEXT: sw $ra, 20($sp) # 4-byte Folded Spill
-; CHECK-NEXT: .cfi_offset 31, -4
-; CHECK-NEXT: bnezc $4, $BB0_2
-; CHECK-NEXT: nop
-; CHECK-NEXT: # %bb.1: # %entry
-; CHECK-NEXT: bc $BB0_3
-; CHECK-NEXT: $BB0_2: # %cond.false
-; CHECK-NEXT: jal g
-; CHECK-NEXT: nop
-; CHECK-NEXT: lw $ra, 20($sp) # 4-byte Folded Reload
-; CHECK-NEXT: jr $ra
-; CHECK-NEXT: addiu $sp, $sp, 24
-; CHECK-NEXT: $BB0_3: # %cond.true
-; CHECK-NEXT: jal f
-; CHECK-NEXT: nop
-; CHECK-NEXT: lw $ra, 20($sp) # 4-byte Folded Reload
-; CHECK-NEXT: jr $ra
-; CHECK-NEXT: addiu $sp, $sp, 24
+; CHECK: bnezc
+; CHECK-NEXT: nop
+
+define i32 @test1(i32 %a) {
entry:
%0 = icmp eq i32 %a, 0
br i1 %0, label %cond.true, label %cond.false
@@ -44,30 +25,11 @@ cond.false:
ret i32 %2
}
+; CHECK-LABEL: test2:
+; CHECK: bgezc
+; CHECK-NEXT: nop
define i32 @test2(i32 %a) {
-; CHECK-LABEL: test2:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: addiu $sp, $sp, -24
-; CHECK-NEXT: .cfi_def_cfa_offset 24
-; CHECK-NEXT: sw $ra, 20($sp) # 4-byte Folded Spill
-; CHECK-NEXT: .cfi_offset 31, -4
-; CHECK-NEXT: bgezc $4, $BB1_2
-; CHECK-NEXT: nop
-; CHECK-NEXT: # %bb.1: # %entry
-; CHECK-NEXT: bc $BB1_3
-; CHECK-NEXT: $BB1_2: # %cond.true
-; CHECK-NEXT: jal f
-; CHECK-NEXT: nop
-; CHECK-NEXT: lw $ra, 20($sp) # 4-byte Folded Reload
-; CHECK-NEXT: jr $ra
-; CHECK-NEXT: addiu $sp, $sp, 24
-; CHECK-NEXT: $BB1_3: # %cond.false
-; CHECK-NEXT: jal g
-; CHECK-NEXT: nop
-; CHECK-NEXT: lw $ra, 20($sp) # 4-byte Folded Reload
-; CHECK-NEXT: jr $ra
-; CHECK-NEXT: addiu $sp, $sp, 24
entry:
%0 = icmp sge i32 %a, 0
br i1 %0, label %cond.true, label %cond.false
@@ -79,30 +41,11 @@ cond.false:
ret i32 %2
}
+; CHECK-LABEL: test3:
+; CHECK: blezc
+; CHECK-NEXT: nop
define i32 @test3(i32 %a) {
-; CHECK-LABEL: test3:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: addiu $sp, $sp, -24
-; CHECK-NEXT: .cfi_def_cfa_offset 24
-; CHECK-NEXT: sw $ra, 20($sp) # 4-byte Folded Spill
-; CHECK-NEXT: .cfi_offset 31, -4
-; CHECK-NEXT: blezc $4, $BB2_2
-; CHECK-NEXT: nop
-; CHECK-NEXT: # %bb.1: # %entry
-; CHECK-NEXT: bc $BB2_3
-; CHECK-NEXT: $BB2_2: # %cond.true
-; CHECK-NEXT: jal f
-; CHECK-NEXT: nop
-; CHECK-NEXT: lw $ra, 20($sp) # 4-byte Folded Reload
-; CHECK-NEXT: jr $ra
-; CHECK-NEXT: addiu $sp, $sp, 24
-; CHECK-NEXT: $BB2_3: # %cond.false
-; CHECK-NEXT: jal g
-; CHECK-NEXT: nop
-; CHECK-NEXT: lw $ra, 20($sp) # 4-byte Folded Reload
-; CHECK-NEXT: jr $ra
-; CHECK-NEXT: addiu $sp, $sp, 24
entry:
%0 = icmp sle i32 %a, 0
br i1 %0, label %cond.true, label %cond.false
@@ -114,30 +57,11 @@ cond.false:
ret i32 %2
}
+; CHECK-LABEL: test4:
+; CHECK: bgtzc
+; CHECK-NEXT: nop
define i32 @test4(i32 %a) {
-; CHECK-LABEL: test4:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: addiu $sp, $sp, -24
-; CHECK-NEXT: .cfi_def_cfa_offset 24
-; CHECK-NEXT: sw $ra, 20($sp) # 4-byte Folded Spill
-; CHECK-NEXT: .cfi_offset 31, -4
-; CHECK-NEXT: bgtzc $4, $BB3_2
-; CHECK-NEXT: nop
-; CHECK-NEXT: # %bb.1: # %entry
-; CHECK-NEXT: bc $BB3_3
-; CHECK-NEXT: $BB3_2: # %cond.true
-; CHECK-NEXT: jal f
-; CHECK-NEXT: nop
-; CHECK-NEXT: lw $ra, 20($sp) # 4-byte Folded Reload
-; CHECK-NEXT: jr $ra
-; CHECK-NEXT: addiu $sp, $sp, 24
-; CHECK-NEXT: $BB3_3: # %cond.false
-; CHECK-NEXT: jal g
-; CHECK-NEXT: nop
-; CHECK-NEXT: lw $ra, 20($sp) # 4-byte Folded Reload
-; CHECK-NEXT: jr $ra
-; CHECK-NEXT: addiu $sp, $sp, 24
entry:
%0 = icmp sgt i32 %a, 0
br i1 %0, label %cond.true, label %cond.false
@@ -149,29 +73,11 @@ cond.false:
ret i32 %2
}
-define i32 @test5(i32 %a) {
; CHECK-LABEL: test5:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: addiu $sp, $sp, -24
-; CHECK-NEXT: .cfi_def_cfa_offset 24
-; CHECK-NEXT: sw $ra, 20($sp) # 4-byte Folded Spill
-; CHECK-NEXT: .cfi_offset 31, -4
-; CHECK-NEXT: bgezc $4, $BB4_2
-; CHECK-NEXT: nop
-; CHECK-NEXT: # %bb.1: # %entry
-; CHECK-NEXT: bc $BB4_3
-; CHECK-NEXT: $BB4_2: # %cond.false
-; CHECK-NEXT: jal g
-; CHECK-NEXT: nop
-; CHECK-NEXT: lw $ra, 20($sp) # 4-byte Folded Reload
-; CHECK-NEXT: jr $ra
-; CHECK-NEXT: addiu $sp, $sp, 24
-; CHECK-NEXT: $BB4_3: # %cond.true
-; CHECK-NEXT: jal f
-; CHECK-NEXT: nop
-; CHECK-NEXT: lw $ra, 20($sp) # 4-byte Folded Reload
-; CHECK-NEXT: jr $ra
-; CHECK-NEXT: addiu $sp, $sp, 24
+; CHECK: bgezc
+; CHECK-NEXT: nop
+
+define i32 @test5(i32 %a) {
entry:
%0 = icmp slt i32 %a, 0
br i1 %0, label %cond.true, label %cond.false
@@ -183,30 +89,11 @@ cond.false:
ret i32 %2
}
-define i32 @test6(i32 %a, i32 %b) {
; CHECK-LABEL: test6:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: addiu $sp, $sp, -24
-; CHECK-NEXT: .cfi_def_cfa_offset 24
-; CHECK-NEXT: sw $ra, 20($sp) # 4-byte Folded Spill
-; CHECK-NEXT: .cfi_offset 31, -4
-; CHECK-NEXT: sltu $1, $5, $4
-; CHECK-NEXT: bnezc $1, $BB5_2
-; CHECK-NEXT: nop
-; CHECK-NEXT: # %bb.1: # %entry
-; CHECK-NEXT: bc $BB5_3
-; CHECK-NEXT: $BB5_2: # %cond.true
-; CHECK-NEXT: jal f
-; CHECK-NEXT: nop
-; CHECK-NEXT: lw $ra, 20($sp) # 4-byte Folded Reload
-; CHECK-NEXT: jr $ra
-; CHECK-NEXT: addiu $sp, $sp, 24
-; CHECK-NEXT: $BB5_3: # %cond.false
-; CHECK-NEXT: jal g
-; CHECK-NEXT: nop
-; CHECK-NEXT: lw $ra, 20($sp) # 4-byte Folded Reload
-; CHECK-NEXT: jr $ra
-; CHECK-NEXT: addiu $sp, $sp, 24
+; CHECK: bnezc
+; CHECK-NEXT: nop
+
+define i32 @test6(i32 %a, i32 %b) {
entry:
%0 = icmp ugt i32 %a, %b
br i1 %0, label %cond.true, label %cond.false
@@ -218,31 +105,11 @@ cond.false:
ret i32 %2
}
+; CHECK-LABEL: test7:
+; CHECK: beqzc
+; CHECK-NEXT: nop
define i32 @test7(i32 %a, i32 %b) {
-; CHECK-LABEL: test7:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: addiu $sp, $sp, -24
-; CHECK-NEXT: .cfi_def_cfa_offset 24
-; CHECK-NEXT: sw $ra, 20($sp) # 4-byte Folded Spill
-; CHECK-NEXT: .cfi_offset 31, -4
-; CHECK-NEXT: sltu $1, $4, $5
-; CHECK-NEXT: bnezc $1, $BB6_2
-; CHECK-NEXT: nop
-; CHECK-NEXT: # %bb.1: # %entry
-; CHECK-NEXT: bc $BB6_3
-; CHECK-NEXT: $BB6_2: # %cond.false
-; CHECK-NEXT: jal g
-; CHECK-NEXT: nop
-; CHECK-NEXT: lw $ra, 20($sp) # 4-byte Folded Reload
-; CHECK-NEXT: jr $ra
-; CHECK-NEXT: addiu $sp, $sp, 24
-; CHECK-NEXT: $BB6_3: # %cond.true
-; CHECK-NEXT: jal f
-; CHECK-NEXT: nop
-; CHECK-NEXT: lw $ra, 20($sp) # 4-byte Folded Reload
-; CHECK-NEXT: jr $ra
-; CHECK-NEXT: addiu $sp, $sp, 24
entry:
%0 = icmp uge i32 %a, %b
br i1 %0, label %cond.true, label %cond.false
@@ -254,31 +121,11 @@ cond.false:
ret i32 %2
}
+; CHECK-LABEL: test8:
+; CHECK: bnezc
+; CHECK-NEXT: nop
define i32 @test8(i32 %a, i32 %b) {
-; CHECK-LABEL: test8:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: addiu $sp, $sp, -24
-; CHECK-NEXT: .cfi_def_cfa_offset 24
-; CHECK-NEXT: sw $ra, 20($sp) # 4-byte Folded Spill
-; CHECK-NEXT: .cfi_offset 31, -4
-; CHECK-NEXT: sltu $1, $4, $5
-; CHECK-NEXT: bnezc $1, $BB7_2
-; CHECK-NEXT: nop
-; CHECK-NEXT: # %bb.1: # %entry
-; CHECK-NEXT: bc $BB7_3
-; CHECK-NEXT: $BB7_2: # %cond.true
-; CHECK-NEXT: jal f
-; CHECK-NEXT: nop
-; CHECK-NEXT: lw $ra, 20($sp) # 4-byte Folded Reload
-; CHECK-NEXT: jr $ra
-; CHECK-NEXT: addiu $sp, $sp, 24
-; CHECK-NEXT: $BB7_3: # %cond.false
-; CHECK-NEXT: jal g
-; CHECK-NEXT: nop
-; CHECK-NEXT: lw $ra, 20($sp) # 4-byte Folded Reload
-; CHECK-NEXT: jr $ra
-; CHECK-NEXT: addiu $sp, $sp, 24
entry:
%0 = icmp ult i32 %a, %b
br i1 %0, label %cond.true, label %cond.false
@@ -290,31 +137,11 @@ cond.false:
ret i32 %2
}
+; CHECK-LABEL: test9:
+; CHECK: beqzc
+; CHECK-NEXT: nop
define i32 @test9(i32 %a, i32 %b) {
-; CHECK-LABEL: test9:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: addiu $sp, $sp, -24
-; CHECK-NEXT: .cfi_def_cfa_offset 24
-; CHECK-NEXT: sw $ra, 20($sp) # 4-byte Folded Spill
-; CHECK-NEXT: .cfi_offset 31, -4
-; CHECK-NEXT: sltu $1, $5, $4
-; CHECK-NEXT: bnezc $1, $BB8_2
-; CHECK-NEXT: nop
-; CHECK-NEXT: # %bb.1: # %entry
-; CHECK-NEXT: bc $BB8_3
-; CHECK-NEXT: $BB8_2: # %cond.false
-; CHECK-NEXT: jal g
-; CHECK-NEXT: nop
-; CHECK-NEXT: lw $ra, 20($sp) # 4-byte Folded Reload
-; CHECK-NEXT: jr $ra
-; CHECK-NEXT: addiu $sp, $sp, 24
-; CHECK-NEXT: $BB8_3: # %cond.true
-; CHECK-NEXT: jal f
-; CHECK-NEXT: nop
-; CHECK-NEXT: lw $ra, 20($sp) # 4-byte Folded Reload
-; CHECK-NEXT: jr $ra
-; CHECK-NEXT: addiu $sp, $sp, 24
entry:
%0 = icmp ule i32 %a, %b
br i1 %0, label %cond.true, label %cond.false
diff --git a/llvm/test/CodeGen/Mips/seleq.ll b/llvm/test/CodeGen/Mips/seleq.ll
index 34565ea27275..7d1e034d68c7 100644
--- a/llvm/test/CodeGen/Mips/seleq.ll
+++ b/llvm/test/CodeGen/Mips/seleq.ll
@@ -1,5 +1,4 @@
-; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
-; RUN: llc -march=mipsel -mattr=mips16 -relocation-model=pic < %s | FileCheck %s -check-prefix=MIPS16
+; RUN: llc -march=mipsel -mattr=mips16 -relocation-model=pic < %s | FileCheck %s -check-prefix=16
@t = global i32 10, align 4
@f = global i32 199, align 4
@@ -12,74 +11,6 @@
@z4 = common global i32 0, align 4
define void @calc_seleq() nounwind {
-; MIPS16-LABEL: calc_seleq:
-; MIPS16: # %bb.0: # %entry
-; MIPS16-NEXT: lui $2, %hi(_gp_disp)
-; MIPS16-NEXT: addiu $2, $2, %lo(_gp_disp)
-; MIPS16-NEXT: li $2, %hi(_gp_disp)
-; MIPS16-NEXT: addiu $3, $pc, %lo(_gp_disp)
-; MIPS16-NEXT: sll $2, $2, 16
-; MIPS16-NEXT: addu $2, $3, $2
-; MIPS16-NEXT: lw $4, %got(b)($2)
-; MIPS16-NEXT: lw $5, 0($4)
-; MIPS16-NEXT: lw $3, %got(a)($2)
-; MIPS16-NEXT: lw $6, 0($3)
-; MIPS16-NEXT: cmp $6, $5
-; MIPS16-NEXT: bteqz $BB0_2 # 16 bit inst
-; MIPS16-NEXT: # %bb.1: # %cond.false
-; MIPS16-NEXT: lw $5, %got(t)($2)
-; MIPS16-NEXT: lw $5, 0($5)
-; MIPS16-NEXT: b $BB0_3 # 16 bit inst
-; MIPS16-NEXT: $BB0_2: # %cond.true
-; MIPS16-NEXT: lw $5, %got(f)($2)
-; MIPS16-NEXT: lw $5, 0($5)
-; MIPS16-NEXT: $BB0_3: # %cond.end
-; MIPS16-NEXT: lw $6, %got(z1)($2)
-; MIPS16-NEXT: sw $5, 0($6)
-; MIPS16-NEXT: lw $5, 0($3)
-; MIPS16-NEXT: lw $4, 0($4)
-; MIPS16-NEXT: cmp $4, $5
-; MIPS16-NEXT: bteqz $BB0_5 # 16 bit inst
-; MIPS16-NEXT: # %bb.4: # %cond.false3
-; MIPS16-NEXT: lw $4, %got(t)($2)
-; MIPS16-NEXT: lw $4, 0($4)
-; MIPS16-NEXT: b $BB0_6 # 16 bit inst
-; MIPS16-NEXT: $BB0_5: # %cond.true2
-; MIPS16-NEXT: lw $4, %got(f)($2)
-; MIPS16-NEXT: lw $4, 0($4)
-; MIPS16-NEXT: $BB0_6: # %cond.end4
-; MIPS16-NEXT: lw $5, %got(z2)($2)
-; MIPS16-NEXT: sw $4, 0($5)
-; MIPS16-NEXT: lw $5, 0($3)
-; MIPS16-NEXT: lw $4, %got(c)($2)
-; MIPS16-NEXT: lw $6, 0($4)
-; MIPS16-NEXT: cmp $6, $5
-; MIPS16-NEXT: bteqz $BB0_8 # 16 bit inst
-; MIPS16-NEXT: # %bb.7: # %cond.false8
-; MIPS16-NEXT: lw $5, %got(f)($2)
-; MIPS16-NEXT: lw $5, 0($5)
-; MIPS16-NEXT: b $BB0_9 # 16 bit inst
-; MIPS16-NEXT: $BB0_8: # %cond.true7
-; MIPS16-NEXT: lw $5, %got(t)($2)
-; MIPS16-NEXT: lw $5, 0($5)
-; MIPS16-NEXT: $BB0_9: # %cond.end9
-; MIPS16-NEXT: lw $6, %got(z3)($2)
-; MIPS16-NEXT: sw $5, 0($6)
-; MIPS16-NEXT: lw $4, 0($4)
-; MIPS16-NEXT: lw $3, 0($3)
-; MIPS16-NEXT: cmp $3, $4
-; MIPS16-NEXT: bteqz $BB0_11 # 16 bit inst
-; MIPS16-NEXT: # %bb.10: # %cond.false13
-; MIPS16-NEXT: lw $3, %got(f)($2)
-; MIPS16-NEXT: lw $3, 0($3)
-; MIPS16-NEXT: b $BB0_12 # 16 bit inst
-; MIPS16-NEXT: $BB0_11: # %cond.true12
-; MIPS16-NEXT: lw $3, %got(t)($2)
-; MIPS16-NEXT: lw $3, 0($3)
-; MIPS16-NEXT: $BB0_12: # %cond.end14
-; MIPS16-NEXT: lw $2, %got(z4)($2)
-; MIPS16-NEXT: sw $3, 0($2)
-; MIPS16-NEXT: jrc $ra
entry:
%0 = load i32, i32* @a, align 4
%1 = load i32, i32* @b, align 4
@@ -149,3 +80,16 @@ cond.end14: ; preds = %cond.false13, %cond
}
attributes #0 = { nounwind "target-cpu"="mips32" "target-features"="+o32,+mips32" }
+
+; 16: cmp ${{[0-9]+}}, ${{[0-9]+}}
+; 16: btnez $BB{{[0-9]+}}_{{[0-9]}}
+
+; 16: cmp ${{[0-9]+}}, ${{[0-9]+}}
+; 16: btnez $BB{{[0-9]+}}_{{[0-9]}}
+
+; 16: cmp ${{[0-9]+}}, ${{[0-9]+}}
+; 16: btnez $BB{{[0-9]+}}_{{[0-9]}}
+
+; 16: cmp ${{[0-9]+}}, ${{[0-9]+}}
+; 16: btnez $BB{{[0-9]+}}_{{[0-9]}}
+
diff --git a/llvm/test/CodeGen/Mips/selle.ll b/llvm/test/CodeGen/Mips/selle.ll
index ffad4ba1c349..8925aac10c4d 100644
--- a/llvm/test/CodeGen/Mips/selle.ll
+++ b/llvm/test/CodeGen/Mips/selle.ll
@@ -1,5 +1,4 @@
-; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
-; RUN: llc -march=mipsel -mattr=mips16 -relocation-model=pic < %s | FileCheck %s -check-prefix=MIPS16
+; RUN: llc -march=mipsel -mattr=mips16 -relocation-model=pic < %s | FileCheck %s -check-prefix=16
@t = global i32 10, align 4
@f = global i32 199, align 4
@@ -13,74 +12,6 @@
@.str = private unnamed_addr constant [5 x i8] c"%i \0A\00", align 1
define void @calc_z() nounwind "target-cpu"="mips16" "target-features"="+mips16,+o32" {
-; MIPS16-LABEL: calc_z:
-; MIPS16: # %bb.0: # %entry
-; MIPS16-NEXT: lui $2, %hi(_gp_disp)
-; MIPS16-NEXT: addiu $2, $2, %lo(_gp_disp)
-; MIPS16-NEXT: li $2, %hi(_gp_disp)
-; MIPS16-NEXT: addiu $3, $pc, %lo(_gp_disp)
-; MIPS16-NEXT: sll $2, $2, 16
-; MIPS16-NEXT: addu $2, $3, $2
-; MIPS16-NEXT: lw $3, %got(a)($2)
-; MIPS16-NEXT: lw $5, 0($3)
-; MIPS16-NEXT: lw $4, %got(b)($2)
-; MIPS16-NEXT: lw $6, 0($4)
-; MIPS16-NEXT: slt $6, $5
-; MIPS16-NEXT: bteqz $BB0_2 # 16 bit inst
-; MIPS16-NEXT: # %bb.1: # %cond.false
-; MIPS16-NEXT: lw $5, %got(f)($2)
-; MIPS16-NEXT: lw $5, 0($5)
-; MIPS16-NEXT: b $BB0_3 # 16 bit inst
-; MIPS16-NEXT: $BB0_2: # %cond.true
-; MIPS16-NEXT: lw $5, %got(t)($2)
-; MIPS16-NEXT: lw $5, 0($5)
-; MIPS16-NEXT: $BB0_3: # %cond.end
-; MIPS16-NEXT: lw $6, %got(z1)($2)
-; MIPS16-NEXT: sw $5, 0($6)
-; MIPS16-NEXT: lw $4, 0($4)
-; MIPS16-NEXT: lw $5, 0($3)
-; MIPS16-NEXT: slt $5, $4
-; MIPS16-NEXT: bteqz $BB0_5 # 16 bit inst
-; MIPS16-NEXT: # %bb.4: # %cond.false3
-; MIPS16-NEXT: lw $4, %got(t)($2)
-; MIPS16-NEXT: lw $4, 0($4)
-; MIPS16-NEXT: b $BB0_6 # 16 bit inst
-; MIPS16-NEXT: $BB0_5: # %cond.true2
-; MIPS16-NEXT: lw $4, %got(f)($2)
-; MIPS16-NEXT: lw $4, 0($4)
-; MIPS16-NEXT: $BB0_6: # %cond.end4
-; MIPS16-NEXT: lw $5, %got(z2)($2)
-; MIPS16-NEXT: sw $4, 0($5)
-; MIPS16-NEXT: lw $4, %got(c)($2)
-; MIPS16-NEXT: lw $5, 0($4)
-; MIPS16-NEXT: lw $6, 0($3)
-; MIPS16-NEXT: slt $6, $5
-; MIPS16-NEXT: bteqz $BB0_8 # 16 bit inst
-; MIPS16-NEXT: # %bb.7: # %cond.false8
-; MIPS16-NEXT: lw $5, %got(f)($2)
-; MIPS16-NEXT: lw $5, 0($5)
-; MIPS16-NEXT: b $BB0_9 # 16 bit inst
-; MIPS16-NEXT: $BB0_8: # %cond.true7
-; MIPS16-NEXT: lw $5, %got(t)($2)
-; MIPS16-NEXT: lw $5, 0($5)
-; MIPS16-NEXT: $BB0_9: # %cond.end9
-; MIPS16-NEXT: lw $6, %got(z3)($2)
-; MIPS16-NEXT: sw $5, 0($6)
-; MIPS16-NEXT: lw $3, 0($3)
-; MIPS16-NEXT: lw $4, 0($4)
-; MIPS16-NEXT: slt $4, $3
-; MIPS16-NEXT: bteqz $BB0_11 # 16 bit inst
-; MIPS16-NEXT: # %bb.10: # %cond.false13
-; MIPS16-NEXT: lw $3, %got(f)($2)
-; MIPS16-NEXT: lw $3, 0($3)
-; MIPS16-NEXT: b $BB0_12 # 16 bit inst
-; MIPS16-NEXT: $BB0_11: # %cond.true12
-; MIPS16-NEXT: lw $3, %got(t)($2)
-; MIPS16-NEXT: lw $3, 0($3)
-; MIPS16-NEXT: $BB0_12: # %cond.end14
-; MIPS16-NEXT: lw $2, %got(z4)($2)
-; MIPS16-NEXT: sw $3, 0($2)
-; MIPS16-NEXT: jrc $ra
entry:
%0 = load i32, i32* @a, align 4
%1 = load i32, i32* @b, align 4
@@ -149,6 +80,17 @@ cond.end14: ; preds = %cond.false13, %cond
ret void
}
+; 16: slt ${{[0-9]+}}, ${{[0-9]+}}
+; 16: btnez $BB{{[0-9]+}}_{{[0-9]}}
+
+; 16: slt ${{[0-9]+}}, ${{[0-9]+}}
+; 16: btnez $BB{{[0-9]+}}_{{[0-9]}}
+
+; 16: slt ${{[0-9]+}}, ${{[0-9]+}}
+; 16: btnez $BB{{[0-9]+}}_{{[0-9]}}
+
+; 16: slt ${{[0-9]+}}, ${{[0-9]+}}
+; 16: btnez $BB{{[0-9]+}}_{{[0-9]}}
attributes #0 = { nounwind "target-cpu"="mips16" "target-features"="+mips16,+o32" }
attributes #1 = { "target-cpu"="mips16" "target-features"="+mips16,+o32" }
diff --git a/llvm/test/CodeGen/PowerPC/brcond.ll b/llvm/test/CodeGen/PowerPC/brcond.ll
index 3df169dd64da..b8c98427f107 100644
--- a/llvm/test/CodeGen/PowerPC/brcond.ll
+++ b/llvm/test/CodeGen/PowerPC/brcond.ll
@@ -1,22 +1,12 @@
-; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc -verify-machineinstrs -mtriple=powerpc64-unknown-unknown \
; RUN: -ppc-reduce-cr-logicals=false < %s | FileCheck %s
; RUN: llc -verify-machineinstrs -mtriple=powerpc64le-unknown-unknown \
; RUN: -ppc-reduce-cr-logicals=false < %s | FileCheck %s
define signext i32 @testi32slt(i32 signext %c1, i32 signext %c2, i32 signext %c3, i32 signext %c4, i32 signext %a1, i32 signext %a2) #0 {
-; CHECK-LABEL: testi32slt:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: cmpw 5, 6
-; CHECK-NEXT: cmpw 1, 3, 4
-; CHECK-NEXT: crorc 20, 2, 6
-; CHECK-NEXT: bc 12, 20, .LBB0_2
-; CHECK-NEXT: # %bb.1: # %iftrue
-; CHECK-NEXT: extsw 3, 7
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB0_2: # %iffalse
-; CHECK-NEXT: extsw 3, 8
-; CHECK-NEXT: blr
+; CHECK-LABEL: testi32slt
+; CHECK: crorc [[REG:[0-9]+]], 2, 6
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = icmp eq i32 %c3, %c4
%cmp3tmp = icmp eq i32 %c1, %c2
@@ -29,18 +19,9 @@ iffalse:
}
define signext i32 @testi32ult(i32 signext %c1, i32 signext %c2, i32 signext %c3, i32 signext %c4, i32 signext %a1, i32 signext %a2) #0 {
-; CHECK-LABEL: testi32ult:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: cmpw 5, 6
-; CHECK-NEXT: cmpw 1, 3, 4
-; CHECK-NEXT: crorc 20, 6, 2
-; CHECK-NEXT: bc 12, 20, .LBB1_2
-; CHECK-NEXT: # %bb.1: # %iftrue
-; CHECK-NEXT: extsw 3, 7
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB1_2: # %iffalse
-; CHECK-NEXT: extsw 3, 8
-; CHECK-NEXT: blr
+; CHECK-LABEL: testi32ult
+; CHECK: crorc [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = icmp eq i32 %c3, %c4
%cmp3tmp = icmp eq i32 %c1, %c2
@@ -53,18 +34,9 @@ iffalse:
}
define signext i32 @testi32sle(i32 signext %c1, i32 signext %c2, i32 signext %c3, i32 signext %c4, i32 signext %a1, i32 signext %a2) #0 {
-; CHECK-LABEL: testi32sle:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: cmpw 5, 6
-; CHECK-NEXT: cmpw 1, 3, 4
-; CHECK-NEXT: crandc 20, 2, 6
-; CHECK-NEXT: bc 4, 20, .LBB2_2
-; CHECK-NEXT: # %bb.1: # %iffalse
-; CHECK-NEXT: extsw 3, 8
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB2_2: # %iftrue
-; CHECK-NEXT: extsw 3, 7
-; CHECK-NEXT: blr
+; CHECK-LABEL: testi32sle
+; CHECK: crandc [[REG:[0-9]+]], 2, 6
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = icmp eq i32 %c3, %c4
%cmp3tmp = icmp eq i32 %c1, %c2
@@ -77,18 +49,9 @@ iffalse:
}
define signext i32 @testi32ule(i32 signext %c1, i32 signext %c2, i32 signext %c3, i32 signext %c4, i32 signext %a1, i32 signext %a2) #0 {
-; CHECK-LABEL: testi32ule:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: cmpw 5, 6
-; CHECK-NEXT: cmpw 1, 3, 4
-; CHECK-NEXT: crandc 20, 6, 2
-; CHECK-NEXT: bc 4, 20, .LBB3_2
-; CHECK-NEXT: # %bb.1: # %iffalse
-; CHECK-NEXT: extsw 3, 8
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB3_2: # %iftrue
-; CHECK-NEXT: extsw 3, 7
-; CHECK-NEXT: blr
+; CHECK-LABEL: testi32ule
+; CHECK: crandc [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = icmp eq i32 %c3, %c4
%cmp3tmp = icmp eq i32 %c1, %c2
@@ -102,17 +65,8 @@ iffalse:
define signext i32 @testi32eq(i32 signext %c1, i32 signext %c2, i32 signext %c3, i32 signext %c4, i32 signext %a1, i32 signext %a2) #0 {
; CHECK-LABEL: testi32eq:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: cmpw 5, 6
-; CHECK-NEXT: cmpw 1, 3, 4
-; CHECK-NEXT: crxor 20, 6, 2
-; CHECK-NEXT: bc 4, 20, .LBB4_2
-; CHECK-NEXT: # %bb.1: # %iffalse
-; CHECK-NEXT: extsw 3, 8
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB4_2: # %iftrue
-; CHECK-NEXT: extsw 3, 7
-; CHECK-NEXT: blr
+; CHECK: crxor [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = icmp eq i32 %c3, %c4
%cmp3tmp = icmp eq i32 %c1, %c2
@@ -126,17 +80,8 @@ iffalse:
define signext i32 @testi32sge(i32 signext %c1, i32 signext %c2, i32 signext %c3, i32 signext %c4, i32 signext %a1, i32 signext %a2) #0 {
; CHECK-LABEL: testi32sge:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: cmpw 5, 6
-; CHECK-NEXT: cmpw 1, 3, 4
-; CHECK-NEXT: crandc 20, 6, 2
-; CHECK-NEXT: bc 4, 20, .LBB5_2
-; CHECK-NEXT: # %bb.1: # %iffalse
-; CHECK-NEXT: extsw 3, 8
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB5_2: # %iftrue
-; CHECK-NEXT: extsw 3, 7
-; CHECK-NEXT: blr
+; CHECK: crandc [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = icmp eq i32 %c3, %c4
%cmp3tmp = icmp eq i32 %c1, %c2
@@ -150,17 +95,8 @@ iffalse:
define signext i32 @testi32uge(i32 signext %c1, i32 signext %c2, i32 signext %c3, i32 signext %c4, i32 signext %a1, i32 signext %a2) #0 {
; CHECK-LABEL: testi32uge:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: cmpw 5, 6
-; CHECK-NEXT: cmpw 1, 3, 4
-; CHECK-NEXT: crandc 20, 2, 6
-; CHECK-NEXT: bc 4, 20, .LBB6_2
-; CHECK-NEXT: # %bb.1: # %iffalse
-; CHECK-NEXT: extsw 3, 8
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB6_2: # %iftrue
-; CHECK-NEXT: extsw 3, 7
-; CHECK-NEXT: blr
+; CHECK: crandc [[REG:[0-9]+]], 2, 6
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = icmp eq i32 %c3, %c4
%cmp3tmp = icmp eq i32 %c1, %c2
@@ -174,17 +110,8 @@ iffalse:
define signext i32 @testi32sgt(i32 signext %c1, i32 signext %c2, i32 signext %c3, i32 signext %c4, i32 signext %a1, i32 signext %a2) #0 {
; CHECK-LABEL: testi32sgt:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: cmpw 5, 6
-; CHECK-NEXT: cmpw 1, 3, 4
-; CHECK-NEXT: crorc 20, 6, 2
-; CHECK-NEXT: bc 12, 20, .LBB7_2
-; CHECK-NEXT: # %bb.1: # %iftrue
-; CHECK-NEXT: extsw 3, 7
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB7_2: # %iffalse
-; CHECK-NEXT: extsw 3, 8
-; CHECK-NEXT: blr
+; CHECK: crorc [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = icmp eq i32 %c3, %c4
%cmp3tmp = icmp eq i32 %c1, %c2
@@ -198,17 +125,8 @@ iffalse:
define signext i32 @testi32ugt(i32 signext %c1, i32 signext %c2, i32 signext %c3, i32 signext %c4, i32 signext %a1, i32 signext %a2) #0 {
; CHECK-LABEL: testi32ugt:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: cmpw 5, 6
-; CHECK-NEXT: cmpw 1, 3, 4
-; CHECK-NEXT: crorc 20, 2, 6
-; CHECK-NEXT: bc 12, 20, .LBB8_2
-; CHECK-NEXT: # %bb.1: # %iftrue
-; CHECK-NEXT: extsw 3, 7
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB8_2: # %iffalse
-; CHECK-NEXT: extsw 3, 8
-; CHECK-NEXT: blr
+; CHECK: crorc [[REG:[0-9]+]], 2, 6
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = icmp eq i32 %c3, %c4
%cmp3tmp = icmp eq i32 %c1, %c2
@@ -222,17 +140,8 @@ iffalse:
define signext i32 @testi32ne(i32 signext %c1, i32 signext %c2, i32 signext %c3, i32 signext %c4, i32 signext %a1, i32 signext %a2) #0 {
; CHECK-LABEL: testi32ne:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: cmpw 5, 6
-; CHECK-NEXT: cmpw 1, 3, 4
-; CHECK-NEXT: creqv 20, 6, 2
-; CHECK-NEXT: bc 12, 20, .LBB9_2
-; CHECK-NEXT: # %bb.1: # %iftrue
-; CHECK-NEXT: extsw 3, 7
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB9_2: # %iffalse
-; CHECK-NEXT: extsw 3, 8
-; CHECK-NEXT: blr
+; CHECK: creqv [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = icmp eq i32 %c3, %c4
%cmp3tmp = icmp eq i32 %c1, %c2
@@ -245,18 +154,9 @@ iffalse:
}
define i64 @testi64slt(i64 %c1, i64 %c2, i64 %c3, i64 %c4, i64 %a1, i64 %a2) #0 {
-; CHECK-LABEL: testi64slt:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: cmpd 5, 6
-; CHECK-NEXT: cmpd 1, 3, 4
-; CHECK-NEXT: crorc 20, 2, 6
-; CHECK-NEXT: bc 12, 20, .LBB10_2
-; CHECK-NEXT: # %bb.1: # %iftrue
-; CHECK-NEXT: mr 3, 7
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB10_2: # %iffalse
-; CHECK-NEXT: mr 3, 8
-; CHECK-NEXT: blr
+; CHECK-LABEL: testi64slt
+; CHECK: crorc [[REG:[0-9]+]], 2, 6
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = icmp eq i64 %c3, %c4
%cmp3tmp = icmp eq i64 %c1, %c2
@@ -269,18 +169,9 @@ iffalse:
}
define i64 @testi64ult(i64 %c1, i64 %c2, i64 %c3, i64 %c4, i64 %a1, i64 %a2) #0 {
-; CHECK-LABEL: testi64ult:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: cmpd 5, 6
-; CHECK-NEXT: cmpd 1, 3, 4
-; CHECK-NEXT: crorc 20, 6, 2
-; CHECK-NEXT: bc 12, 20, .LBB11_2
-; CHECK-NEXT: # %bb.1: # %iftrue
-; CHECK-NEXT: mr 3, 7
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB11_2: # %iffalse
-; CHECK-NEXT: mr 3, 8
-; CHECK-NEXT: blr
+; CHECK-LABEL: testi64ult
+; CHECK: crorc [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = icmp eq i64 %c3, %c4
%cmp3tmp = icmp eq i64 %c1, %c2
@@ -293,18 +184,9 @@ iffalse:
}
define i64 @testi64sle(i64 %c1, i64 %c2, i64 %c3, i64 %c4, i64 %a1, i64 %a2) #0 {
-; CHECK-LABEL: testi64sle:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: cmpd 5, 6
-; CHECK-NEXT: cmpd 1, 3, 4
-; CHECK-NEXT: crandc 20, 2, 6
-; CHECK-NEXT: bc 4, 20, .LBB12_2
-; CHECK-NEXT: # %bb.1: # %iffalse
-; CHECK-NEXT: mr 3, 8
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB12_2: # %iftrue
-; CHECK-NEXT: mr 3, 7
-; CHECK-NEXT: blr
+; CHECK-LABEL: testi64sle
+; CHECK: crandc [[REG:[0-9]+]], 2, 6
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = icmp eq i64 %c3, %c4
%cmp3tmp = icmp eq i64 %c1, %c2
@@ -317,18 +199,9 @@ iffalse:
}
define i64 @testi64ule(i64 %c1, i64 %c2, i64 %c3, i64 %c4, i64 %a1, i64 %a2) #0 {
-; CHECK-LABEL: testi64ule:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: cmpd 5, 6
-; CHECK-NEXT: cmpd 1, 3, 4
-; CHECK-NEXT: crandc 20, 6, 2
-; CHECK-NEXT: bc 4, 20, .LBB13_2
-; CHECK-NEXT: # %bb.1: # %iffalse
-; CHECK-NEXT: mr 3, 8
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB13_2: # %iftrue
-; CHECK-NEXT: mr 3, 7
-; CHECK-NEXT: blr
+; CHECK-LABEL: testi64ule
+; CHECK: crandc [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = icmp eq i64 %c3, %c4
%cmp3tmp = icmp eq i64 %c1, %c2
@@ -341,18 +214,9 @@ iffalse:
}
define i64 @testi64eq(i64 %c1, i64 %c2, i64 %c3, i64 %c4, i64 %a1, i64 %a2) #0 {
-; CHECK-LABEL: testi64eq:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: cmpd 5, 6
-; CHECK-NEXT: cmpd 1, 3, 4
-; CHECK-NEXT: crxor 20, 6, 2
-; CHECK-NEXT: bc 4, 20, .LBB14_2
-; CHECK-NEXT: # %bb.1: # %iffalse
-; CHECK-NEXT: mr 3, 8
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB14_2: # %iftrue
-; CHECK-NEXT: mr 3, 7
-; CHECK-NEXT: blr
+; CHECK-LABEL: testi64eq
+; CHECK: crxor [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = icmp eq i64 %c3, %c4
%cmp3tmp = icmp eq i64 %c1, %c2
@@ -365,18 +229,9 @@ iffalse:
}
define i64 @testi64sge(i64 %c1, i64 %c2, i64 %c3, i64 %c4, i64 %a1, i64 %a2) #0 {
-; CHECK-LABEL: testi64sge:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: cmpd 5, 6
-; CHECK-NEXT: cmpd 1, 3, 4
-; CHECK-NEXT: crandc 20, 6, 2
-; CHECK-NEXT: bc 4, 20, .LBB15_2
-; CHECK-NEXT: # %bb.1: # %iffalse
-; CHECK-NEXT: mr 3, 8
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB15_2: # %iftrue
-; CHECK-NEXT: mr 3, 7
-; CHECK-NEXT: blr
+; CHECK-LABEL: testi64sge
+; CHECK: crandc [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = icmp eq i64 %c3, %c4
%cmp3tmp = icmp eq i64 %c1, %c2
@@ -389,18 +244,9 @@ iffalse:
}
define i64 @testi64uge(i64 %c1, i64 %c2, i64 %c3, i64 %c4, i64 %a1, i64 %a2) #0 {
-; CHECK-LABEL: testi64uge:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: cmpd 5, 6
-; CHECK-NEXT: cmpd 1, 3, 4
-; CHECK-NEXT: crandc 20, 2, 6
-; CHECK-NEXT: bc 4, 20, .LBB16_2
-; CHECK-NEXT: # %bb.1: # %iffalse
-; CHECK-NEXT: mr 3, 8
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB16_2: # %iftrue
-; CHECK-NEXT: mr 3, 7
-; CHECK-NEXT: blr
+; CHECK-LABEL: testi64uge
+; CHECK: crandc [[REG:[0-9]+]], 2, 6
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = icmp eq i64 %c3, %c4
%cmp3tmp = icmp eq i64 %c1, %c2
@@ -413,18 +259,9 @@ iffalse:
}
define i64 @testi64sgt(i64 %c1, i64 %c2, i64 %c3, i64 %c4, i64 %a1, i64 %a2) #0 {
-; CHECK-LABEL: testi64sgt:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: cmpd 5, 6
-; CHECK-NEXT: cmpd 1, 3, 4
-; CHECK-NEXT: crorc 20, 6, 2
-; CHECK-NEXT: bc 12, 20, .LBB17_2
-; CHECK-NEXT: # %bb.1: # %iftrue
-; CHECK-NEXT: mr 3, 7
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB17_2: # %iffalse
-; CHECK-NEXT: mr 3, 8
-; CHECK-NEXT: blr
+; CHECK-LABEL: testi64sgt
+; CHECK: crorc [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = icmp eq i64 %c3, %c4
%cmp3tmp = icmp eq i64 %c1, %c2
@@ -437,18 +274,9 @@ iffalse:
}
define i64 @testi64ugt(i64 %c1, i64 %c2, i64 %c3, i64 %c4, i64 %a1, i64 %a2) #0 {
-; CHECK-LABEL: testi64ugt:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: cmpd 5, 6
-; CHECK-NEXT: cmpd 1, 3, 4
-; CHECK-NEXT: crorc 20, 2, 6
-; CHECK-NEXT: bc 12, 20, .LBB18_2
-; CHECK-NEXT: # %bb.1: # %iftrue
-; CHECK-NEXT: mr 3, 7
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB18_2: # %iffalse
-; CHECK-NEXT: mr 3, 8
-; CHECK-NEXT: blr
+; CHECK-LABEL: testi64ugt
+; CHECK: crorc [[REG:[0-9]+]], 2, 6
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = icmp eq i64 %c3, %c4
%cmp3tmp = icmp eq i64 %c1, %c2
@@ -461,18 +289,9 @@ iffalse:
}
define i64 @testi64ne(i64 %c1, i64 %c2, i64 %c3, i64 %c4, i64 %a1, i64 %a2) #0 {
-; CHECK-LABEL: testi64ne:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: cmpd 5, 6
-; CHECK-NEXT: cmpd 1, 3, 4
-; CHECK-NEXT: creqv 20, 6, 2
-; CHECK-NEXT: bc 12, 20, .LBB19_2
-; CHECK-NEXT: # %bb.1: # %iftrue
-; CHECK-NEXT: mr 3, 7
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB19_2: # %iffalse
-; CHECK-NEXT: mr 3, 8
-; CHECK-NEXT: blr
+; CHECK-LABEL: testi64ne
+; CHECK: creqv [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = icmp eq i64 %c3, %c4
%cmp3tmp = icmp eq i64 %c1, %c2
@@ -485,18 +304,9 @@ iffalse:
}
define float @testfloatslt(float %c1, float %c2, float %c3, float %c4, float %a1, float %a2) #0 {
-; CHECK-LABEL: testfloatslt:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: fcmpu 0, 3, 4
-; CHECK-NEXT: fcmpu 1, 1, 2
-; CHECK-NEXT: crorc 20, 2, 6
-; CHECK-NEXT: bc 12, 20, .LBB20_2
-; CHECK-NEXT: # %bb.1: # %iftrue
-; CHECK-NEXT: fmr 1, 5
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB20_2: # %iffalse
-; CHECK-NEXT: fmr 1, 6
-; CHECK-NEXT: blr
+; CHECK-LABEL: testfloatslt
+; CHECK: crorc [[REG:[0-9]+]], 2, 6
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = fcmp oeq float %c3, %c4
%cmp3tmp = fcmp oeq float %c1, %c2
@@ -509,18 +319,9 @@ iffalse:
}
define float @testfloatult(float %c1, float %c2, float %c3, float %c4, float %a1, float %a2) #0 {
-; CHECK-LABEL: testfloatult:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: fcmpu 0, 3, 4
-; CHECK-NEXT: fcmpu 1, 1, 2
-; CHECK-NEXT: crorc 20, 6, 2
-; CHECK-NEXT: bc 12, 20, .LBB21_2
-; CHECK-NEXT: # %bb.1: # %iftrue
-; CHECK-NEXT: fmr 1, 5
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB21_2: # %iffalse
-; CHECK-NEXT: fmr 1, 6
-; CHECK-NEXT: blr
+; CHECK-LABEL: testfloatult
+; CHECK: crorc [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = fcmp oeq float %c3, %c4
%cmp3tmp = fcmp oeq float %c1, %c2
@@ -533,18 +334,9 @@ iffalse:
}
define float @testfloatsle(float %c1, float %c2, float %c3, float %c4, float %a1, float %a2) #0 {
-; CHECK-LABEL: testfloatsle:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: fcmpu 0, 3, 4
-; CHECK-NEXT: fcmpu 1, 1, 2
-; CHECK-NEXT: crandc 20, 2, 6
-; CHECK-NEXT: bc 4, 20, .LBB22_2
-; CHECK-NEXT: # %bb.1: # %iffalse
-; CHECK-NEXT: fmr 1, 6
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB22_2: # %iftrue
-; CHECK-NEXT: fmr 1, 5
-; CHECK-NEXT: blr
+; CHECK-LABEL: testfloatsle
+; CHECK: crandc [[REG:[0-9]+]], 2, 6
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = fcmp oeq float %c3, %c4
%cmp3tmp = fcmp oeq float %c1, %c2
@@ -557,18 +349,9 @@ iffalse:
}
define float @testfloatule(float %c1, float %c2, float %c3, float %c4, float %a1, float %a2) #0 {
-; CHECK-LABEL: testfloatule:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: fcmpu 0, 3, 4
-; CHECK-NEXT: fcmpu 1, 1, 2
-; CHECK-NEXT: crandc 20, 6, 2
-; CHECK-NEXT: bc 4, 20, .LBB23_2
-; CHECK-NEXT: # %bb.1: # %iffalse
-; CHECK-NEXT: fmr 1, 6
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB23_2: # %iftrue
-; CHECK-NEXT: fmr 1, 5
-; CHECK-NEXT: blr
+; CHECK-LABEL: testfloatule
+; CHECK: crandc [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = fcmp oeq float %c3, %c4
%cmp3tmp = fcmp oeq float %c1, %c2
@@ -581,18 +364,9 @@ iffalse:
}
define float @testfloateq(float %c1, float %c2, float %c3, float %c4, float %a1, float %a2) #0 {
-; CHECK-LABEL: testfloateq:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: fcmpu 0, 3, 4
-; CHECK-NEXT: fcmpu 1, 1, 2
-; CHECK-NEXT: crxor 20, 6, 2
-; CHECK-NEXT: bc 4, 20, .LBB24_2
-; CHECK-NEXT: # %bb.1: # %iffalse
-; CHECK-NEXT: fmr 1, 6
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB24_2: # %iftrue
-; CHECK-NEXT: fmr 1, 5
-; CHECK-NEXT: blr
+; CHECK-LABEL: testfloateq
+; CHECK: crxor [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = fcmp oeq float %c3, %c4
%cmp3tmp = fcmp oeq float %c1, %c2
@@ -605,18 +379,9 @@ iffalse:
}
define float @testfloatsge(float %c1, float %c2, float %c3, float %c4, float %a1, float %a2) #0 {
-; CHECK-LABEL: testfloatsge:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: fcmpu 0, 3, 4
-; CHECK-NEXT: fcmpu 1, 1, 2
-; CHECK-NEXT: crandc 20, 6, 2
-; CHECK-NEXT: bc 4, 20, .LBB25_2
-; CHECK-NEXT: # %bb.1: # %iffalse
-; CHECK-NEXT: fmr 1, 6
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB25_2: # %iftrue
-; CHECK-NEXT: fmr 1, 5
-; CHECK-NEXT: blr
+; CHECK-LABEL: testfloatsge
+; CHECK: crandc [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = fcmp oeq float %c3, %c4
%cmp3tmp = fcmp oeq float %c1, %c2
@@ -629,18 +394,9 @@ iffalse:
}
define float @testfloatuge(float %c1, float %c2, float %c3, float %c4, float %a1, float %a2) #0 {
-; CHECK-LABEL: testfloatuge:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: fcmpu 0, 3, 4
-; CHECK-NEXT: fcmpu 1, 1, 2
-; CHECK-NEXT: crandc 20, 2, 6
-; CHECK-NEXT: bc 4, 20, .LBB26_2
-; CHECK-NEXT: # %bb.1: # %iffalse
-; CHECK-NEXT: fmr 1, 6
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB26_2: # %iftrue
-; CHECK-NEXT: fmr 1, 5
-; CHECK-NEXT: blr
+; CHECK-LABEL: testfloatuge
+; CHECK: crandc [[REG:[0-9]+]], 2, 6
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = fcmp oeq float %c3, %c4
%cmp3tmp = fcmp oeq float %c1, %c2
@@ -653,18 +409,9 @@ iffalse:
}
define float @testfloatsgt(float %c1, float %c2, float %c3, float %c4, float %a1, float %a2) #0 {
-; CHECK-LABEL: testfloatsgt:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: fcmpu 0, 3, 4
-; CHECK-NEXT: fcmpu 1, 1, 2
-; CHECK-NEXT: crorc 20, 6, 2
-; CHECK-NEXT: bc 12, 20, .LBB27_2
-; CHECK-NEXT: # %bb.1: # %iftrue
-; CHECK-NEXT: fmr 1, 5
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB27_2: # %iffalse
-; CHECK-NEXT: fmr 1, 6
-; CHECK-NEXT: blr
+; CHECK-LABEL: testfloatsgt
+; CHECK: crorc [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = fcmp oeq float %c3, %c4
%cmp3tmp = fcmp oeq float %c1, %c2
@@ -677,18 +424,9 @@ iffalse:
}
define float @testfloatugt(float %c1, float %c2, float %c3, float %c4, float %a1, float %a2) #0 {
-; CHECK-LABEL: testfloatugt:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: fcmpu 0, 3, 4
-; CHECK-NEXT: fcmpu 1, 1, 2
-; CHECK-NEXT: crorc 20, 2, 6
-; CHECK-NEXT: bc 12, 20, .LBB28_2
-; CHECK-NEXT: # %bb.1: # %iftrue
-; CHECK-NEXT: fmr 1, 5
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB28_2: # %iffalse
-; CHECK-NEXT: fmr 1, 6
-; CHECK-NEXT: blr
+; CHECK-LABEL: testfloatugt
+; CHECK: crorc [[REG:[0-9]+]], 2, 6
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = fcmp oeq float %c3, %c4
%cmp3tmp = fcmp oeq float %c1, %c2
@@ -701,18 +439,9 @@ iffalse:
}
define float @testfloatne(float %c1, float %c2, float %c3, float %c4, float %a1, float %a2) #0 {
-; CHECK-LABEL: testfloatne:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: fcmpu 0, 3, 4
-; CHECK-NEXT: fcmpu 1, 1, 2
-; CHECK-NEXT: creqv 20, 6, 2
-; CHECK-NEXT: bc 12, 20, .LBB29_2
-; CHECK-NEXT: # %bb.1: # %iftrue
-; CHECK-NEXT: fmr 1, 5
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB29_2: # %iffalse
-; CHECK-NEXT: fmr 1, 6
-; CHECK-NEXT: blr
+; CHECK-LABEL: testfloatne
+; CHECK: creqv [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = fcmp oeq float %c3, %c4
%cmp3tmp = fcmp oeq float %c1, %c2
@@ -725,18 +454,9 @@ iffalse:
}
define double @testdoubleslt(double %c1, double %c2, double %c3, double %c4, double %a1, double %a2) #0 {
-; CHECK-LABEL: testdoubleslt:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: fcmpu 0, 3, 4
-; CHECK-NEXT: fcmpu 1, 1, 2
-; CHECK-NEXT: crorc 20, 2, 6
-; CHECK-NEXT: bc 12, 20, .LBB30_2
-; CHECK-NEXT: # %bb.1: # %iftrue
-; CHECK-NEXT: fmr 1, 5
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB30_2: # %iffalse
-; CHECK-NEXT: fmr 1, 6
-; CHECK-NEXT: blr
+; CHECK-LABEL: testdoubleslt
+; CHECK: crorc [[REG:[0-9]+]], 2, 6
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = fcmp oeq double %c3, %c4
%cmp3tmp = fcmp oeq double %c1, %c2
@@ -750,17 +470,8 @@ iffalse:
define double @testdoubleult(double %c1, double %c2, double %c3, double %c4, double %a1, double %a2) #0 {
; CHECK-LABEL: testdoubleult:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: fcmpu 0, 3, 4
-; CHECK-NEXT: fcmpu 1, 1, 2
-; CHECK-NEXT: crorc 20, 6, 2
-; CHECK-NEXT: bc 12, 20, .LBB31_2
-; CHECK-NEXT: # %bb.1: # %iftrue
-; CHECK-NEXT: fmr 1, 5
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB31_2: # %iffalse
-; CHECK-NEXT: fmr 1, 6
-; CHECK-NEXT: blr
+; CHECK: crorc [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = fcmp oeq double %c3, %c4
%cmp3tmp = fcmp oeq double %c1, %c2
@@ -773,18 +484,9 @@ iffalse:
}
define double @testdoublesle(double %c1, double %c2, double %c3, double %c4, double %a1, double %a2) #0 {
-; CHECK-LABEL: testdoublesle:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: fcmpu 0, 3, 4
-; CHECK-NEXT: fcmpu 1, 1, 2
-; CHECK-NEXT: crandc 20, 2, 6
-; CHECK-NEXT: bc 4, 20, .LBB32_2
-; CHECK-NEXT: # %bb.1: # %iffalse
-; CHECK-NEXT: fmr 1, 6
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB32_2: # %iftrue
-; CHECK-NEXT: fmr 1, 5
-; CHECK-NEXT: blr
+; CHECK-LABEL: testdoublesle
+; CHECK: crandc [[REG:[0-9]+]], 2, 6
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = fcmp oeq double %c3, %c4
%cmp3tmp = fcmp oeq double %c1, %c2
@@ -798,17 +500,8 @@ iffalse:
define double @testdoubleule(double %c1, double %c2, double %c3, double %c4, double %a1, double %a2) #0 {
; CHECK-LABEL: testdoubleule:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: fcmpu 0, 3, 4
-; CHECK-NEXT: fcmpu 1, 1, 2
-; CHECK-NEXT: crandc 20, 6, 2
-; CHECK-NEXT: bc 4, 20, .LBB33_2
-; CHECK-NEXT: # %bb.1: # %iffalse
-; CHECK-NEXT: fmr 1, 6
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB33_2: # %iftrue
-; CHECK-NEXT: fmr 1, 5
-; CHECK-NEXT: blr
+; CHECK: crandc [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = fcmp oeq double %c3, %c4
%cmp3tmp = fcmp oeq double %c1, %c2
@@ -821,18 +514,9 @@ iffalse:
}
define double @testdoubleeq(double %c1, double %c2, double %c3, double %c4, double %a1, double %a2) #0 {
-; CHECK-LABEL: testdoubleeq:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: fcmpu 0, 3, 4
-; CHECK-NEXT: fcmpu 1, 1, 2
-; CHECK-NEXT: crxor 20, 6, 2
-; CHECK-NEXT: bc 4, 20, .LBB34_2
-; CHECK-NEXT: # %bb.1: # %iffalse
-; CHECK-NEXT: fmr 1, 6
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB34_2: # %iftrue
-; CHECK-NEXT: fmr 1, 5
-; CHECK-NEXT: blr
+; CHECK-LABEL: testdoubleeq
+; CHECK: crxor [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = fcmp oeq double %c3, %c4
%cmp3tmp = fcmp oeq double %c1, %c2
@@ -845,18 +529,9 @@ iffalse:
}
define double @testdoublesge(double %c1, double %c2, double %c3, double %c4, double %a1, double %a2) #0 {
-; CHECK-LABEL: testdoublesge:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: fcmpu 0, 3, 4
-; CHECK-NEXT: fcmpu 1, 1, 2
-; CHECK-NEXT: crandc 20, 6, 2
-; CHECK-NEXT: bc 4, 20, .LBB35_2
-; CHECK-NEXT: # %bb.1: # %iffalse
-; CHECK-NEXT: fmr 1, 6
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB35_2: # %iftrue
-; CHECK-NEXT: fmr 1, 5
-; CHECK-NEXT: blr
+; CHECK-LABEL: testdoublesge
+; CHECK: crandc [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = fcmp oeq double %c3, %c4
%cmp3tmp = fcmp oeq double %c1, %c2
@@ -869,18 +544,9 @@ iffalse:
}
define double @testdoubleuge(double %c1, double %c2, double %c3, double %c4, double %a1, double %a2) #0 {
-; CHECK-LABEL: testdoubleuge:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: fcmpu 0, 3, 4
-; CHECK-NEXT: fcmpu 1, 1, 2
-; CHECK-NEXT: crandc 20, 2, 6
-; CHECK-NEXT: bc 4, 20, .LBB36_2
-; CHECK-NEXT: # %bb.1: # %iffalse
-; CHECK-NEXT: fmr 1, 6
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB36_2: # %iftrue
-; CHECK-NEXT: fmr 1, 5
-; CHECK-NEXT: blr
+; CHECK-LABEL: testdoubleuge
+; CHECK: crandc [[REG:[0-9]+]], 2, 6
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = fcmp oeq double %c3, %c4
%cmp3tmp = fcmp oeq double %c1, %c2
@@ -894,17 +560,8 @@ iffalse:
define double @testdoublesgt(double %c1, double %c2, double %c3, double %c4, double %a1, double %a2) #0 {
; CHECK-LABEL: testdoublesgt:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: fcmpu 0, 3, 4
-; CHECK-NEXT: fcmpu 1, 1, 2
-; CHECK-NEXT: crorc 20, 6, 2
-; CHECK-NEXT: bc 12, 20, .LBB37_2
-; CHECK-NEXT: # %bb.1: # %iftrue
-; CHECK-NEXT: fmr 1, 5
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB37_2: # %iffalse
-; CHECK-NEXT: fmr 1, 6
-; CHECK-NEXT: blr
+; CHECK: crorc [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = fcmp oeq double %c3, %c4
%cmp3tmp = fcmp oeq double %c1, %c2
@@ -917,18 +574,9 @@ iffalse:
}
define double @testdoubleugt(double %c1, double %c2, double %c3, double %c4, double %a1, double %a2) #0 {
-; CHECK-LABEL: testdoubleugt:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: fcmpu 0, 3, 4
-; CHECK-NEXT: fcmpu 1, 1, 2
-; CHECK-NEXT: crorc 20, 2, 6
-; CHECK-NEXT: bc 12, 20, .LBB38_2
-; CHECK-NEXT: # %bb.1: # %iftrue
-; CHECK-NEXT: fmr 1, 5
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB38_2: # %iffalse
-; CHECK-NEXT: fmr 1, 6
-; CHECK-NEXT: blr
+; CHECK-LABEL: testdoubleugt
+; CHECK: crorc [[REG:[0-9]+]], 2, 6
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = fcmp oeq double %c3, %c4
%cmp3tmp = fcmp oeq double %c1, %c2
@@ -941,18 +589,9 @@ iffalse:
}
define double @testdoublene(double %c1, double %c2, double %c3, double %c4, double %a1, double %a2) #0 {
-; CHECK-LABEL: testdoublene:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: fcmpu 0, 3, 4
-; CHECK-NEXT: fcmpu 1, 1, 2
-; CHECK-NEXT: creqv 20, 6, 2
-; CHECK-NEXT: bc 12, 20, .LBB39_2
-; CHECK-NEXT: # %bb.1: # %iftrue
-; CHECK-NEXT: fmr 1, 5
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB39_2: # %iffalse
-; CHECK-NEXT: fmr 1, 6
-; CHECK-NEXT: blr
+; CHECK-LABEL: testdoublene
+; CHECK: creqv [[REG:[0-9]+]], 6, 2
+; CHECK: bc 12, [[REG]], {{\.[a-zA-Z0-9_]+}}
entry:
%cmp1 = fcmp oeq double %c3, %c4
%cmp3tmp = fcmp oeq double %c1, %c2
diff --git a/llvm/test/CodeGen/PowerPC/memCmpUsedInZeroEqualityComparison.ll b/llvm/test/CodeGen/PowerPC/memCmpUsedInZeroEqualityComparison.ll
index c60efc240160..ce2f93871359 100644
--- a/llvm/test/CodeGen/PowerPC/memCmpUsedInZeroEqualityComparison.ll
+++ b/llvm/test/CodeGen/PowerPC/memCmpUsedInZeroEqualityComparison.ll
@@ -43,12 +43,12 @@ define signext i32 @zeroEqualityTest01(i8* %x, i8* %y) {
; CHECK-NEXT: ld 3, 8(3)
; CHECK-NEXT: ld 4, 8(4)
; CHECK-NEXT: cmpld 3, 4
+; CHECK-NEXT: li 3, 0
; CHECK-NEXT: beq 0, .LBB1_3
; CHECK-NEXT: .LBB1_2: # %res_block
; CHECK-NEXT: li 3, 1
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB1_3:
-; CHECK-NEXT: li 3, 0
+; CHECK-NEXT: .LBB1_3: # %endblock
+; CHECK-NEXT: clrldi 3, 3, 32
; CHECK-NEXT: blr
%call = tail call signext i32 @memcmp(i8* %x, i8* %y, i64 16)
%not.tobool = icmp ne i32 %call, 0
@@ -73,12 +73,12 @@ define signext i32 @zeroEqualityTest03(i8* %x, i8* %y) {
; CHECK-NEXT: lbz 3, 6(3)
; CHECK-NEXT: lbz 4, 6(4)
; CHECK-NEXT: cmplw 3, 4
+; CHECK-NEXT: li 3, 0
; CHECK-NEXT: beq 0, .LBB2_4
; CHECK-NEXT: .LBB2_3: # %res_block
; CHECK-NEXT: li 3, 1
-; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB2_4:
-; CHECK-NEXT: li 3, 0
+; CHECK-NEXT: .LBB2_4: # %endblock
+; CHECK-NEXT: clrldi 3, 3, 32
; CHECK-NEXT: blr
%call = tail call signext i32 @memcmp(i8* %x, i8* %y, i64 7)
%not.lnot = icmp ne i32 %call, 0
@@ -136,16 +136,14 @@ define signext i32 @equalityFoldOneConstant(i8* %X) {
; CHECK-NEXT: sldi 4, 4, 32
; CHECK-NEXT: ori 4, 4, 2
; CHECK-NEXT: cmpld 3, 4
-; CHECK-NEXT: beq 0, .LBB6_4
+; CHECK-NEXT: li 3, 0
+; CHECK-NEXT: beq 0, .LBB6_3
; CHECK-NEXT: .LBB6_2: # %res_block
; CHECK-NEXT: li 3, 1
; CHECK-NEXT: .LBB6_3: # %endblock
; CHECK-NEXT: cntlzw 3, 3
; CHECK-NEXT: srwi 3, 3, 5
; CHECK-NEXT: blr
-; CHECK-NEXT: .LBB6_4:
-; CHECK-NEXT: li 3, 0
-; CHECK-NEXT: b .LBB6_3
%call = tail call signext i32 @memcmp(i8* bitcast ([15 x i32]* @zeroEqualityTest04.buffer1 to i8*), i8* %X, i64 16)
%not.tobool = icmp eq i32 %call, 0
%cond = zext i1 %not.tobool to i32
diff --git a/llvm/test/CodeGen/PowerPC/redundant-copy-after-tail-dup.ll b/llvm/test/CodeGen/PowerPC/redundant-copy-after-tail-dup.ll
index c65acff5f0e1..f6506b3c87f8 100644
--- a/llvm/test/CodeGen/PowerPC/redundant-copy-after-tail-dup.ll
+++ b/llvm/test/CodeGen/PowerPC/redundant-copy-after-tail-dup.ll
@@ -17,28 +17,24 @@
define dso_local i1 @t(%class.A* %this, i32 %color, i32 %vertex) local_unnamed_addr {
; CHECK-P9-LABEL: t:
; CHECK-P9: # %bb.0: # %entry
-; CHECK-P9-NEXT: bc 12, 4*cr5+lt, .LBB0_3
+; CHECK-P9-NEXT: li r5, 1
+; CHECK-P9-NEXT: bc 12, 4*cr5+lt, .LBB0_4
; CHECK-P9-NEXT: # %bb.1: # %land.lhs.true
-; CHECK-P9-NEXT: li r4, 1
-; CHECK-P9-NEXT: bc 4, 4*cr5+lt, .LBB0_4
-; CHECK-P9-NEXT: # %bb.2: # %cleanup16
-; CHECK-P9-NEXT: mr r3, r4
-; CHECK-P9-NEXT: blr
-; CHECK-P9-NEXT: .LBB0_3: # %lor.lhs.false
-; CHECK-P9-NEXT: cmplwi r4, 0
-; CHECK-P9-NEXT: beq cr0, .LBB0_6
-; CHECK-P9-NEXT: .LBB0_4: # %for.inc
+; CHECK-P9-NEXT: bc 12, 4*cr5+lt, .LBB0_5
+; CHECK-P9-NEXT: .LBB0_2: # %for.inc
; CHECK-P9-NEXT: lhz r3, 5308(r3)
; CHECK-P9-NEXT: cmplwi r3, 2
-; CHECK-P9-NEXT: bge- cr0, .LBB0_7
-; CHECK-P9-NEXT: # %bb.5: # %land.lhs.true.1
+; CHECK-P9-NEXT: bge- cr0, .LBB0_6
+; CHECK-P9-NEXT: # %bb.3: # %land.lhs.true.1
; CHECK-P9-NEXT: li r3, 0
; CHECK-P9-NEXT: blr
-; CHECK-P9-NEXT: .LBB0_6:
-; CHECK-P9-NEXT: li r4, 1
-; CHECK-P9-NEXT: mr r3, r4
+; CHECK-P9-NEXT: .LBB0_4: # %lor.lhs.false
+; CHECK-P9-NEXT: cmplwi r4, 0
+; CHECK-P9-NEXT: bne cr0, .LBB0_2
+; CHECK-P9-NEXT: .LBB0_5: # %cleanup16
+; CHECK-P9-NEXT: mr r3, r5
; CHECK-P9-NEXT: blr
-; CHECK-P9-NEXT: .LBB0_7: # %lor.lhs.false.1
+; CHECK-P9-NEXT: .LBB0_6: # %lor.lhs.false.1
entry:
br i1 undef, label %land.lhs.true, label %lor.lhs.false
diff --git a/llvm/test/CodeGen/RISCV/branch.ll b/llvm/test/CodeGen/RISCV/branch.ll
index 562b0fe5cf07..e83449928032 100644
--- a/llvm/test/CodeGen/RISCV/branch.ll
+++ b/llvm/test/CodeGen/RISCV/branch.ll
@@ -6,42 +6,41 @@ define void @foo(i32 %a, i32 *%b, i1 %c) nounwind {
; RV32I-LABEL: foo:
; RV32I: # %bb.0:
; RV32I-NEXT: lw a3, 0(a1)
-; RV32I-NEXT: beq a3, a0, .LBB0_2
+; RV32I-NEXT: beq a3, a0, .LBB0_12
; RV32I-NEXT: # %bb.1: # %test2
; RV32I-NEXT: lw a3, 0(a1)
-; RV32I-NEXT: beq a3, a0, .LBB0_3
-; RV32I-NEXT: .LBB0_2: # %end
-; RV32I-NEXT: ret
-; RV32I-NEXT: .LBB0_3: # %test3
+; RV32I-NEXT: bne a3, a0, .LBB0_12
+; RV32I-NEXT: # %bb.2: # %test3
; RV32I-NEXT: lw a3, 0(a1)
-; RV32I-NEXT: blt a3, a0, .LBB0_2
-; RV32I-NEXT: # %bb.4: # %test4
+; RV32I-NEXT: blt a3, a0, .LBB0_12
+; RV32I-NEXT: # %bb.3: # %test4
; RV32I-NEXT: lw a3, 0(a1)
-; RV32I-NEXT: bge a3, a0, .LBB0_2
-; RV32I-NEXT: # %bb.5: # %test5
+; RV32I-NEXT: bge a3, a0, .LBB0_12
+; RV32I-NEXT: # %bb.4: # %test5
; RV32I-NEXT: lw a3, 0(a1)
-; RV32I-NEXT: bltu a3, a0, .LBB0_2
-; RV32I-NEXT: # %bb.6: # %test6
+; RV32I-NEXT: bltu a3, a0, .LBB0_12
+; RV32I-NEXT: # %bb.5: # %test6
; RV32I-NEXT: lw a3, 0(a1)
-; RV32I-NEXT: bgeu a3, a0, .LBB0_2
-; RV32I-NEXT: # %bb.7: # %test7
+; RV32I-NEXT: bgeu a3, a0, .LBB0_12
+; RV32I-NEXT: # %bb.6: # %test7
; RV32I-NEXT: lw a3, 0(a1)
-; RV32I-NEXT: blt a0, a3, .LBB0_2
-; RV32I-NEXT: # %bb.8: # %test8
+; RV32I-NEXT: blt a0, a3, .LBB0_12
+; RV32I-NEXT: # %bb.7: # %test8
; RV32I-NEXT: lw a3, 0(a1)
-; RV32I-NEXT: bge a0, a3, .LBB0_2
-; RV32I-NEXT: # %bb.9: # %test9
+; RV32I-NEXT: bge a0, a3, .LBB0_12
+; RV32I-NEXT: # %bb.8: # %test9
; RV32I-NEXT: lw a3, 0(a1)
-; RV32I-NEXT: bltu a0, a3, .LBB0_2
-; RV32I-NEXT: # %bb.10: # %test10
+; RV32I-NEXT: bltu a0, a3, .LBB0_12
+; RV32I-NEXT: # %bb.9: # %test10
; RV32I-NEXT: lw a3, 0(a1)
-; RV32I-NEXT: bgeu a0, a3, .LBB0_2
-; RV32I-NEXT: # %bb.11: # %test11
+; RV32I-NEXT: bgeu a0, a3, .LBB0_12
+; RV32I-NEXT: # %bb.10: # %test11
; RV32I-NEXT: lw a0, 0(a1)
; RV32I-NEXT: andi a0, a2, 1
-; RV32I-NEXT: bnez a0, .LBB0_2
-; RV32I-NEXT: # %bb.12: # %test12
+; RV32I-NEXT: bnez a0, .LBB0_12
+; RV32I-NEXT: # %bb.11: # %test12
; RV32I-NEXT: lw a0, 0(a1)
+; RV32I-NEXT: .LBB0_12: # %end
; RV32I-NEXT: ret
%val1 = load volatile i32, i32* %b
%tst1 = icmp eq i32 %val1, %a
diff --git a/llvm/test/CodeGen/RISCV/rv64m-w-insts-legalization.ll b/llvm/test/CodeGen/RISCV/rv64m-w-insts-legalization.ll
index 5635c9ee6872..8af41d260844 100644
--- a/llvm/test/CodeGen/RISCV/rv64m-w-insts-legalization.ll
+++ b/llvm/test/CodeGen/RISCV/rv64m-w-insts-legalization.ll
@@ -5,13 +5,15 @@ define signext i32 @mulw(i32 signext %s, i32 signext %n, i32 signext %k) nounwin
; CHECK-LABEL: mulw:
; CHECK: # %bb.0: # %entry
; CHECK-NEXT: addi a2, zero, 1
-; CHECK-NEXT: bge a0, a1, .LBB0_2
-; CHECK-NEXT: .LBB0_1: # %for.body
+; CHECK-NEXT: bge a0, a1, .LBB0_3
+; CHECK-NEXT: # %bb.1: # %for.body.preheader
+; CHECK-NEXT: addi a2, zero, 1
+; CHECK-NEXT: .LBB0_2: # %for.body
; CHECK-NEXT: # =>This Inner Loop Header: Depth=1
; CHECK-NEXT: mulw a2, a0, a2
; CHECK-NEXT: addiw a0, a0, 1
-; CHECK-NEXT: blt a0, a1, .LBB0_1
-; CHECK-NEXT: .LBB0_2: # %for.cond.cleanup
+; CHECK-NEXT: blt a0, a1, .LBB0_2
+; CHECK-NEXT: .LBB0_3: # %for.cond.cleanup
; CHECK-NEXT: mv a0, a2
; CHECK-NEXT: ret
entry:
diff --git a/llvm/test/CodeGen/SystemZ/int-cmp-37.ll b/llvm/test/CodeGen/SystemZ/int-cmp-37.ll
index 6126000a8978..28cabf7f5bb7 100644
--- a/llvm/test/CodeGen/SystemZ/int-cmp-37.ll
+++ b/llvm/test/CodeGen/SystemZ/int-cmp-37.ll
@@ -1,4 +1,3 @@
-; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; Test 32-bit comparisons in which the second operand is zero-extended
; from a PC-relative i16.
;
@@ -10,16 +9,9 @@
; Check unsigned comparison.
define i32 @f1(i32 %src1) {
; CHECK-LABEL: f1:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: clhrl %r2, g
-; CHECK-NEXT: jhe .LBB0_2
-; CHECK-NEXT: # %bb.1: # %exit
-; CHECK-NEXT: ahi %r2, 1
-; CHECK-NEXT: br %r14
-; CHECK-NEXT: .LBB0_2: # %mulb
-; CHECK-NEXT: msr %r2, %r2
-; CHECK-NEXT: ahi %r2, 1
-; CHECK-NEXT: br %r14
+; CHECK: clhrl %r2, g
+; CHECK-NEXT: jl
+; CHECK: br %r14
entry:
%val = load i16, i16 *@g
%src2 = zext i16 %val to i32
@@ -37,16 +29,8 @@ exit:
; Check signed comparison.
define i32 @f2(i32 %src1) {
; CHECK-LABEL: f2:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: llhrl %r0, g
-; CHECK-NEXT: crjhe %r2, %r0, .LBB1_2
-; CHECK-NEXT: # %bb.1: # %exit
-; CHECK-NEXT: ahi %r2, 1
-; CHECK-NEXT: br %r14
-; CHECK-NEXT: .LBB1_2: # %mulb
-; CHECK-NEXT: msr %r2, %r2
-; CHECK-NEXT: ahi %r2, 1
-; CHECK-NEXT: br %r14
+; CHECK-NOT: clhrl
+; CHECK: br %r14
entry:
%val = load i16, i16 *@g
%src2 = zext i16 %val to i32
@@ -64,14 +48,9 @@ exit:
; Check equality.
define i32 @f3(i32 %src1) {
; CHECK-LABEL: f3:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: clhrl %r2, g
-; CHECK-NEXT: je .LBB2_2
-; CHECK-NEXT: # %bb.1: # %mulb
-; CHECK-NEXT: msr %r2, %r2
-; CHECK-NEXT: .LBB2_2: # %exit
-; CHECK-NEXT: ahi %r2, 1
-; CHECK-NEXT: br %r14
+; CHECK: clhrl %r2, g
+; CHECK-NEXT: je
+; CHECK: br %r14
entry:
%val = load i16, i16 *@g
%src2 = zext i16 %val to i32
@@ -89,16 +68,9 @@ exit:
; Check inequality.
define i32 @f4(i32 %src1) {
; CHECK-LABEL: f4:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: clhrl %r2, g
-; CHECK-NEXT: je .LBB3_2
-; CHECK-NEXT: # %bb.1: # %exit
-; CHECK-NEXT: ahi %r2, 1
-; CHECK-NEXT: br %r14
-; CHECK-NEXT: .LBB3_2: # %mulb
-; CHECK-NEXT: msr %r2, %r2
-; CHECK-NEXT: ahi %r2, 1
-; CHECK-NEXT: br %r14
+; CHECK: clhrl %r2, g
+; CHECK-NEXT: jlh
+; CHECK: br %r14
entry:
%val = load i16, i16 *@g
%src2 = zext i16 %val to i32
@@ -116,17 +88,10 @@ exit:
; Repeat f1 with an unaligned address.
define i32 @f5(i32 %src1) {
; CHECK-LABEL: f5:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: lgrl %r1, h at GOT
-; CHECK-NEXT: llh %r0, 0(%r1)
-; CHECK-NEXT: clrjhe %r2, %r0, .LBB4_2
-; CHECK-NEXT: # %bb.1: # %exit
-; CHECK-NEXT: ahi %r2, 1
-; CHECK-NEXT: br %r14
-; CHECK-NEXT: .LBB4_2: # %mulb
-; CHECK-NEXT: msr %r2, %r2
-; CHECK-NEXT: ahi %r2, 1
-; CHECK-NEXT: br %r14
+; CHECK: lgrl [[REG:%r[0-5]]], h at GOT
+; CHECK: llh [[VAL:%r[0-5]]], 0([[REG]])
+; CHECK: clrjl %r2, [[VAL]],
+; CHECK: br %r14
entry:
%val = load i16, i16 *@h, align 1
%src2 = zext i16 %val to i32
@@ -144,16 +109,9 @@ exit:
; Check the comparison can be reversed if that allows CLHRL to be used.
define i32 @f6(i32 %src2) {
; CHECK-LABEL: f6:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: clhrl %r2, g
-; CHECK-NEXT: jle .LBB5_2
-; CHECK-NEXT: # %bb.1: # %exit
-; CHECK-NEXT: ahi %r2, 1
-; CHECK-NEXT: br %r14
-; CHECK-NEXT: .LBB5_2: # %mulb
-; CHECK-NEXT: msr %r2, %r2
-; CHECK-NEXT: ahi %r2, 1
-; CHECK-NEXT: br %r14
+; CHECK: clhrl %r2, g
+; CHECK-NEXT: jh {{\.L.*}}
+; CHECK: br %r14
entry:
%val = load i16, i16 *@g
%src1 = zext i16 %val to i32
diff --git a/llvm/test/CodeGen/SystemZ/int-cmp-40.ll b/llvm/test/CodeGen/SystemZ/int-cmp-40.ll
index f14f48e1d3d0..09bf903be7b2 100644
--- a/llvm/test/CodeGen/SystemZ/int-cmp-40.ll
+++ b/llvm/test/CodeGen/SystemZ/int-cmp-40.ll
@@ -1,4 +1,3 @@
-; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; Test 64-bit comparisons in which the second operand is zero-extended
; from a PC-relative i16.
;
@@ -10,16 +9,9 @@
; Check unsigned comparison.
define i64 @f1(i64 %src1) {
; CHECK-LABEL: f1:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: clghrl %r2, g
-; CHECK-NEXT: jhe .LBB0_2
-; CHECK-NEXT: # %bb.1: # %exit
-; CHECK-NEXT: la %r2, 1(%r2)
-; CHECK-NEXT: br %r14
-; CHECK-NEXT: .LBB0_2: # %mulb
-; CHECK-NEXT: msgr %r2, %r2
-; CHECK-NEXT: la %r2, 1(%r2)
-; CHECK-NEXT: br %r14
+; CHECK: clghrl %r2, g
+; CHECK-NEXT: jl
+; CHECK: br %r14
entry:
%val = load i16, i16 *@g
%src2 = zext i16 %val to i64
@@ -37,16 +29,8 @@ exit:
; Check signed comparison.
define i64 @f2(i64 %src1) {
; CHECK-LABEL: f2:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: llghrl %r0, g
-; CHECK-NEXT: cgrjhe %r2, %r0, .LBB1_2
-; CHECK-NEXT: # %bb.1: # %exit
-; CHECK-NEXT: la %r2, 1(%r2)
-; CHECK-NEXT: br %r14
-; CHECK-NEXT: .LBB1_2: # %mulb
-; CHECK-NEXT: msgr %r2, %r2
-; CHECK-NEXT: la %r2, 1(%r2)
-; CHECK-NEXT: br %r14
+; CHECK-NOT: clghrl
+; CHECK: br %r14
entry:
%val = load i16, i16 *@g
%src2 = zext i16 %val to i64
@@ -64,14 +48,9 @@ exit:
; Check equality.
define i64 @f3(i64 %src1) {
; CHECK-LABEL: f3:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: clghrl %r2, g
-; CHECK-NEXT: je .LBB2_2
-; CHECK-NEXT: # %bb.1: # %mulb
-; CHECK-NEXT: msgr %r2, %r2
-; CHECK-NEXT: .LBB2_2: # %exit
-; CHECK-NEXT: la %r2, 1(%r2)
-; CHECK-NEXT: br %r14
+; CHECK: clghrl %r2, g
+; CHECK-NEXT: je
+; CHECK: br %r14
entry:
%val = load i16, i16 *@g
%src2 = zext i16 %val to i64
@@ -89,16 +68,9 @@ exit:
; Check inequality.
define i64 @f4(i64 %src1) {
; CHECK-LABEL: f4:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: clghrl %r2, g
-; CHECK-NEXT: je .LBB3_2
-; CHECK-NEXT: # %bb.1: # %exit
-; CHECK-NEXT: la %r2, 1(%r2)
-; CHECK-NEXT: br %r14
-; CHECK-NEXT: .LBB3_2: # %mulb
-; CHECK-NEXT: msgr %r2, %r2
-; CHECK-NEXT: la %r2, 1(%r2)
-; CHECK-NEXT: br %r14
+; CHECK: clghrl %r2, g
+; CHECK-NEXT: jlh
+; CHECK: br %r14
entry:
%val = load i16, i16 *@g
%src2 = zext i16 %val to i64
@@ -116,17 +88,10 @@ exit:
; Repeat f1 with an unaligned address.
define i64 @f5(i64 %src1) {
; CHECK-LABEL: f5:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: lgrl %r1, h at GOT
-; CHECK-NEXT: llgh %r0, 0(%r1)
-; CHECK-NEXT: clgrjhe %r2, %r0, .LBB4_2
-; CHECK-NEXT: # %bb.1: # %exit
-; CHECK-NEXT: la %r2, 1(%r2)
-; CHECK-NEXT: br %r14
-; CHECK-NEXT: .LBB4_2: # %mulb
-; CHECK-NEXT: msgr %r2, %r2
-; CHECK-NEXT: la %r2, 1(%r2)
-; CHECK-NEXT: br %r14
+; CHECK: lgrl [[REG:%r[0-5]]], h at GOT
+; CHECK: llgh [[VAL:%r[0-5]]], 0([[REG]])
+; CHECK: clgrjl %r2, [[VAL]],
+; CHECK: br %r14
entry:
%val = load i16, i16 *@h, align 1
%src2 = zext i16 %val to i64
@@ -144,16 +109,9 @@ exit:
; Check the comparison can be reversed if that allows CLGHRL to be used.
define i64 @f6(i64 %src2) {
; CHECK-LABEL: f6:
-; CHECK: # %bb.0: # %entry
-; CHECK-NEXT: clghrl %r2, g
-; CHECK-NEXT: jle .LBB5_2
-; CHECK-NEXT: # %bb.1: # %exit
-; CHECK-NEXT: la %r2, 1(%r2)
-; CHECK-NEXT: br %r14
-; CHECK-NEXT: .LBB5_2: # %mulb
-; CHECK-NEXT: msgr %r2, %r2
-; CHECK-NEXT: la %r2, 1(%r2)
-; CHECK-NEXT: br %r14
+; CHECK: clghrl %r2, g
+; CHECK-NEXT: jh {{\.L.*}}
+; CHECK: br %r14
entry:
%val = load i16, i16 *@g
%src1 = zext i16 %val to i64
diff --git a/llvm/test/CodeGen/Thumb2/LowOverheadLoops/mve-float-loops.ll b/llvm/test/CodeGen/Thumb2/LowOverheadLoops/mve-float-loops.ll
index 9ad5cdf60ce3..261222f60f17 100644
--- a/llvm/test/CodeGen/Thumb2/LowOverheadLoops/mve-float-loops.ll
+++ b/llvm/test/CodeGen/Thumb2/LowOverheadLoops/mve-float-loops.ll
@@ -903,7 +903,8 @@ define arm_aapcs_vfpcc void @float_int_int_mul(i32* nocapture readonly %a, i32*
; CHECK-NEXT: le lr, .LBB4_4
; CHECK-NEXT: @ %bb.5: @ %middle.block
; CHECK-NEXT: cmp r12, r3
-; CHECK-NEXT: beq .LBB4_8
+; CHECK-NEXT: it eq
+; CHECK-NEXT: popeq {r4, r5, r6, pc}
; CHECK-NEXT: .LBB4_6: @ %for.body.preheader11
; CHECK-NEXT: sub.w lr, r3, r12
; CHECK-NEXT: add.w r0, r0, r12, lsl #2
diff --git a/llvm/test/CodeGen/Thumb2/LowOverheadLoops/varying-outer-2d-reduction.ll b/llvm/test/CodeGen/Thumb2/LowOverheadLoops/varying-outer-2d-reduction.ll
index 0bd7ac870974..32a1c17dbbff 100644
--- a/llvm/test/CodeGen/Thumb2/LowOverheadLoops/varying-outer-2d-reduction.ll
+++ b/llvm/test/CodeGen/Thumb2/LowOverheadLoops/varying-outer-2d-reduction.ll
@@ -17,7 +17,7 @@
define dso_local void @varying_outer_2d_reduction(i16* nocapture readonly %Input, i16* nocapture %Output, i16 signext %Size, i16 signext %N, i16 signext %Scale) local_unnamed_addr {
; ENABLED-LABEL: varying_outer_2d_reduction:
; ENABLED: @ %bb.0: @ %entry
-; ENABLED-NEXT: push.w {r4, r5, r6, r7, r9, r10, r11, lr}
+; ENABLED-NEXT: push.w {r4, r5, r6, r7, r8, r9, r10, lr}
; ENABLED-NEXT: sub sp, #4
; ENABLED-NEXT: cmp r3, #1
; ENABLED-NEXT: str r0, [sp] @ 4-byte Spill
@@ -54,7 +54,7 @@ define dso_local void @varying_outer_2d_reduction(i16* nocapture readonly %Input
; ENABLED-NEXT: subs r0, #4
; ENABLED-NEXT: subs r4, r2, r6
; ENABLED-NEXT: vmov.i32 q0, #0x0
-; ENABLED-NEXT: add.w r11, r7, r0, lsr #2
+; ENABLED-NEXT: add.w r8, r7, r0, lsr #2
; ENABLED-NEXT: mov r7, r10
; ENABLED-NEXT: dlstp.32 lr, r4
; ENABLED-NEXT: ldr r0, [sp] @ 4-byte Reload
@@ -63,9 +63,9 @@ define dso_local void @varying_outer_2d_reduction(i16* nocapture readonly %Input
; ENABLED-NEXT: @ => This Inner Loop Header: Depth=2
; ENABLED-NEXT: vldrh.s32 q1, [r0], #8
; ENABLED-NEXT: vldrh.s32 q2, [r7], #8
-; ENABLED-NEXT: mov lr, r11
+; ENABLED-NEXT: mov lr, r8
; ENABLED-NEXT: vmul.i32 q1, q2, q1
-; ENABLED-NEXT: sub.w r11, r11, #1
+; ENABLED-NEXT: sub.w r8, r8, #1
; ENABLED-NEXT: vshl.s32 q1, r5
; ENABLED-NEXT: vadd.i32 q0, q1, q0
; ENABLED-NEXT: letp lr, .LBB0_6
@@ -75,11 +75,11 @@ define dso_local void @varying_outer_2d_reduction(i16* nocapture readonly %Input
; ENABLED-NEXT: b .LBB0_3
; ENABLED-NEXT: .LBB0_8: @ %for.end17
; ENABLED-NEXT: add sp, #4
-; ENABLED-NEXT: pop.w {r4, r5, r6, r7, r9, r10, r11, pc}
+; ENABLED-NEXT: pop.w {r4, r5, r6, r7, r8, r9, r10, pc}
;
; NOREDUCTIONS-LABEL: varying_outer_2d_reduction:
; NOREDUCTIONS: @ %bb.0: @ %entry
-; NOREDUCTIONS-NEXT: push.w {r4, r5, r6, r7, r9, r10, r11, lr}
+; NOREDUCTIONS-NEXT: push.w {r4, r5, r6, r7, r8, r9, r10, lr}
; NOREDUCTIONS-NEXT: sub sp, #4
; NOREDUCTIONS-NEXT: cmp r3, #1
; NOREDUCTIONS-NEXT: str r0, [sp] @ 4-byte Spill
@@ -116,7 +116,7 @@ define dso_local void @varying_outer_2d_reduction(i16* nocapture readonly %Input
; NOREDUCTIONS-NEXT: subs r0, #4
; NOREDUCTIONS-NEXT: subs r4, r2, r6
; NOREDUCTIONS-NEXT: vmov.i32 q0, #0x0
-; NOREDUCTIONS-NEXT: add.w r11, r7, r0, lsr #2
+; NOREDUCTIONS-NEXT: add.w r8, r7, r0, lsr #2
; NOREDUCTIONS-NEXT: mov r7, r10
; NOREDUCTIONS-NEXT: dlstp.32 lr, r4
; NOREDUCTIONS-NEXT: ldr r0, [sp] @ 4-byte Reload
@@ -125,9 +125,9 @@ define dso_local void @varying_outer_2d_reduction(i16* nocapture readonly %Input
; NOREDUCTIONS-NEXT: @ => This Inner Loop Header: Depth=2
; NOREDUCTIONS-NEXT: vldrh.s32 q1, [r0], #8
; NOREDUCTIONS-NEXT: vldrh.s32 q2, [r7], #8
-; NOREDUCTIONS-NEXT: mov lr, r11
+; NOREDUCTIONS-NEXT: mov lr, r8
; NOREDUCTIONS-NEXT: vmul.i32 q1, q2, q1
-; NOREDUCTIONS-NEXT: sub.w r11, r11, #1
+; NOREDUCTIONS-NEXT: sub.w r8, r8, #1
; NOREDUCTIONS-NEXT: vshl.s32 q1, r5
; NOREDUCTIONS-NEXT: vadd.i32 q0, q1, q0
; NOREDUCTIONS-NEXT: letp lr, .LBB0_6
@@ -137,7 +137,8 @@ define dso_local void @varying_outer_2d_reduction(i16* nocapture readonly %Input
; NOREDUCTIONS-NEXT: b .LBB0_3
; NOREDUCTIONS-NEXT: .LBB0_8: @ %for.end17
; NOREDUCTIONS-NEXT: add sp, #4
-; NOREDUCTIONS-NEXT: pop.w {r4, r5, r6, r7, r9, r10, r11, pc}
+; NOREDUCTIONS-NEXT: pop.w {r4, r5, r6, r7, r8, r9, r10, pc}
+;
entry:
%conv = sext i16 %N to i32
%cmp36 = icmp sgt i16 %N, 0
diff --git a/llvm/test/CodeGen/Thumb2/mve-postinc-distribute.ll b/llvm/test/CodeGen/Thumb2/mve-postinc-distribute.ll
index a3b88cfc2313..2db5bf59ecfa 100644
--- a/llvm/test/CodeGen/Thumb2/mve-postinc-distribute.ll
+++ b/llvm/test/CodeGen/Thumb2/mve-postinc-distribute.ll
@@ -302,7 +302,8 @@ define void @fma8(float* noalias nocapture readonly %A, float* noalias nocapture
; CHECK-NEXT: le lr, .LBB2_4
; CHECK-NEXT: @ %bb.5: @ %middle.block
; CHECK-NEXT: cmp r12, r3
-; CHECK-NEXT: beq .LBB2_8
+; CHECK-NEXT: it eq
+; CHECK-NEXT: popeq {r4, r5, r6, pc}
; CHECK-NEXT: .LBB2_6: @ %for.body.preheader12
; CHECK-NEXT: sub.w lr, r3, r12
; CHECK-NEXT: add.w r0, r0, r12, lsl #2
diff --git a/llvm/test/CodeGen/Thumb2/mve-postinc-lsr.ll b/llvm/test/CodeGen/Thumb2/mve-postinc-lsr.ll
index c99bcf7dccbb..646124e0cf98 100644
--- a/llvm/test/CodeGen/Thumb2/mve-postinc-lsr.ll
+++ b/llvm/test/CodeGen/Thumb2/mve-postinc-lsr.ll
@@ -36,7 +36,8 @@ define void @fma(float* noalias nocapture readonly %A, float* noalias nocapture
; CHECK-NEXT: le lr, .LBB0_4
; CHECK-NEXT: @ %bb.5: @ %middle.block
; CHECK-NEXT: cmp r12, r3
-; CHECK-NEXT: beq .LBB0_8
+; CHECK-NEXT: it eq
+; CHECK-NEXT: popeq {r4, r5, r6, pc}
; CHECK-NEXT: .LBB0_6: @ %for.body.preheader12
; CHECK-NEXT: sub.w lr, r3, r12
; CHECK-NEXT: add.w r0, r0, r12, lsl #2
@@ -219,7 +220,7 @@ define i8* @test(i8* nocapture readonly %input_row, i8* nocapture readonly %inpu
; CHECK-NEXT: ldr r3, [sp, #64]
; CHECK-NEXT: mov.w r11, #0
; CHECK-NEXT: ldr r1, [sp, #16] @ 4-byte Reload
-; CHECK-NEXT: ldr r4, [sp, #56]
+; CHECK-NEXT: ldr.w r9, [sp, #56]
; CHECK-NEXT: add.w r0, r1, r3, lsl #1
; CHECK-NEXT: str r0, [sp, #8] @ 4-byte Spill
; CHECK-NEXT: adds r0, r1, r3
@@ -228,56 +229,56 @@ define i8* @test(i8* nocapture readonly %input_row, i8* nocapture readonly %inpu
; CHECK-NEXT: add r0, r1
; CHECK-NEXT: str r0, [sp] @ 4-byte Spill
; CHECK-NEXT: adds r0, r3, #7
-; CHECK-NEXT: lsr.w r9, r0, #3
+; CHECK-NEXT: lsrs r0, r0, #3
; CHECK-NEXT: b .LBB2_5
; CHECK-NEXT: .LBB2_3: @ in Loop: Header=BB2_5 Depth=1
-; CHECK-NEXT: mov r8, r10
-; CHECK-NEXT: mov r12, r10
-; CHECK-NEXT: mov r6, r10
+; CHECK-NEXT: mov r8, r12
+; CHECK-NEXT: mov r10, r12
+; CHECK-NEXT: mov r6, r12
; CHECK-NEXT: .LBB2_4: @ %for.cond.cleanup23
; CHECK-NEXT: @ in Loop: Header=BB2_5 Depth=1
-; CHECK-NEXT: ldr r1, [sp, #72]
-; CHECK-NEXT: add.w r0, r12, r8
-; CHECK-NEXT: add r0, r6
-; CHECK-NEXT: add r0, r10
-; CHECK-NEXT: strb.w r0, [r1, r11]
+; CHECK-NEXT: ldr r3, [sp, #72]
+; CHECK-NEXT: add.w r1, r10, r8
+; CHECK-NEXT: add r1, r6
+; CHECK-NEXT: add r1, r12
+; CHECK-NEXT: strb.w r1, [r3, r11]
; CHECK-NEXT: add.w r11, r11, #1
; CHECK-NEXT: cmp r11, r2
; CHECK-NEXT: beq .LBB2_8
; CHECK-NEXT: .LBB2_5: @ %for.body
; CHECK-NEXT: @ =>This Loop Header: Depth=1
; CHECK-NEXT: @ Child Loop BB2_7 Depth 2
-; CHECK-NEXT: ldr r0, [sp, #68]
-; CHECK-NEXT: subs.w lr, r9, r9
-; CHECK-NEXT: ldr.w r10, [r0, r11, lsl #2]
+; CHECK-NEXT: ldr r1, [sp, #68]
+; CHECK-NEXT: subs.w lr, r0, r0
+; CHECK-NEXT: ldr.w r12, [r1, r11, lsl #2]
; CHECK-NEXT: ble .LBB2_3
; CHECK-NEXT: @ %bb.6: @ %for.body24.preheader
; CHECK-NEXT: @ in Loop: Header=BB2_5 Depth=1
-; CHECK-NEXT: ldr r1, [sp, #64]
-; CHECK-NEXT: mov r6, r10
-; CHECK-NEXT: ldr r0, [sp, #12] @ 4-byte Reload
+; CHECK-NEXT: ldr r3, [sp, #64]
+; CHECK-NEXT: mov r6, r12
+; CHECK-NEXT: ldr r1, [sp, #12] @ 4-byte Reload
; CHECK-NEXT: dls lr, lr
-; CHECK-NEXT: ldr r3, [sp, #8] @ 4-byte Reload
-; CHECK-NEXT: mov r12, r10
-; CHECK-NEXT: mla r7, r11, r1, r0
+; CHECK-NEXT: ldr r5, [sp, #8] @ 4-byte Reload
+; CHECK-NEXT: mov r10, r12
+; CHECK-NEXT: mla r7, r11, r3, r1
; CHECK-NEXT: ldr r1, [sp, #16] @ 4-byte Reload
-; CHECK-NEXT: ldrd r5, r0, [sp] @ 8-byte Folded Reload
-; CHECK-NEXT: mov r8, r10
+; CHECK-NEXT: ldrd r4, r3, [sp] @ 8-byte Folded Reload
+; CHECK-NEXT: mov r8, r12
; CHECK-NEXT: .LBB2_7: @ %for.body24
; CHECK-NEXT: @ Parent Loop BB2_5 Depth=1
; CHECK-NEXT: @ => This Inner Loop Header: Depth=2
-; CHECK-NEXT: vldrb.s16 q0, [r5], #8
-; CHECK-NEXT: vadd.i16 q1, q0, r4
+; CHECK-NEXT: vldrb.s16 q0, [r4], #8
+; CHECK-NEXT: vadd.i16 q1, q0, r9
; CHECK-NEXT: vldrb.s16 q0, [r7], #8
-; CHECK-NEXT: vmlava.s16 r10, q0, q1
-; CHECK-NEXT: vldrb.s16 q1, [r3], #8
-; CHECK-NEXT: vadd.i16 q1, q1, r4
-; CHECK-NEXT: vmlava.s16 r6, q0, q1
-; CHECK-NEXT: vldrb.s16 q1, [r0], #8
-; CHECK-NEXT: vadd.i16 q1, q1, r4
; CHECK-NEXT: vmlava.s16 r12, q0, q1
+; CHECK-NEXT: vldrb.s16 q1, [r5], #8
+; CHECK-NEXT: vadd.i16 q1, q1, r9
+; CHECK-NEXT: vmlava.s16 r6, q0, q1
+; CHECK-NEXT: vldrb.s16 q1, [r3], #8
+; CHECK-NEXT: vadd.i16 q1, q1, r9
+; CHECK-NEXT: vmlava.s16 r10, q0, q1
; CHECK-NEXT: vldrb.s16 q1, [r1], #8
-; CHECK-NEXT: vadd.i16 q1, q1, r4
+; CHECK-NEXT: vadd.i16 q1, q1, r9
; CHECK-NEXT: vmlava.s16 r8, q0, q1
; CHECK-NEXT: le lr, .LBB2_7
; CHECK-NEXT: b .LBB2_4
@@ -402,7 +403,7 @@ define i8* @test_optsize(i8* nocapture readonly %input_row, i8* nocapture readon
; CHECK-NEXT: ldr r3, [sp, #64]
; CHECK-NEXT: mov.w r11, #0
; CHECK-NEXT: ldr r1, [sp, #16] @ 4-byte Reload
-; CHECK-NEXT: ldr r4, [sp, #56]
+; CHECK-NEXT: ldr.w r9, [sp, #56]
; CHECK-NEXT: add.w r0, r1, r3, lsl #1
; CHECK-NEXT: str r0, [sp, #8] @ 4-byte Spill
; CHECK-NEXT: adds r0, r1, r3
@@ -411,55 +412,55 @@ define i8* @test_optsize(i8* nocapture readonly %input_row, i8* nocapture readon
; CHECK-NEXT: add r0, r1
; CHECK-NEXT: str r0, [sp] @ 4-byte Spill
; CHECK-NEXT: adds r0, r3, #7
-; CHECK-NEXT: lsr.w r9, r0, #3
+; CHECK-NEXT: lsrs r0, r0, #3
; CHECK-NEXT: .LBB3_3: @ %for.body
; CHECK-NEXT: @ =>This Loop Header: Depth=1
; CHECK-NEXT: @ Child Loop BB3_5 Depth 2
-; CHECK-NEXT: ldr r0, [sp, #68]
-; CHECK-NEXT: subs.w lr, r9, r9
-; CHECK-NEXT: ldr.w r10, [r0, r11, lsl #2]
+; CHECK-NEXT: ldr r1, [sp, #68]
+; CHECK-NEXT: subs.w lr, r0, r0
+; CHECK-NEXT: ldr.w r12, [r1, r11, lsl #2]
; CHECK-NEXT: ble .LBB3_6
; CHECK-NEXT: @ %bb.4: @ %for.body24.preheader
; CHECK-NEXT: @ in Loop: Header=BB3_3 Depth=1
-; CHECK-NEXT: ldr r1, [sp, #64]
-; CHECK-NEXT: mov r6, r10
-; CHECK-NEXT: ldr r0, [sp, #12] @ 4-byte Reload
+; CHECK-NEXT: ldr r3, [sp, #64]
+; CHECK-NEXT: mov r6, r12
+; CHECK-NEXT: ldr r1, [sp, #12] @ 4-byte Reload
; CHECK-NEXT: dls lr, lr
-; CHECK-NEXT: ldr r3, [sp, #8] @ 4-byte Reload
-; CHECK-NEXT: mov r12, r10
-; CHECK-NEXT: mla r7, r11, r1, r0
+; CHECK-NEXT: ldr r5, [sp, #8] @ 4-byte Reload
+; CHECK-NEXT: mov r10, r12
+; CHECK-NEXT: mla r7, r11, r3, r1
; CHECK-NEXT: ldr r1, [sp, #16] @ 4-byte Reload
-; CHECK-NEXT: ldrd r5, r0, [sp] @ 8-byte Folded Reload
-; CHECK-NEXT: mov r8, r10
+; CHECK-NEXT: ldrd r4, r3, [sp] @ 8-byte Folded Reload
+; CHECK-NEXT: mov r8, r12
; CHECK-NEXT: .LBB3_5: @ %for.body24
; CHECK-NEXT: @ Parent Loop BB3_3 Depth=1
; CHECK-NEXT: @ => This Inner Loop Header: Depth=2
-; CHECK-NEXT: vldrb.s16 q0, [r5], #8
-; CHECK-NEXT: vadd.i16 q1, q0, r4
+; CHECK-NEXT: vldrb.s16 q0, [r4], #8
+; CHECK-NEXT: vadd.i16 q1, q0, r9
; CHECK-NEXT: vldrb.s16 q0, [r7], #8
-; CHECK-NEXT: vmlava.s16 r10, q0, q1
-; CHECK-NEXT: vldrb.s16 q1, [r3], #8
-; CHECK-NEXT: vadd.i16 q1, q1, r4
-; CHECK-NEXT: vmlava.s16 r6, q0, q1
-; CHECK-NEXT: vldrb.s16 q1, [r0], #8
-; CHECK-NEXT: vadd.i16 q1, q1, r4
; CHECK-NEXT: vmlava.s16 r12, q0, q1
+; CHECK-NEXT: vldrb.s16 q1, [r5], #8
+; CHECK-NEXT: vadd.i16 q1, q1, r9
+; CHECK-NEXT: vmlava.s16 r6, q0, q1
+; CHECK-NEXT: vldrb.s16 q1, [r3], #8
+; CHECK-NEXT: vadd.i16 q1, q1, r9
+; CHECK-NEXT: vmlava.s16 r10, q0, q1
; CHECK-NEXT: vldrb.s16 q1, [r1], #8
-; CHECK-NEXT: vadd.i16 q1, q1, r4
+; CHECK-NEXT: vadd.i16 q1, q1, r9
; CHECK-NEXT: vmlava.s16 r8, q0, q1
; CHECK-NEXT: le lr, .LBB3_5
; CHECK-NEXT: b .LBB3_7
; CHECK-NEXT: .LBB3_6: @ in Loop: Header=BB3_3 Depth=1
-; CHECK-NEXT: mov r8, r10
-; CHECK-NEXT: mov r12, r10
-; CHECK-NEXT: mov r6, r10
+; CHECK-NEXT: mov r8, r12
+; CHECK-NEXT: mov r10, r12
+; CHECK-NEXT: mov r6, r12
; CHECK-NEXT: .LBB3_7: @ %for.cond.cleanup23
; CHECK-NEXT: @ in Loop: Header=BB3_3 Depth=1
-; CHECK-NEXT: ldr r1, [sp, #72]
-; CHECK-NEXT: add.w r0, r12, r8
-; CHECK-NEXT: add r0, r6
-; CHECK-NEXT: add r0, r10
-; CHECK-NEXT: strb.w r0, [r1, r11]
+; CHECK-NEXT: ldr r3, [sp, #72]
+; CHECK-NEXT: add.w r1, r10, r8
+; CHECK-NEXT: add r1, r6
+; CHECK-NEXT: add r1, r12
+; CHECK-NEXT: strb.w r1, [r3, r11]
; CHECK-NEXT: add.w r11, r11, #1
; CHECK-NEXT: cmp r11, r2
; CHECK-NEXT: bne .LBB3_3
@@ -736,8 +737,8 @@ define i8* @signext(i8* %input_row, i8* %input_col, i16 zeroext %output_ch, i16
; CHECK-NEXT: ldr r1, [sp, #16] @ 4-byte Reload
; CHECK-NEXT: mov r8, r12
; CHECK-NEXT: mla r5, r11, r3, r0
-; CHECK-NEXT: ldr r7, [sp, #8] @ 4-byte Reload
-; CHECK-NEXT: ldrd r4, r0, [sp] @ 8-byte Folded Reload
+; CHECK-NEXT: ldr r0, [sp, #8] @ 4-byte Reload
+; CHECK-NEXT: ldrd r4, r7, [sp] @ 8-byte Folded Reload
; CHECK-NEXT: mov r10, r12
; CHECK-NEXT: .LBB5_7: @ %for.body24
; CHECK-NEXT: @ Parent Loop BB5_5 Depth=1
@@ -746,10 +747,10 @@ define i8* @signext(i8* %input_row, i8* %input_col, i16 zeroext %output_ch, i16
; CHECK-NEXT: vadd.i16 q2, q1, q0
; CHECK-NEXT: vldrb.s16 q1, [r5], #8
; CHECK-NEXT: vmlava.s16 r12, q1, q2
-; CHECK-NEXT: vldrb.s16 q2, [r7], #8
+; CHECK-NEXT: vldrb.s16 q2, [r0], #8
; CHECK-NEXT: vadd.i16 q2, q2, q0
; CHECK-NEXT: vmlava.s16 r6, q1, q2
-; CHECK-NEXT: vldrb.s16 q2, [r0], #8
+; CHECK-NEXT: vldrb.s16 q2, [r7], #8
; CHECK-NEXT: vadd.i16 q2, q2, q0
; CHECK-NEXT: vmlava.s16 r8, q1, q2
; CHECK-NEXT: vldrb.s16 q2, [r1], #8
@@ -907,8 +908,8 @@ define i8* @signext_optsize(i8* %input_row, i8* %input_col, i16 zeroext %output_
; CHECK-NEXT: ldr r1, [sp, #16] @ 4-byte Reload
; CHECK-NEXT: mov r8, r12
; CHECK-NEXT: mla r5, r11, r3, r0
-; CHECK-NEXT: ldr r7, [sp, #8] @ 4-byte Reload
-; CHECK-NEXT: ldrd r4, r0, [sp] @ 8-byte Folded Reload
+; CHECK-NEXT: ldr r0, [sp, #8] @ 4-byte Reload
+; CHECK-NEXT: ldrd r4, r7, [sp] @ 8-byte Folded Reload
; CHECK-NEXT: mov r10, r12
; CHECK-NEXT: .LBB6_5: @ %for.body24
; CHECK-NEXT: @ Parent Loop BB6_3 Depth=1
@@ -917,10 +918,10 @@ define i8* @signext_optsize(i8* %input_row, i8* %input_col, i16 zeroext %output_
; CHECK-NEXT: vadd.i16 q2, q1, q0
; CHECK-NEXT: vldrb.s16 q1, [r5], #8
; CHECK-NEXT: vmlava.s16 r12, q1, q2
-; CHECK-NEXT: vldrb.s16 q2, [r7], #8
+; CHECK-NEXT: vldrb.s16 q2, [r0], #8
; CHECK-NEXT: vadd.i16 q2, q2, q0
; CHECK-NEXT: vmlava.s16 r6, q1, q2
-; CHECK-NEXT: vldrb.s16 q2, [r0], #8
+; CHECK-NEXT: vldrb.s16 q2, [r7], #8
; CHECK-NEXT: vadd.i16 q2, q2, q0
; CHECK-NEXT: vmlava.s16 r8, q1, q2
; CHECK-NEXT: vldrb.s16 q2, [r1], #8
diff --git a/llvm/test/CodeGen/Thumb2/mve-satmul-loops.ll b/llvm/test/CodeGen/Thumb2/mve-satmul-loops.ll
index 3ec0e4644276..9897b607d6b3 100644
--- a/llvm/test/CodeGen/Thumb2/mve-satmul-loops.ll
+++ b/llvm/test/CodeGen/Thumb2/mve-satmul-loops.ll
@@ -20,50 +20,50 @@ define arm_aapcs_vfpcc void @ssatmul_s_q31(i32* nocapture readonly %pSrcA, i32*
; CHECK-NEXT: @ %bb.2:
; CHECK-NEXT: movs r7, #0
; CHECK-NEXT: mov r12, r0
-; CHECK-NEXT: mov r6, r1
+; CHECK-NEXT: mov r9, r1
; CHECK-NEXT: mov r11, r2
; CHECK-NEXT: b .LBB0_6
; CHECK-NEXT: .LBB0_3: @ %vector.ph
; CHECK-NEXT: str r3, [sp, #4] @ 4-byte Spill
; CHECK-NEXT: bic r3, r3, #1
; CHECK-NEXT: subs r7, r3, #2
-; CHECK-NEXT: movs r6, #1
; CHECK-NEXT: adr r4, .LCPI0_0
+; CHECK-NEXT: movs r6, #1
; CHECK-NEXT: add.w r11, r2, r3, lsl #2
; CHECK-NEXT: add.w lr, r6, r7, lsr #1
-; CHECK-NEXT: add.w r6, r1, r3, lsl #2
+; CHECK-NEXT: add.w r9, r1, r3, lsl #2
; CHECK-NEXT: add.w r12, r0, r3, lsl #2
; CHECK-NEXT: vldrw.u32 q0, [r4]
-; CHECK-NEXT: mvn r10, #-2147483648
; CHECK-NEXT: vmvn.i32 q1, #0x80000000
+; CHECK-NEXT: mov.w r10, #-1
; CHECK-NEXT: dls lr, lr
; CHECK-NEXT: str r3, [sp] @ 4-byte Spill
; CHECK-NEXT: .LBB0_4: @ %vector.body
; CHECK-NEXT: @ =>This Inner Loop Header: Depth=1
-; CHECK-NEXT: ldrd r4, r8, [r0]
+; CHECK-NEXT: ldrd r4, r5, [r0]
; CHECK-NEXT: adds r0, #8
-; CHECK-NEXT: ldrd r7, r5, [r1]
+; CHECK-NEXT: ldrd r7, r6, [r1]
; CHECK-NEXT: adds r1, #8
-; CHECK-NEXT: smull r8, r5, r5, r8
+; CHECK-NEXT: smull r8, r5, r6, r5
; CHECK-NEXT: smull r4, r7, r7, r4
; CHECK-NEXT: asrl r8, r5, #31
; CHECK-NEXT: asrl r4, r7, #31
-; CHECK-NEXT: rsbs.w r9, r4, #-2147483648
+; CHECK-NEXT: rsbs.w r3, r4, #-2147483648
; CHECK-NEXT: vmov.32 q4[0], r4
-; CHECK-NEXT: mov.w r9, #-1
-; CHECK-NEXT: sbcs.w r3, r9, r7
-; CHECK-NEXT: mov.w r3, #0
+; CHECK-NEXT: sbcs.w r3, r10, r7
; CHECK-NEXT: vmov.32 q4[1], r7
+; CHECK-NEXT: mov.w r3, #0
+; CHECK-NEXT: vmov.32 q4[2], r8
; CHECK-NEXT: it lt
; CHECK-NEXT: movlt r3, #1
; CHECK-NEXT: cmp r3, #0
; CHECK-NEXT: csetm r3, ne
-; CHECK-NEXT: vmov.32 q4[2], r8
-; CHECK-NEXT: vmov.32 q2[0], r3
; CHECK-NEXT: vmov.32 q4[3], r5
+; CHECK-NEXT: vmov.32 q2[0], r3
; CHECK-NEXT: vmov.32 q2[1], r3
; CHECK-NEXT: rsbs.w r3, r8, #-2147483648
-; CHECK-NEXT: sbcs.w r3, r9, r5
+; CHECK-NEXT: sbcs.w r3, r10, r5
+; CHECK-NEXT: mvn r5, #-2147483648
; CHECK-NEXT: mov.w r3, #0
; CHECK-NEXT: it lt
; CHECK-NEXT: movlt r3, #1
@@ -76,7 +76,7 @@ define arm_aapcs_vfpcc void @ssatmul_s_q31(i32* nocapture readonly %pSrcA, i32*
; CHECK-NEXT: vorr q2, q2, q3
; CHECK-NEXT: vmov r4, s8
; CHECK-NEXT: vmov r3, s9
-; CHECK-NEXT: subs.w r4, r4, r10
+; CHECK-NEXT: subs r4, r4, r5
; CHECK-NEXT: sbcs r3, r3, #0
; CHECK-NEXT: vmov r4, s10
; CHECK-NEXT: mov.w r3, #0
@@ -87,7 +87,7 @@ define arm_aapcs_vfpcc void @ssatmul_s_q31(i32* nocapture readonly %pSrcA, i32*
; CHECK-NEXT: vmov.32 q3[0], r3
; CHECK-NEXT: vmov.32 q3[1], r3
; CHECK-NEXT: vmov r3, s11
-; CHECK-NEXT: subs.w r4, r4, r10
+; CHECK-NEXT: subs r4, r4, r5
; CHECK-NEXT: sbcs r3, r3, #0
; CHECK-NEXT: mov.w r3, #0
; CHECK-NEXT: it lt
@@ -116,7 +116,7 @@ define arm_aapcs_vfpcc void @ssatmul_s_q31(i32* nocapture readonly %pSrcA, i32*
; CHECK-NEXT: .LBB0_7: @ %for.body
; CHECK-NEXT: @ =>This Inner Loop Header: Depth=1
; CHECK-NEXT: ldr r3, [r12], #4
-; CHECK-NEXT: ldr r4, [r6], #4
+; CHECK-NEXT: ldr r4, [r9], #4
; CHECK-NEXT: smull r4, r3, r4, r3
; CHECK-NEXT: asrl r4, r3, #31
; CHECK-NEXT: subs r5, r1, r4
@@ -908,41 +908,36 @@ for.body: ; preds = %for.body.preheader,
define arm_aapcs_vfpcc void @usatmul_4_q31(i32* nocapture readonly %pSrcA, i32* nocapture readonly %pSrcB, i32* noalias nocapture %pDst, i32 %N) {
; CHECK-LABEL: usatmul_4_q31:
; CHECK: @ %bb.0: @ %entry
-; CHECK-NEXT: .save {r4, r5, r6, r7, r8, r9, r10, r11, lr}
-; CHECK-NEXT: push.w {r4, r5, r6, r7, r8, r9, r10, r11, lr}
-; CHECK-NEXT: .pad #4
-; CHECK-NEXT: sub sp, #4
+; CHECK-NEXT: .save {r4, r5, r6, r7, r8, r9, r10, lr}
+; CHECK-NEXT: push.w {r4, r5, r6, r7, r8, r9, r10, lr}
; CHECK-NEXT: .vsave {d8, d9}
; CHECK-NEXT: vpush {d8, d9}
; CHECK-NEXT: cmp r3, #0
; CHECK-NEXT: beq.w .LBB4_8
; CHECK-NEXT: @ %bb.1: @ %for.body.preheader
-; CHECK-NEXT: mov.w r11, #0
+; CHECK-NEXT: mov.w r8, #0
; CHECK-NEXT: cmp r3, #3
; CHECK-NEXT: bhi .LBB4_3
; CHECK-NEXT: @ %bb.2:
; CHECK-NEXT: mov r12, r0
-; CHECK-NEXT: mov r8, r1
-; CHECK-NEXT: mov r9, r2
+; CHECK-NEXT: mov r9, r1
+; CHECK-NEXT: mov r10, r2
; CHECK-NEXT: b .LBB4_6
; CHECK-NEXT: .LBB4_3: @ %vector.ph
-; CHECK-NEXT: bic r11, r3, #3
+; CHECK-NEXT: bic r8, r3, #3
; CHECK-NEXT: movs r6, #1
-; CHECK-NEXT: sub.w r7, r11, #4
-; CHECK-NEXT: add.w r9, r2, r11, lsl #2
-; CHECK-NEXT: add.w r8, r1, r11, lsl #2
-; CHECK-NEXT: add.w r7, r6, r7, lsr #2
-; CHECK-NEXT: add.w r12, r0, r11, lsl #2
-; CHECK-NEXT: mov r10, r7
+; CHECK-NEXT: sub.w r7, r8, #4
+; CHECK-NEXT: add.w r10, r2, r8, lsl #2
+; CHECK-NEXT: add.w r9, r1, r8, lsl #2
+; CHECK-NEXT: add.w lr, r6, r7, lsr #2
+; CHECK-NEXT: add.w r12, r0, r8, lsl #2
+; CHECK-NEXT: dls lr, lr
; CHECK-NEXT: .LBB4_4: @ %vector.body
; CHECK-NEXT: @ =>This Inner Loop Header: Depth=1
; CHECK-NEXT: vldrw.u32 q0, [r0], #16
; CHECK-NEXT: vldrw.u32 q2, [r1], #16
-; CHECK-NEXT: mov lr, r10
; CHECK-NEXT: vmov.f32 s4, s2
-; CHECK-NEXT: sub.w lr, lr, #1
; CHECK-NEXT: vmov.f32 s12, s10
-; CHECK-NEXT: mov r10, lr
; CHECK-NEXT: vmov.f32 s6, s3
; CHECK-NEXT: vmov.f32 s14, s11
; CHECK-NEXT: vmullb.u32 q4, q3, q1
@@ -1007,31 +1002,28 @@ define arm_aapcs_vfpcc void @usatmul_4_q31(i32* nocapture readonly %pSrcA, i32*
; CHECK-NEXT: vmov.f32 s2, s4
; CHECK-NEXT: vmov.f32 s3, s6
; CHECK-NEXT: vstrb.8 q0, [r2], #16
-; CHECK-NEXT: cmp.w lr, #0
-; CHECK-NEXT: bne.w .LBB4_4
-; CHECK-NEXT: b .LBB4_5
-; CHECK-NEXT: .LBB4_5: @ %middle.block
-; CHECK-NEXT: cmp r11, r3
+; CHECK-NEXT: le lr, .LBB4_4
+; CHECK-NEXT: @ %bb.5: @ %middle.block
+; CHECK-NEXT: cmp r8, r3
; CHECK-NEXT: beq .LBB4_8
; CHECK-NEXT: .LBB4_6: @ %for.body.preheader21
-; CHECK-NEXT: sub.w lr, r3, r11
+; CHECK-NEXT: sub.w lr, r3, r8
; CHECK-NEXT: dls lr, lr
; CHECK-NEXT: .LBB4_7: @ %for.body
; CHECK-NEXT: @ =>This Inner Loop Header: Depth=1
; CHECK-NEXT: ldr r0, [r12], #4
-; CHECK-NEXT: ldr r1, [r8], #4
+; CHECK-NEXT: ldr r1, [r9], #4
; CHECK-NEXT: umull r0, r1, r1, r0
; CHECK-NEXT: lsrl r0, r1, #31
; CHECK-NEXT: subs.w r2, r0, #-1
; CHECK-NEXT: sbcs r1, r1, #0
; CHECK-NEXT: it hs
; CHECK-NEXT: movhs.w r0, #-1
-; CHECK-NEXT: str r0, [r9], #4
+; CHECK-NEXT: str r0, [r10], #4
; CHECK-NEXT: le lr, .LBB4_7
; CHECK-NEXT: .LBB4_8: @ %for.cond.cleanup
; CHECK-NEXT: vpop {d8, d9}
-; CHECK-NEXT: add sp, #4
-; CHECK-NEXT: pop.w {r4, r5, r6, r7, r8, r9, r10, r11, pc}
+; CHECK-NEXT: pop.w {r4, r5, r6, r7, r8, r9, r10, pc}
entry:
%cmp8 = icmp eq i32 %N, 0
br i1 %cmp8, label %for.cond.cleanup, label %for.body.preheader
@@ -1144,7 +1136,8 @@ define arm_aapcs_vfpcc void @ssatmul_4_q15(i16* nocapture readonly %pSrcA, i16*
; CHECK-NEXT: le lr, .LBB5_4
; CHECK-NEXT: @ %bb.5: @ %middle.block
; CHECK-NEXT: cmp r5, r3
-; CHECK-NEXT: beq .LBB5_8
+; CHECK-NEXT: it eq
+; CHECK-NEXT: popeq {r4, r5, r6, pc}
; CHECK-NEXT: .LBB5_6: @ %for.body.preheader21
; CHECK-NEXT: movw r0, #32768
; CHECK-NEXT: sub.w lr, r3, r5
@@ -1284,7 +1277,8 @@ define arm_aapcs_vfpcc void @ssatmul_8_q15(i16* nocapture readonly %pSrcA, i16*
; CHECK-NEXT: le lr, .LBB6_4
; CHECK-NEXT: @ %bb.5: @ %middle.block
; CHECK-NEXT: cmp r5, r3
-; CHECK-NEXT: beq .LBB6_8
+; CHECK-NEXT: it eq
+; CHECK-NEXT: popeq {r4, r5, r6, pc}
; CHECK-NEXT: .LBB6_6: @ %for.body.preheader21
; CHECK-NEXT: movw r0, #32768
; CHECK-NEXT: sub.w lr, r3, r5
@@ -1421,7 +1415,8 @@ define arm_aapcs_vfpcc void @ssatmul_8i_q15(i16* nocapture readonly %pSrcA, i16*
; CHECK-NEXT: le lr, .LBB7_4
; CHECK-NEXT: @ %bb.5: @ %middle.block
; CHECK-NEXT: cmp r5, r3
-; CHECK-NEXT: beq .LBB7_8
+; CHECK-NEXT: it eq
+; CHECK-NEXT: popeq {r4, r5, r6, pc}
; CHECK-NEXT: .LBB7_6: @ %for.body.preheader21
; CHECK-NEXT: movw r0, #32768
; CHECK-NEXT: sub.w lr, r3, r5
@@ -1964,7 +1959,8 @@ define arm_aapcs_vfpcc void @usatmul_4_q15(i16* nocapture readonly %pSrcA, i16*
; CHECK-NEXT: le lr, .LBB11_4
; CHECK-NEXT: @ %bb.5: @ %middle.block
; CHECK-NEXT: cmp r5, r3
-; CHECK-NEXT: beq .LBB11_8
+; CHECK-NEXT: it eq
+; CHECK-NEXT: popeq {r4, r5, r6, pc}
; CHECK-NEXT: .LBB11_6: @ %for.body.preheader21
; CHECK-NEXT: sub.w lr, r3, r5
; CHECK-NEXT: movw r0, #65535
@@ -2097,7 +2093,8 @@ define arm_aapcs_vfpcc void @usatmul_8_q15(i16* nocapture readonly %pSrcA, i16*
; CHECK-NEXT: le lr, .LBB12_4
; CHECK-NEXT: @ %bb.5: @ %middle.block
; CHECK-NEXT: cmp r5, r3
-; CHECK-NEXT: beq .LBB12_8
+; CHECK-NEXT: it eq
+; CHECK-NEXT: popeq {r4, r5, r6, pc}
; CHECK-NEXT: .LBB12_6: @ %for.body.preheader21
; CHECK-NEXT: sub.w lr, r3, r5
; CHECK-NEXT: movw r0, #65535
@@ -2232,7 +2229,8 @@ define arm_aapcs_vfpcc void @ssatmul_4_q7(i8* nocapture readonly %pSrcA, i8* noc
; CHECK-NEXT: le lr, .LBB13_4
; CHECK-NEXT: @ %bb.5: @ %middle.block
; CHECK-NEXT: cmp r5, r3
-; CHECK-NEXT: beq .LBB13_8
+; CHECK-NEXT: it eq
+; CHECK-NEXT: popeq {r4, r5, r6, pc}
; CHECK-NEXT: .LBB13_6: @ %for.body.preheader21
; CHECK-NEXT: sub.w lr, r3, r5
; CHECK-NEXT: dls lr, lr
@@ -2366,7 +2364,8 @@ define arm_aapcs_vfpcc void @ssatmul_8_q7(i8* nocapture readonly %pSrcA, i8* noc
; CHECK-NEXT: le lr, .LBB14_4
; CHECK-NEXT: @ %bb.5: @ %middle.block
; CHECK-NEXT: cmp r5, r3
-; CHECK-NEXT: beq .LBB14_8
+; CHECK-NEXT: it eq
+; CHECK-NEXT: popeq {r4, r5, r6, pc}
; CHECK-NEXT: .LBB14_6: @ %for.body.preheader23
; CHECK-NEXT: sub.w lr, r3, r5
; CHECK-NEXT: mvn r0, #127
@@ -2505,7 +2504,8 @@ define arm_aapcs_vfpcc void @ssatmul_16_q7(i8* nocapture readonly %pSrcA, i8* no
; CHECK-NEXT: le lr, .LBB15_4
; CHECK-NEXT: @ %bb.5: @ %middle.block
; CHECK-NEXT: cmp r5, r3
-; CHECK-NEXT: beq .LBB15_8
+; CHECK-NEXT: it eq
+; CHECK-NEXT: popeq {r4, r5, r6, pc}
; CHECK-NEXT: .LBB15_6: @ %for.body.preheader23
; CHECK-NEXT: sub.w lr, r3, r5
; CHECK-NEXT: mvn r0, #127
@@ -2641,7 +2641,8 @@ define arm_aapcs_vfpcc void @ssatmul_16i_q7(i8* nocapture readonly %pSrcA, i8* n
; CHECK-NEXT: le lr, .LBB16_4
; CHECK-NEXT: @ %bb.5: @ %middle.block
; CHECK-NEXT: cmp r5, r3
-; CHECK-NEXT: beq .LBB16_8
+; CHECK-NEXT: it eq
+; CHECK-NEXT: popeq {r4, r5, r6, pc}
; CHECK-NEXT: .LBB16_6: @ %for.body.preheader23
; CHECK-NEXT: sub.w lr, r3, r5
; CHECK-NEXT: mvn r0, #127
@@ -3421,7 +3422,8 @@ define arm_aapcs_vfpcc void @usatmul_8_q7(i8* nocapture readonly %pSrcA, i8* noc
; CHECK-NEXT: le lr, .LBB20_4
; CHECK-NEXT: @ %bb.5: @ %middle.block
; CHECK-NEXT: cmp r5, r3
-; CHECK-NEXT: beq .LBB20_8
+; CHECK-NEXT: it eq
+; CHECK-NEXT: popeq {r4, r5, r6, pc}
; CHECK-NEXT: .LBB20_6: @ %for.body.preheader23
; CHECK-NEXT: sub.w lr, r3, r5
; CHECK-NEXT: dls lr, lr
@@ -3555,7 +3557,8 @@ define arm_aapcs_vfpcc void @usatmul_16_q7(i8* nocapture readonly %pSrcA, i8* no
; CHECK-NEXT: le lr, .LBB21_4
; CHECK-NEXT: @ %bb.5: @ %middle.block
; CHECK-NEXT: cmp r5, r3
-; CHECK-NEXT: beq .LBB21_8
+; CHECK-NEXT: it eq
+; CHECK-NEXT: popeq {r4, r5, r6, pc}
; CHECK-NEXT: .LBB21_6: @ %for.body.preheader23
; CHECK-NEXT: sub.w lr, r3, r5
; CHECK-NEXT: dls lr, lr
diff --git a/llvm/test/CodeGen/Thumb2/thumb2-branch.ll b/llvm/test/CodeGen/Thumb2/thumb2-branch.ll
index e52bab2b1105..332ed50ede6f 100644
--- a/llvm/test/CodeGen/Thumb2/thumb2-branch.ll
+++ b/llvm/test/CodeGen/Thumb2/thumb2-branch.ll
@@ -8,7 +8,7 @@ declare void @foo()
define i32 @f1(i32 %a, i32 %b, i32* %v) {
entry:
; CHECK-LABEL: f1:
-; CHECK: beq LBB
+; CHECK: bne LBB
%tmp = icmp eq i32 %a, %b ; <i1> [#uses=1]
br i1 %tmp, label %cond_true, label %return
@@ -59,7 +59,7 @@ return: ; preds = %entry
define i32 @f4(i32 %a, i32 %b, i32* %v) {
entry:
; CHECK-LABEL: f4:
-; CHECK: bhs LBB
+; CHECK: blo LBB
%tmp = icmp uge i32 %a, %b ; <i1> [#uses=1]
br i1 %tmp, label %cond_true, label %return
diff --git a/llvm/test/CodeGen/X86/3addr-16bit.ll b/llvm/test/CodeGen/X86/3addr-16bit.ll
index e8171e2767df..882150a6e39a 100644
--- a/llvm/test/CodeGen/X86/3addr-16bit.ll
+++ b/llvm/test/CodeGen/X86/3addr-16bit.ll
@@ -12,11 +12,8 @@ define zeroext i16 @test1(i16 zeroext %c, i16 zeroext %k) nounwind ssp {
; X64-NEXT: movl %esi, %eax
; X64-NEXT: incl %eax
; X64-NEXT: cmpw %di, %si
-; X64-NEXT: je LBB0_1
-; X64-NEXT: ## %bb.2: ## %bb1
-; X64-NEXT: movzwl %ax, %eax
-; X64-NEXT: retq
-; X64-NEXT: LBB0_1: ## %bb
+; X64-NEXT: jne LBB0_2
+; X64-NEXT: ## %bb.1: ## %bb
; X64-NEXT: pushq %rbx
; X64-NEXT: movzwl %ax, %ebx
; X64-NEXT: movl %ebx, %edi
@@ -24,6 +21,9 @@ define zeroext i16 @test1(i16 zeroext %c, i16 zeroext %k) nounwind ssp {
; X64-NEXT: movl %ebx, %eax
; X64-NEXT: popq %rbx
; X64-NEXT: retq
+; X64-NEXT: LBB0_2: ## %bb1
+; X64-NEXT: movzwl %ax, %eax
+; X64-NEXT: retq
;
; X32-LABEL: test1:
; X32: ## %bb.0: ## %entry
@@ -33,15 +33,15 @@ define zeroext i16 @test1(i16 zeroext %c, i16 zeroext %k) nounwind ssp {
; X32-NEXT: movl %ecx, %eax
; X32-NEXT: incl %eax
; X32-NEXT: cmpw {{[0-9]+}}(%esp), %cx
-; X32-NEXT: je LBB0_1
-; X32-NEXT: ## %bb.2: ## %bb1
-; X32-NEXT: movzwl %ax, %eax
-; X32-NEXT: jmp LBB0_3
-; X32-NEXT: LBB0_1: ## %bb
+; X32-NEXT: jne LBB0_2
+; X32-NEXT: ## %bb.1: ## %bb
; X32-NEXT: movzwl %ax, %esi
; X32-NEXT: movl %esi, (%esp)
; X32-NEXT: calll _foo
; X32-NEXT: movl %esi, %eax
+; X32-NEXT: jmp LBB0_3
+; X32-NEXT: LBB0_2: ## %bb1
+; X32-NEXT: movzwl %ax, %eax
; X32-NEXT: LBB0_3: ## %bb1
; X32-NEXT: addl $8, %esp
; X32-NEXT: popl %esi
@@ -65,11 +65,8 @@ define zeroext i16 @test2(i16 zeroext %c, i16 zeroext %k) nounwind ssp {
; X64-NEXT: movl %esi, %eax
; X64-NEXT: decl %eax
; X64-NEXT: cmpw %di, %si
-; X64-NEXT: je LBB1_1
-; X64-NEXT: ## %bb.2: ## %bb1
-; X64-NEXT: movzwl %ax, %eax
-; X64-NEXT: retq
-; X64-NEXT: LBB1_1: ## %bb
+; X64-NEXT: jne LBB1_2
+; X64-NEXT: ## %bb.1: ## %bb
; X64-NEXT: pushq %rbx
; X64-NEXT: movzwl %ax, %ebx
; X64-NEXT: movl %ebx, %edi
@@ -77,6 +74,9 @@ define zeroext i16 @test2(i16 zeroext %c, i16 zeroext %k) nounwind ssp {
; X64-NEXT: movl %ebx, %eax
; X64-NEXT: popq %rbx
; X64-NEXT: retq
+; X64-NEXT: LBB1_2: ## %bb1
+; X64-NEXT: movzwl %ax, %eax
+; X64-NEXT: retq
;
; X32-LABEL: test2:
; X32: ## %bb.0: ## %entry
@@ -86,15 +86,15 @@ define zeroext i16 @test2(i16 zeroext %c, i16 zeroext %k) nounwind ssp {
; X32-NEXT: movl %ecx, %eax
; X32-NEXT: decl %eax
; X32-NEXT: cmpw {{[0-9]+}}(%esp), %cx
-; X32-NEXT: je LBB1_1
-; X32-NEXT: ## %bb.2: ## %bb1
-; X32-NEXT: movzwl %ax, %eax
-; X32-NEXT: jmp LBB1_3
-; X32-NEXT: LBB1_1: ## %bb
+; X32-NEXT: jne LBB1_2
+; X32-NEXT: ## %bb.1: ## %bb
; X32-NEXT: movzwl %ax, %esi
; X32-NEXT: movl %esi, (%esp)
; X32-NEXT: calll _foo
; X32-NEXT: movl %esi, %eax
+; X32-NEXT: jmp LBB1_3
+; X32-NEXT: LBB1_2: ## %bb1
+; X32-NEXT: movzwl %ax, %eax
; X32-NEXT: LBB1_3: ## %bb1
; X32-NEXT: addl $8, %esp
; X32-NEXT: popl %esi
@@ -120,11 +120,8 @@ define zeroext i16 @test3(i16 zeroext %c, i16 zeroext %k) nounwind ssp {
; X64-NEXT: movl %esi, %eax
; X64-NEXT: addl $2, %eax
; X64-NEXT: cmpw %di, %si
-; X64-NEXT: je LBB2_1
-; X64-NEXT: ## %bb.2: ## %bb1
-; X64-NEXT: movzwl %ax, %eax
-; X64-NEXT: retq
-; X64-NEXT: LBB2_1: ## %bb
+; X64-NEXT: jne LBB2_2
+; X64-NEXT: ## %bb.1: ## %bb
; X64-NEXT: pushq %rbx
; X64-NEXT: movzwl %ax, %ebx
; X64-NEXT: movl %ebx, %edi
@@ -132,6 +129,9 @@ define zeroext i16 @test3(i16 zeroext %c, i16 zeroext %k) nounwind ssp {
; X64-NEXT: movl %ebx, %eax
; X64-NEXT: popq %rbx
; X64-NEXT: retq
+; X64-NEXT: LBB2_2: ## %bb1
+; X64-NEXT: movzwl %ax, %eax
+; X64-NEXT: retq
;
; X32-LABEL: test3:
; X32: ## %bb.0: ## %entry
@@ -141,15 +141,15 @@ define zeroext i16 @test3(i16 zeroext %c, i16 zeroext %k) nounwind ssp {
; X32-NEXT: movl %ecx, %eax
; X32-NEXT: addl $2, %eax
; X32-NEXT: cmpw {{[0-9]+}}(%esp), %cx
-; X32-NEXT: je LBB2_1
-; X32-NEXT: ## %bb.2: ## %bb1
-; X32-NEXT: movzwl %ax, %eax
-; X32-NEXT: jmp LBB2_3
-; X32-NEXT: LBB2_1: ## %bb
+; X32-NEXT: jne LBB2_2
+; X32-NEXT: ## %bb.1: ## %bb
; X32-NEXT: movzwl %ax, %esi
; X32-NEXT: movl %esi, (%esp)
; X32-NEXT: calll _foo
; X32-NEXT: movl %esi, %eax
+; X32-NEXT: jmp LBB2_3
+; X32-NEXT: LBB2_2: ## %bb1
+; X32-NEXT: movzwl %ax, %eax
; X32-NEXT: LBB2_3: ## %bb1
; X32-NEXT: addl $8, %esp
; X32-NEXT: popl %esi
@@ -173,11 +173,8 @@ define zeroext i16 @test4(i16 zeroext %c, i16 zeroext %k) nounwind ssp {
; X64-NEXT: movl %esi, %eax
; X64-NEXT: addl %edi, %eax
; X64-NEXT: cmpw %di, %si
-; X64-NEXT: je LBB3_1
-; X64-NEXT: ## %bb.2: ## %bb1
-; X64-NEXT: movzwl %ax, %eax
-; X64-NEXT: retq
-; X64-NEXT: LBB3_1: ## %bb
+; X64-NEXT: jne LBB3_2
+; X64-NEXT: ## %bb.1: ## %bb
; X64-NEXT: pushq %rbx
; X64-NEXT: movzwl %ax, %ebx
; X64-NEXT: movl %ebx, %edi
@@ -185,6 +182,9 @@ define zeroext i16 @test4(i16 zeroext %c, i16 zeroext %k) nounwind ssp {
; X64-NEXT: movl %ebx, %eax
; X64-NEXT: popq %rbx
; X64-NEXT: retq
+; X64-NEXT: LBB3_2: ## %bb1
+; X64-NEXT: movzwl %ax, %eax
+; X64-NEXT: retq
;
; X32-LABEL: test4:
; X32: ## %bb.0: ## %entry
@@ -195,15 +195,15 @@ define zeroext i16 @test4(i16 zeroext %c, i16 zeroext %k) nounwind ssp {
; X32-NEXT: movl %edx, %eax
; X32-NEXT: addl %ecx, %eax
; X32-NEXT: cmpw %cx, %dx
-; X32-NEXT: je LBB3_1
-; X32-NEXT: ## %bb.2: ## %bb1
-; X32-NEXT: movzwl %ax, %eax
-; X32-NEXT: jmp LBB3_3
-; X32-NEXT: LBB3_1: ## %bb
+; X32-NEXT: jne LBB3_2
+; X32-NEXT: ## %bb.1: ## %bb
; X32-NEXT: movzwl %ax, %esi
; X32-NEXT: movl %esi, (%esp)
; X32-NEXT: calll _foo
; X32-NEXT: movl %esi, %eax
+; X32-NEXT: jmp LBB3_3
+; X32-NEXT: LBB3_2: ## %bb1
+; X32-NEXT: movzwl %ax, %eax
; X32-NEXT: LBB3_3: ## %bb1
; X32-NEXT: addl $8, %esp
; X32-NEXT: popl %esi
diff --git a/llvm/test/CodeGen/X86/absolute-cmp.ll b/llvm/test/CodeGen/X86/absolute-cmp.ll
index b5fd426039f2..b4f158aa7c91 100644
--- a/llvm/test/CodeGen/X86/absolute-cmp.ll
+++ b/llvm/test/CodeGen/X86/absolute-cmp.ll
@@ -17,16 +17,12 @@ define void @foo8(i64 %val) {
; NOPIC-NEXT: .cfi_def_cfa_offset 16
; NOPIC-NEXT: cmpq $cmp8 at ABS8, %rdi # encoding: [0x48,0x83,0xff,A]
; NOPIC-NEXT: # fixup A - offset: 3, value: cmp8 at ABS8, kind: FK_Data_1
-; NOPIC-NEXT: jbe .LBB0_1 # encoding: [0x76,A]
-; NOPIC-NEXT: # fixup A - offset: 1, value: .LBB0_1-1, kind: FK_PCRel_1
-; NOPIC-NEXT: # %bb.2: # %f
-; NOPIC-NEXT: popq %rax # encoding: [0x58]
-; NOPIC-NEXT: .cfi_def_cfa_offset 8
-; NOPIC-NEXT: retq # encoding: [0xc3]
-; NOPIC-NEXT: .LBB0_1: # %t
-; NOPIC-NEXT: .cfi_def_cfa_offset 16
+; NOPIC-NEXT: ja .LBB0_2 # encoding: [0x77,A]
+; NOPIC-NEXT: # fixup A - offset: 1, value: .LBB0_2-1, kind: FK_PCRel_1
+; NOPIC-NEXT: # %bb.1: # %t
; NOPIC-NEXT: callq f # encoding: [0xe8,A,A,A,A]
; NOPIC-NEXT: # fixup A - offset: 1, value: f-4, kind: reloc_branch_4byte_pcrel
+; NOPIC-NEXT: .LBB0_2: # %f
; NOPIC-NEXT: popq %rax # encoding: [0x58]
; NOPIC-NEXT: .cfi_def_cfa_offset 8
; NOPIC-NEXT: retq # encoding: [0xc3]
@@ -37,16 +33,12 @@ define void @foo8(i64 %val) {
; PIC-NEXT: .cfi_def_cfa_offset 16
; PIC-NEXT: cmpq $cmp8 at ABS8, %rdi # encoding: [0x48,0x83,0xff,A]
; PIC-NEXT: # fixup A - offset: 3, value: cmp8 at ABS8, kind: FK_Data_1
-; PIC-NEXT: jbe .LBB0_1 # encoding: [0x76,A]
-; PIC-NEXT: # fixup A - offset: 1, value: .LBB0_1-1, kind: FK_PCRel_1
-; PIC-NEXT: # %bb.2: # %f
-; PIC-NEXT: popq %rax # encoding: [0x58]
-; PIC-NEXT: .cfi_def_cfa_offset 8
-; PIC-NEXT: retq # encoding: [0xc3]
-; PIC-NEXT: .LBB0_1: # %t
-; PIC-NEXT: .cfi_def_cfa_offset 16
+; PIC-NEXT: ja .LBB0_2 # encoding: [0x77,A]
+; PIC-NEXT: # fixup A - offset: 1, value: .LBB0_2-1, kind: FK_PCRel_1
+; PIC-NEXT: # %bb.1: # %t
; PIC-NEXT: callq f at PLT # encoding: [0xe8,A,A,A,A]
; PIC-NEXT: # fixup A - offset: 1, value: f at PLT-4, kind: FK_PCRel_4
+; PIC-NEXT: .LBB0_2: # %f
; PIC-NEXT: popq %rax # encoding: [0x58]
; PIC-NEXT: .cfi_def_cfa_offset 8
; PIC-NEXT: retq # encoding: [0xc3]
@@ -68,16 +60,12 @@ define void @foo32(i64 %val) {
; NOPIC-NEXT: .cfi_def_cfa_offset 16
; NOPIC-NEXT: cmpq $cmp32, %rdi # encoding: [0x48,0x81,0xff,A,A,A,A]
; NOPIC-NEXT: # fixup A - offset: 3, value: cmp32, kind: reloc_signed_4byte
-; NOPIC-NEXT: jbe .LBB1_1 # encoding: [0x76,A]
-; NOPIC-NEXT: # fixup A - offset: 1, value: .LBB1_1-1, kind: FK_PCRel_1
-; NOPIC-NEXT: # %bb.2: # %f
-; NOPIC-NEXT: popq %rax # encoding: [0x58]
-; NOPIC-NEXT: .cfi_def_cfa_offset 8
-; NOPIC-NEXT: retq # encoding: [0xc3]
-; NOPIC-NEXT: .LBB1_1: # %t
-; NOPIC-NEXT: .cfi_def_cfa_offset 16
+; NOPIC-NEXT: ja .LBB1_2 # encoding: [0x77,A]
+; NOPIC-NEXT: # fixup A - offset: 1, value: .LBB1_2-1, kind: FK_PCRel_1
+; NOPIC-NEXT: # %bb.1: # %t
; NOPIC-NEXT: callq f # encoding: [0xe8,A,A,A,A]
; NOPIC-NEXT: # fixup A - offset: 1, value: f-4, kind: reloc_branch_4byte_pcrel
+; NOPIC-NEXT: .LBB1_2: # %f
; NOPIC-NEXT: popq %rax # encoding: [0x58]
; NOPIC-NEXT: .cfi_def_cfa_offset 8
; NOPIC-NEXT: retq # encoding: [0xc3]
@@ -88,16 +76,12 @@ define void @foo32(i64 %val) {
; PIC-NEXT: .cfi_def_cfa_offset 16
; PIC-NEXT: cmpq $cmp32, %rdi # encoding: [0x48,0x81,0xff,A,A,A,A]
; PIC-NEXT: # fixup A - offset: 3, value: cmp32, kind: reloc_signed_4byte
-; PIC-NEXT: jbe .LBB1_1 # encoding: [0x76,A]
-; PIC-NEXT: # fixup A - offset: 1, value: .LBB1_1-1, kind: FK_PCRel_1
-; PIC-NEXT: # %bb.2: # %f
-; PIC-NEXT: popq %rax # encoding: [0x58]
-; PIC-NEXT: .cfi_def_cfa_offset 8
-; PIC-NEXT: retq # encoding: [0xc3]
-; PIC-NEXT: .LBB1_1: # %t
-; PIC-NEXT: .cfi_def_cfa_offset 16
+; PIC-NEXT: ja .LBB1_2 # encoding: [0x77,A]
+; PIC-NEXT: # fixup A - offset: 1, value: .LBB1_2-1, kind: FK_PCRel_1
+; PIC-NEXT: # %bb.1: # %t
; PIC-NEXT: callq f at PLT # encoding: [0xe8,A,A,A,A]
; PIC-NEXT: # fixup A - offset: 1, value: f at PLT-4, kind: FK_PCRel_4
+; PIC-NEXT: .LBB1_2: # %f
; PIC-NEXT: popq %rax # encoding: [0x58]
; PIC-NEXT: .cfi_def_cfa_offset 8
; PIC-NEXT: retq # encoding: [0xc3]
diff --git a/llvm/test/CodeGen/X86/atomic-flags.ll b/llvm/test/CodeGen/X86/atomic-flags.ll
index 6565c107f7a9..e0c4a915965c 100644
--- a/llvm/test/CodeGen/X86/atomic-flags.ll
+++ b/llvm/test/CodeGen/X86/atomic-flags.ll
@@ -1,55 +1,20 @@
-; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
-; RUN: llc < %s -mtriple=x86_64-unknown-unknown -verify-machineinstrs | FileCheck %s --check-prefixes=CHECK,X86-64
-; RUN: llc < %s -mtriple=i686-unknown-unknown -verify-machineinstrs | FileCheck %s --check-prefixes=CHECK,X86-32
+; RUN: llc < %s -mtriple=x86_64-unknown-unknown -verify-machineinstrs | FileCheck %s
+; RUN: llc < %s -mtriple=i686-unknown-unknown -verify-machineinstrs | FileCheck %s
; Make sure that flags are properly preserved despite atomic optimizations.
define i32 @atomic_and_flags_1(i8* %p, i32 %a, i32 %b) {
+; CHECK-LABEL: atomic_and_flags_1:
+
; Generate flags value, and use it.
-; X86-64-LABEL: atomic_and_flags_1:
-; X86-64: # %bb.0:
-; X86-64-NEXT: cmpl %edx, %esi
-; X86-64-NEXT: je .LBB0_1
-; X86-64-NEXT: # %bb.3: # %L2
-; X86-64-NEXT: movl $2, %eax
-; X86-64-NEXT: retq
-; X86-64-NEXT: .LBB0_1: # %L1
-; X86-64-NEXT: incb (%rdi)
-; X86-64-NEXT: cmpl %edx, %esi
-; X86-64-NEXT: je .LBB0_4
-; X86-64-NEXT: # %bb.2: # %L4
-; X86-64-NEXT: movl $4, %eax
-; X86-64-NEXT: retq
-; X86-64-NEXT: .LBB0_4: # %L3
-; X86-64-NEXT: movl $3, %eax
-; X86-64-NEXT: retq
-;
-; X86-32-LABEL: atomic_and_flags_1:
-; X86-32: # %bb.0:
-; X86-32-NEXT: movl {{[0-9]+}}(%esp), %eax
-; X86-32-NEXT: movl {{[0-9]+}}(%esp), %ecx
-; X86-32-NEXT: cmpl %eax, %ecx
-; X86-32-NEXT: je .LBB0_1
-; X86-32-NEXT: # %bb.3: # %L2
-; X86-32-NEXT: movl $2, %eax
-; X86-32-NEXT: retl
-; X86-32-NEXT: .LBB0_1: # %L1
-; X86-32-NEXT: movl {{[0-9]+}}(%esp), %edx
-; X86-32-NEXT: incb (%edx)
-; X86-32-NEXT: cmpl %eax, %ecx
-; X86-32-NEXT: je .LBB0_4
-; X86-32-NEXT: # %bb.2: # %L4
-; X86-32-NEXT: movl $4, %eax
-; X86-32-NEXT: retl
-; X86-32-NEXT: .LBB0_4: # %L3
-; X86-32-NEXT: movl $3, %eax
-; X86-32-NEXT: retl
+ ; CHECK: cmpl
+ ; CHECK-NEXT: jne
%cmp = icmp eq i32 %a, %b
br i1 %cmp, label %L1, label %L2
L1:
; The following pattern will get folded.
-
+ ; CHECK: incb
%1 = load atomic i8, i8* %p seq_cst, align 1
%2 = add i8 %1, 1 ; This forces the INC instruction to be generated.
store atomic i8 %2, i8* %p release, align 1
@@ -58,7 +23,8 @@ L1:
; somehow. This test checks that cmpl gets emitted again, but any
; rematerialization would work (the optimizer used to clobber the flags with
; the add).
-
+ ; CHECK-NEXT: cmpl
+ ; CHECK-NEXT: jne
br i1 %cmp, label %L3, label %L4
L2:
@@ -73,51 +39,18 @@ L4:
; Same as above, but using 2 as immediate to avoid the INC instruction.
define i32 @atomic_and_flags_2(i8* %p, i32 %a, i32 %b) {
-; X86-64-LABEL: atomic_and_flags_2:
-; X86-64: # %bb.0:
-; X86-64-NEXT: cmpl %edx, %esi
-; X86-64-NEXT: je .LBB1_1
-; X86-64-NEXT: # %bb.3: # %L2
-; X86-64-NEXT: movl $2, %eax
-; X86-64-NEXT: retq
-; X86-64-NEXT: .LBB1_1: # %L1
-; X86-64-NEXT: addb $2, (%rdi)
-; X86-64-NEXT: cmpl %edx, %esi
-; X86-64-NEXT: je .LBB1_4
-; X86-64-NEXT: # %bb.2: # %L4
-; X86-64-NEXT: movl $4, %eax
-; X86-64-NEXT: retq
-; X86-64-NEXT: .LBB1_4: # %L3
-; X86-64-NEXT: movl $3, %eax
-; X86-64-NEXT: retq
-;
-; X86-32-LABEL: atomic_and_flags_2:
-; X86-32: # %bb.0:
-; X86-32-NEXT: movl {{[0-9]+}}(%esp), %eax
-; X86-32-NEXT: movl {{[0-9]+}}(%esp), %ecx
-; X86-32-NEXT: cmpl %eax, %ecx
-; X86-32-NEXT: je .LBB1_1
-; X86-32-NEXT: # %bb.3: # %L2
-; X86-32-NEXT: movl $2, %eax
-; X86-32-NEXT: retl
-; X86-32-NEXT: .LBB1_1: # %L1
-; X86-32-NEXT: movl {{[0-9]+}}(%esp), %edx
-; X86-32-NEXT: addb $2, (%edx)
-; X86-32-NEXT: cmpl %eax, %ecx
-; X86-32-NEXT: je .LBB1_4
-; X86-32-NEXT: # %bb.2: # %L4
-; X86-32-NEXT: movl $4, %eax
-; X86-32-NEXT: retl
-; X86-32-NEXT: .LBB1_4: # %L3
-; X86-32-NEXT: movl $3, %eax
-; X86-32-NEXT: retl
+; CHECK-LABEL: atomic_and_flags_2:
+ ; CHECK: cmpl
+ ; CHECK-NEXT: jne
%cmp = icmp eq i32 %a, %b
br i1 %cmp, label %L1, label %L2
L1:
+ ; CHECK: addb
%1 = load atomic i8, i8* %p seq_cst, align 1
%2 = add i8 %1, 2
store atomic i8 %2, i8* %p release, align 1
-
+ ; CHECK-NEXT: cmpl
+ ; CHECK-NEXT: jne
br i1 %cmp, label %L3, label %L4
L2:
ret i32 2
diff --git a/llvm/test/CodeGen/X86/atomic-unordered.ll b/llvm/test/CodeGen/X86/atomic-unordered.ll
index 3e5abc97b89a..7a1f34c65c18 100644
--- a/llvm/test/CodeGen/X86/atomic-unordered.ll
+++ b/llvm/test/CodeGen/X86/atomic-unordered.ll
@@ -2716,24 +2716,24 @@ define i1 @fold_cmp_over_fence(i32* %p, i32 %v1) {
; CHECK-O3-CUR-NEXT: movl (%rdi), %eax
; CHECK-O3-CUR-NEXT: mfence
; CHECK-O3-CUR-NEXT: cmpl %eax, %esi
-; CHECK-O3-CUR-NEXT: je .LBB116_1
-; CHECK-O3-CUR-NEXT: # %bb.2: # %untaken
-; CHECK-O3-CUR-NEXT: xorl %eax, %eax
-; CHECK-O3-CUR-NEXT: retq
-; CHECK-O3-CUR-NEXT: .LBB116_1: # %taken
+; CHECK-O3-CUR-NEXT: jne .LBB116_2
+; CHECK-O3-CUR-NEXT: # %bb.1: # %taken
; CHECK-O3-CUR-NEXT: movb $1, %al
; CHECK-O3-CUR-NEXT: retq
+; CHECK-O3-CUR-NEXT: .LBB116_2: # %untaken
+; CHECK-O3-CUR-NEXT: xorl %eax, %eax
+; CHECK-O3-CUR-NEXT: retq
;
; CHECK-O3-EX-LABEL: fold_cmp_over_fence:
; CHECK-O3-EX: # %bb.0:
; CHECK-O3-EX-NEXT: cmpl (%rdi), %esi
; CHECK-O3-EX-NEXT: mfence
-; CHECK-O3-EX-NEXT: je .LBB116_1
-; CHECK-O3-EX-NEXT: # %bb.2: # %untaken
-; CHECK-O3-EX-NEXT: xorl %eax, %eax
-; CHECK-O3-EX-NEXT: retq
-; CHECK-O3-EX-NEXT: .LBB116_1: # %taken
+; CHECK-O3-EX-NEXT: jne .LBB116_2
+; CHECK-O3-EX-NEXT: # %bb.1: # %taken
; CHECK-O3-EX-NEXT: movb $1, %al
+; CHECK-O3-EX-NEXT: retq
+; CHECK-O3-EX-NEXT: .LBB116_2: # %untaken
+; CHECK-O3-EX-NEXT: xorl %eax, %eax
; CHECK-O3-EX-NEXT: retq
%v2 = load atomic i32, i32* %p unordered, align 4
fence seq_cst
diff --git a/llvm/test/CodeGen/X86/bmi.ll b/llvm/test/CodeGen/X86/bmi.ll
index f8c7300186df..9c3a74fccbe6 100644
--- a/llvm/test/CodeGen/X86/bmi.ll
+++ b/llvm/test/CodeGen/X86/bmi.ll
@@ -1249,20 +1249,20 @@ define void @pr42118_i32(i32 %x) {
; X86-LABEL: pr42118_i32:
; X86: # %bb.0:
; X86-NEXT: blsrl {{[0-9]+}}(%esp), %eax
-; X86-NEXT: je .LBB50_2
-; X86-NEXT: # %bb.1:
-; X86-NEXT: retl
-; X86-NEXT: .LBB50_2:
+; X86-NEXT: jne .LBB50_1
+; X86-NEXT: # %bb.2:
; X86-NEXT: jmp bar # TAILCALL
+; X86-NEXT: .LBB50_1:
+; X86-NEXT: retl
;
; X64-LABEL: pr42118_i32:
; X64: # %bb.0:
; X64-NEXT: blsrl %edi, %eax
-; X64-NEXT: je .LBB50_2
-; X64-NEXT: # %bb.1:
-; X64-NEXT: retq
-; X64-NEXT: .LBB50_2:
+; X64-NEXT: jne .LBB50_1
+; X64-NEXT: # %bb.2:
; X64-NEXT: jmp bar # TAILCALL
+; X64-NEXT: .LBB50_1:
+; X64-NEXT: retq
%tmp = sub i32 0, %x
%tmp1 = and i32 %tmp, %x
%cmp = icmp eq i32 %tmp1, %x
@@ -1289,25 +1289,25 @@ define void @pr42118_i64(i64 %x) {
; X86-NEXT: andl %eax, %edx
; X86-NEXT: andl %ecx, %esi
; X86-NEXT: orl %edx, %esi
-; X86-NEXT: je .LBB51_2
-; X86-NEXT: # %bb.1:
+; X86-NEXT: jne .LBB51_1
+; X86-NEXT: # %bb.2:
; X86-NEXT: popl %esi
; X86-NEXT: .cfi_def_cfa_offset 4
-; X86-NEXT: retl
-; X86-NEXT: .LBB51_2:
+; X86-NEXT: jmp bar # TAILCALL
+; X86-NEXT: .LBB51_1:
; X86-NEXT: .cfi_def_cfa_offset 8
; X86-NEXT: popl %esi
; X86-NEXT: .cfi_def_cfa_offset 4
-; X86-NEXT: jmp bar # TAILCALL
+; X86-NEXT: retl
;
; X64-LABEL: pr42118_i64:
; X64: # %bb.0:
; X64-NEXT: blsrq %rdi, %rax
-; X64-NEXT: je .LBB51_2
-; X64-NEXT: # %bb.1:
-; X64-NEXT: retq
-; X64-NEXT: .LBB51_2:
+; X64-NEXT: jne .LBB51_1
+; X64-NEXT: # %bb.2:
; X64-NEXT: jmp bar # TAILCALL
+; X64-NEXT: .LBB51_1:
+; X64-NEXT: retq
%tmp = sub i64 0, %x
%tmp1 = and i64 %tmp, %x
%cmp = icmp eq i64 %tmp1, %x
diff --git a/llvm/test/CodeGen/X86/bt.ll b/llvm/test/CodeGen/X86/bt.ll
index 48fef1e8270c..59df366c6e4a 100644
--- a/llvm/test/CodeGen/X86/bt.ll
+++ b/llvm/test/CodeGen/X86/bt.ll
@@ -611,23 +611,21 @@ define void @query3(i32 %x, i32 %n) nounwind {
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
; X86-NEXT: btl %ecx, %eax
-; X86-NEXT: jb .LBB16_1
-; X86-NEXT: # %bb.2: # %UnifiedReturnBlock
-; X86-NEXT: retl
-; X86-NEXT: .LBB16_1: # %bb
+; X86-NEXT: jae .LBB16_2
+; X86-NEXT: # %bb.1: # %bb
; X86-NEXT: calll foo
+; X86-NEXT: .LBB16_2: # %UnifiedReturnBlock
; X86-NEXT: retl
;
; X64-LABEL: query3:
; X64: # %bb.0: # %entry
; X64-NEXT: btl %esi, %edi
-; X64-NEXT: jb .LBB16_1
-; X64-NEXT: # %bb.2: # %UnifiedReturnBlock
-; X64-NEXT: retq
-; X64-NEXT: .LBB16_1: # %bb
+; X64-NEXT: jae .LBB16_2
+; X64-NEXT: # %bb.1: # %bb
; X64-NEXT: pushq %rax
; X64-NEXT: callq foo
; X64-NEXT: popq %rax
+; X64-NEXT: .LBB16_2: # %UnifiedReturnBlock
; X64-NEXT: retq
entry:
%tmp29 = shl i32 1, %n
@@ -649,23 +647,21 @@ define void @query3b(i32 %x, i32 %n) nounwind {
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
; X86-NEXT: btl %ecx, %eax
-; X86-NEXT: jb .LBB17_1
-; X86-NEXT: # %bb.2: # %UnifiedReturnBlock
-; X86-NEXT: retl
-; X86-NEXT: .LBB17_1: # %bb
+; X86-NEXT: jae .LBB17_2
+; X86-NEXT: # %bb.1: # %bb
; X86-NEXT: calll foo
+; X86-NEXT: .LBB17_2: # %UnifiedReturnBlock
; X86-NEXT: retl
;
; X64-LABEL: query3b:
; X64: # %bb.0: # %entry
; X64-NEXT: btl %esi, %edi
-; X64-NEXT: jb .LBB17_1
-; X64-NEXT: # %bb.2: # %UnifiedReturnBlock
-; X64-NEXT: retq
-; X64-NEXT: .LBB17_1: # %bb
+; X64-NEXT: jae .LBB17_2
+; X64-NEXT: # %bb.1: # %bb
; X64-NEXT: pushq %rax
; X64-NEXT: callq foo
; X64-NEXT: popq %rax
+; X64-NEXT: .LBB17_2: # %UnifiedReturnBlock
; X64-NEXT: retq
entry:
%tmp29 = shl i32 1, %n
@@ -687,23 +683,21 @@ define void @query3x(i32 %x, i32 %n) nounwind {
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
; X86-NEXT: btl %ecx, %eax
-; X86-NEXT: jb .LBB18_1
-; X86-NEXT: # %bb.2: # %UnifiedReturnBlock
-; X86-NEXT: retl
-; X86-NEXT: .LBB18_1: # %bb
+; X86-NEXT: jae .LBB18_2
+; X86-NEXT: # %bb.1: # %bb
; X86-NEXT: calll foo
+; X86-NEXT: .LBB18_2: # %UnifiedReturnBlock
; X86-NEXT: retl
;
; X64-LABEL: query3x:
; X64: # %bb.0: # %entry
; X64-NEXT: btl %esi, %edi
-; X64-NEXT: jb .LBB18_1
-; X64-NEXT: # %bb.2: # %UnifiedReturnBlock
-; X64-NEXT: retq
-; X64-NEXT: .LBB18_1: # %bb
+; X64-NEXT: jae .LBB18_2
+; X64-NEXT: # %bb.1: # %bb
; X64-NEXT: pushq %rax
; X64-NEXT: callq foo
; X64-NEXT: popq %rax
+; X64-NEXT: .LBB18_2: # %UnifiedReturnBlock
; X64-NEXT: retq
entry:
%tmp29 = shl i32 1, %n
@@ -725,23 +719,21 @@ define void @query3bx(i32 %x, i32 %n) nounwind {
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
; X86-NEXT: btl %ecx, %eax
-; X86-NEXT: jb .LBB19_1
-; X86-NEXT: # %bb.2: # %UnifiedReturnBlock
-; X86-NEXT: retl
-; X86-NEXT: .LBB19_1: # %bb
+; X86-NEXT: jae .LBB19_2
+; X86-NEXT: # %bb.1: # %bb
; X86-NEXT: calll foo
+; X86-NEXT: .LBB19_2: # %UnifiedReturnBlock
; X86-NEXT: retl
;
; X64-LABEL: query3bx:
; X64: # %bb.0: # %entry
; X64-NEXT: btl %esi, %edi
-; X64-NEXT: jb .LBB19_1
-; X64-NEXT: # %bb.2: # %UnifiedReturnBlock
-; X64-NEXT: retq
-; X64-NEXT: .LBB19_1: # %bb
+; X64-NEXT: jae .LBB19_2
+; X64-NEXT: # %bb.1: # %bb
; X64-NEXT: pushq %rax
; X64-NEXT: callq foo
; X64-NEXT: popq %rax
+; X64-NEXT: .LBB19_2: # %UnifiedReturnBlock
; X64-NEXT: retq
entry:
%tmp29 = shl i32 1, %n
diff --git a/llvm/test/CodeGen/X86/conditional-tailcall-pgso.ll b/llvm/test/CodeGen/X86/conditional-tailcall-pgso.ll
index d3cb6daecf9c..65bd1dad21a8 100644
--- a/llvm/test/CodeGen/X86/conditional-tailcall-pgso.ll
+++ b/llvm/test/CodeGen/X86/conditional-tailcall-pgso.ll
@@ -63,21 +63,21 @@ define void @f_non_leaf(i32 %x, i32 %y) !prof !14 {
; CHECK32-NEXT: #APP
; CHECK32-NEXT: #NO_APP
; CHECK32-NEXT: cmpl {{[0-9]+}}(%esp), %eax # encoding: [0x3b,0x44,0x24,0x0c]
-; CHECK32-NEXT: je .LBB1_1 # encoding: [0x74,A]
-; CHECK32-NEXT: # fixup A - offset: 1, value: .LBB1_1-1, kind: FK_PCRel_1
-; CHECK32-NEXT: # %bb.2: # %bb2
+; CHECK32-NEXT: jne .LBB1_2 # encoding: [0x75,A]
+; CHECK32-NEXT: # fixup A - offset: 1, value: .LBB1_2-1, kind: FK_PCRel_1
+; CHECK32-NEXT: # %bb.1: # %bb1
; CHECK32-NEXT: popl %ebx # encoding: [0x5b]
; CHECK32-NEXT: .cfi_def_cfa_offset 4
-; CHECK32-NEXT: jmp bar # TAILCALL
+; CHECK32-NEXT: jmp foo # TAILCALL
; CHECK32-NEXT: # encoding: [0xeb,A]
-; CHECK32-NEXT: # fixup A - offset: 1, value: bar-1, kind: FK_PCRel_1
-; CHECK32-NEXT: .LBB1_1: # %bb1
+; CHECK32-NEXT: # fixup A - offset: 1, value: foo-1, kind: FK_PCRel_1
+; CHECK32-NEXT: .LBB1_2: # %bb2
; CHECK32-NEXT: .cfi_def_cfa_offset 8
; CHECK32-NEXT: popl %ebx # encoding: [0x5b]
; CHECK32-NEXT: .cfi_def_cfa_offset 4
-; CHECK32-NEXT: jmp foo # TAILCALL
+; CHECK32-NEXT: jmp bar # TAILCALL
; CHECK32-NEXT: # encoding: [0xeb,A]
-; CHECK32-NEXT: # fixup A - offset: 1, value: foo-1, kind: FK_PCRel_1
+; CHECK32-NEXT: # fixup A - offset: 1, value: bar-1, kind: FK_PCRel_1
;
; CHECK64-LABEL: f_non_leaf:
; CHECK64: # %bb.0: # %entry
@@ -87,21 +87,21 @@ define void @f_non_leaf(i32 %x, i32 %y) !prof !14 {
; CHECK64-NEXT: #APP
; CHECK64-NEXT: #NO_APP
; CHECK64-NEXT: cmpl %esi, %edi # encoding: [0x39,0xf7]
-; CHECK64-NEXT: je .LBB1_1 # encoding: [0x74,A]
-; CHECK64-NEXT: # fixup A - offset: 1, value: .LBB1_1-1, kind: FK_PCRel_1
-; CHECK64-NEXT: # %bb.2: # %bb2
+; CHECK64-NEXT: jne .LBB1_2 # encoding: [0x75,A]
+; CHECK64-NEXT: # fixup A - offset: 1, value: .LBB1_2-1, kind: FK_PCRel_1
+; CHECK64-NEXT: # %bb.1: # %bb1
; CHECK64-NEXT: popq %rbx # encoding: [0x5b]
; CHECK64-NEXT: .cfi_def_cfa_offset 8
-; CHECK64-NEXT: jmp bar # TAILCALL
+; CHECK64-NEXT: jmp foo # TAILCALL
; CHECK64-NEXT: # encoding: [0xeb,A]
-; CHECK64-NEXT: # fixup A - offset: 1, value: bar-1, kind: FK_PCRel_1
-; CHECK64-NEXT: .LBB1_1: # %bb1
+; CHECK64-NEXT: # fixup A - offset: 1, value: foo-1, kind: FK_PCRel_1
+; CHECK64-NEXT: .LBB1_2: # %bb2
; CHECK64-NEXT: .cfi_def_cfa_offset 16
; CHECK64-NEXT: popq %rbx # encoding: [0x5b]
; CHECK64-NEXT: .cfi_def_cfa_offset 8
-; CHECK64-NEXT: jmp foo # TAILCALL
+; CHECK64-NEXT: jmp bar # TAILCALL
; CHECK64-NEXT: # encoding: [0xeb,A]
-; CHECK64-NEXT: # fixup A - offset: 1, value: foo-1, kind: FK_PCRel_1
+; CHECK64-NEXT: # fixup A - offset: 1, value: bar-1, kind: FK_PCRel_1
;
; WIN64-LABEL: f_non_leaf:
; WIN64: # %bb.0: # %entry
@@ -111,19 +111,19 @@ define void @f_non_leaf(i32 %x, i32 %y) !prof !14 {
; WIN64-NEXT: #APP
; WIN64-NEXT: #NO_APP
; WIN64-NEXT: cmpl %edx, %ecx # encoding: [0x39,0xd1]
-; WIN64-NEXT: je .LBB1_1 # encoding: [0x74,A]
-; WIN64-NEXT: # fixup A - offset: 1, value: .LBB1_1-1, kind: FK_PCRel_1
-; WIN64-NEXT: # %bb.2: # %bb2
+; WIN64-NEXT: jne .LBB1_2 # encoding: [0x75,A]
+; WIN64-NEXT: # fixup A - offset: 1, value: .LBB1_2-1, kind: FK_PCRel_1
+; WIN64-NEXT: # %bb.1: # %bb1
; WIN64-NEXT: popq %rbx # encoding: [0x5b]
-; WIN64-NEXT: jmp bar # TAILCALL
+; WIN64-NEXT: jmp foo # TAILCALL
; WIN64-NEXT: # encoding: [0xeb,A]
-; WIN64-NEXT: # fixup A - offset: 1, value: bar-1, kind: FK_PCRel_1
-; WIN64-NEXT: .LBB1_1: # %bb1
+; WIN64-NEXT: # fixup A - offset: 1, value: foo-1, kind: FK_PCRel_1
+; WIN64-NEXT: .LBB1_2: # %bb2
; WIN64-NEXT: nop # encoding: [0x90]
; WIN64-NEXT: popq %rbx # encoding: [0x5b]
-; WIN64-NEXT: jmp foo # TAILCALL
+; WIN64-NEXT: jmp bar # TAILCALL
; WIN64-NEXT: # encoding: [0xeb,A]
-; WIN64-NEXT: # fixup A - offset: 1, value: foo-1, kind: FK_PCRel_1
+; WIN64-NEXT: # fixup A - offset: 1, value: bar-1, kind: FK_PCRel_1
; WIN64-NEXT: .seh_handlerdata
; WIN64-NEXT: .text
; WIN64-NEXT: .seh_endproc
diff --git a/llvm/test/CodeGen/X86/conditional-tailcall.ll b/llvm/test/CodeGen/X86/conditional-tailcall.ll
index 66c57697aa72..17078413a824 100644
--- a/llvm/test/CodeGen/X86/conditional-tailcall.ll
+++ b/llvm/test/CodeGen/X86/conditional-tailcall.ll
@@ -63,21 +63,21 @@ define void @f_non_leaf(i32 %x, i32 %y) optsize {
; CHECK32-NEXT: #APP
; CHECK32-NEXT: #NO_APP
; CHECK32-NEXT: cmpl {{[0-9]+}}(%esp), %eax # encoding: [0x3b,0x44,0x24,0x0c]
-; CHECK32-NEXT: je .LBB1_1 # encoding: [0x74,A]
-; CHECK32-NEXT: # fixup A - offset: 1, value: .LBB1_1-1, kind: FK_PCRel_1
-; CHECK32-NEXT: # %bb.2: # %bb2
+; CHECK32-NEXT: jne .LBB1_2 # encoding: [0x75,A]
+; CHECK32-NEXT: # fixup A - offset: 1, value: .LBB1_2-1, kind: FK_PCRel_1
+; CHECK32-NEXT: # %bb.1: # %bb1
; CHECK32-NEXT: popl %ebx # encoding: [0x5b]
; CHECK32-NEXT: .cfi_def_cfa_offset 4
-; CHECK32-NEXT: jmp bar # TAILCALL
+; CHECK32-NEXT: jmp foo # TAILCALL
; CHECK32-NEXT: # encoding: [0xeb,A]
-; CHECK32-NEXT: # fixup A - offset: 1, value: bar-1, kind: FK_PCRel_1
-; CHECK32-NEXT: .LBB1_1: # %bb1
+; CHECK32-NEXT: # fixup A - offset: 1, value: foo-1, kind: FK_PCRel_1
+; CHECK32-NEXT: .LBB1_2: # %bb2
; CHECK32-NEXT: .cfi_def_cfa_offset 8
; CHECK32-NEXT: popl %ebx # encoding: [0x5b]
; CHECK32-NEXT: .cfi_def_cfa_offset 4
-; CHECK32-NEXT: jmp foo # TAILCALL
+; CHECK32-NEXT: jmp bar # TAILCALL
; CHECK32-NEXT: # encoding: [0xeb,A]
-; CHECK32-NEXT: # fixup A - offset: 1, value: foo-1, kind: FK_PCRel_1
+; CHECK32-NEXT: # fixup A - offset: 1, value: bar-1, kind: FK_PCRel_1
;
; CHECK64-LABEL: f_non_leaf:
; CHECK64: # %bb.0: # %entry
@@ -87,21 +87,21 @@ define void @f_non_leaf(i32 %x, i32 %y) optsize {
; CHECK64-NEXT: #APP
; CHECK64-NEXT: #NO_APP
; CHECK64-NEXT: cmpl %esi, %edi # encoding: [0x39,0xf7]
-; CHECK64-NEXT: je .LBB1_1 # encoding: [0x74,A]
-; CHECK64-NEXT: # fixup A - offset: 1, value: .LBB1_1-1, kind: FK_PCRel_1
-; CHECK64-NEXT: # %bb.2: # %bb2
+; CHECK64-NEXT: jne .LBB1_2 # encoding: [0x75,A]
+; CHECK64-NEXT: # fixup A - offset: 1, value: .LBB1_2-1, kind: FK_PCRel_1
+; CHECK64-NEXT: # %bb.1: # %bb1
; CHECK64-NEXT: popq %rbx # encoding: [0x5b]
; CHECK64-NEXT: .cfi_def_cfa_offset 8
-; CHECK64-NEXT: jmp bar # TAILCALL
+; CHECK64-NEXT: jmp foo # TAILCALL
; CHECK64-NEXT: # encoding: [0xeb,A]
-; CHECK64-NEXT: # fixup A - offset: 1, value: bar-1, kind: FK_PCRel_1
-; CHECK64-NEXT: .LBB1_1: # %bb1
+; CHECK64-NEXT: # fixup A - offset: 1, value: foo-1, kind: FK_PCRel_1
+; CHECK64-NEXT: .LBB1_2: # %bb2
; CHECK64-NEXT: .cfi_def_cfa_offset 16
; CHECK64-NEXT: popq %rbx # encoding: [0x5b]
; CHECK64-NEXT: .cfi_def_cfa_offset 8
-; CHECK64-NEXT: jmp foo # TAILCALL
+; CHECK64-NEXT: jmp bar # TAILCALL
; CHECK64-NEXT: # encoding: [0xeb,A]
-; CHECK64-NEXT: # fixup A - offset: 1, value: foo-1, kind: FK_PCRel_1
+; CHECK64-NEXT: # fixup A - offset: 1, value: bar-1, kind: FK_PCRel_1
;
; WIN64-LABEL: f_non_leaf:
; WIN64: # %bb.0: # %entry
@@ -111,19 +111,19 @@ define void @f_non_leaf(i32 %x, i32 %y) optsize {
; WIN64-NEXT: #APP
; WIN64-NEXT: #NO_APP
; WIN64-NEXT: cmpl %edx, %ecx # encoding: [0x39,0xd1]
-; WIN64-NEXT: je .LBB1_1 # encoding: [0x74,A]
-; WIN64-NEXT: # fixup A - offset: 1, value: .LBB1_1-1, kind: FK_PCRel_1
-; WIN64-NEXT: # %bb.2: # %bb2
+; WIN64-NEXT: jne .LBB1_2 # encoding: [0x75,A]
+; WIN64-NEXT: # fixup A - offset: 1, value: .LBB1_2-1, kind: FK_PCRel_1
+; WIN64-NEXT: # %bb.1: # %bb1
; WIN64-NEXT: popq %rbx # encoding: [0x5b]
-; WIN64-NEXT: jmp bar # TAILCALL
+; WIN64-NEXT: jmp foo # TAILCALL
; WIN64-NEXT: # encoding: [0xeb,A]
-; WIN64-NEXT: # fixup A - offset: 1, value: bar-1, kind: FK_PCRel_1
-; WIN64-NEXT: .LBB1_1: # %bb1
+; WIN64-NEXT: # fixup A - offset: 1, value: foo-1, kind: FK_PCRel_1
+; WIN64-NEXT: .LBB1_2: # %bb2
; WIN64-NEXT: nop # encoding: [0x90]
; WIN64-NEXT: popq %rbx # encoding: [0x5b]
-; WIN64-NEXT: jmp foo # TAILCALL
+; WIN64-NEXT: jmp bar # TAILCALL
; WIN64-NEXT: # encoding: [0xeb,A]
-; WIN64-NEXT: # fixup A - offset: 1, value: foo-1, kind: FK_PCRel_1
+; WIN64-NEXT: # fixup A - offset: 1, value: bar-1, kind: FK_PCRel_1
; WIN64-NEXT: .seh_handlerdata
; WIN64-NEXT: .text
; WIN64-NEXT: .seh_endproc
diff --git a/llvm/test/CodeGen/X86/fast-isel-cmp-branch2.ll b/llvm/test/CodeGen/X86/fast-isel-cmp-branch2.ll
index 1ba8b00646ef..475d8fcf7f35 100644
--- a/llvm/test/CodeGen/X86/fast-isel-cmp-branch2.ll
+++ b/llvm/test/CodeGen/X86/fast-isel-cmp-branch2.ll
@@ -1,19 +1,11 @@
-; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc < %s -mtriple=x86_64-apple-darwin10 | FileCheck %s
; RUN: llc < %s -fast-isel -fast-isel-abort=1 -mtriple=x86_64-apple-darwin10 | FileCheck %s
define i32 @fcmp_oeq(float %x, float %y) {
-; CHECK-LABEL: fcmp_oeq:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: ucomiss %xmm1, %xmm0
-; CHECK-NEXT: jne LBB0_1
-; CHECK-NEXT: jp LBB0_1
-; CHECK-NEXT: ## %bb.2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB0_1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: fcmp_oeq
+; CHECK: ucomiss %xmm1, %xmm0
+; CHECK-NEXT: jne {{LBB.+_1}}
+; CHECK-NEXT: jp {{LBB.+_1}}
%1 = fcmp oeq float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -23,16 +15,9 @@ bb1:
}
define i32 @fcmp_ogt(float %x, float %y) {
-; CHECK-LABEL: fcmp_ogt:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: ucomiss %xmm1, %xmm0
-; CHECK-NEXT: jbe LBB1_1
-; CHECK-NEXT: ## %bb.2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB1_1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: fcmp_ogt
+; CHECK: ucomiss %xmm1, %xmm0
+; CHECK-NEXT: jbe {{LBB.+_1}}
%1 = fcmp ogt float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -42,16 +27,9 @@ bb1:
}
define i32 @fcmp_oge(float %x, float %y) {
-; CHECK-LABEL: fcmp_oge:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: ucomiss %xmm1, %xmm0
-; CHECK-NEXT: jb LBB2_1
-; CHECK-NEXT: ## %bb.2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB2_1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: fcmp_oge
+; CHECK: ucomiss %xmm1, %xmm0
+; CHECK-NEXT: jb {{LBB.+_1}}
%1 = fcmp oge float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -61,16 +39,9 @@ bb1:
}
define i32 @fcmp_olt(float %x, float %y) {
-; CHECK-LABEL: fcmp_olt:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: ucomiss %xmm0, %xmm1
-; CHECK-NEXT: jbe LBB3_1
-; CHECK-NEXT: ## %bb.2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB3_1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: fcmp_olt
+; CHECK: ucomiss %xmm0, %xmm1
+; CHECK-NEXT: jbe {{LBB.+_1}}
%1 = fcmp olt float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -80,16 +51,9 @@ bb1:
}
define i32 @fcmp_ole(float %x, float %y) {
-; CHECK-LABEL: fcmp_ole:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: ucomiss %xmm0, %xmm1
-; CHECK-NEXT: jb LBB4_1
-; CHECK-NEXT: ## %bb.2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB4_1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: fcmp_ole
+; CHECK: ucomiss %xmm0, %xmm1
+; CHECK-NEXT: jb {{LBB.+_1}}
%1 = fcmp ole float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -99,16 +63,9 @@ bb1:
}
define i32 @fcmp_one(float %x, float %y) {
-; CHECK-LABEL: fcmp_one:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: ucomiss %xmm1, %xmm0
-; CHECK-NEXT: je LBB5_1
-; CHECK-NEXT: ## %bb.2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB5_1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: fcmp_one
+; CHECK: ucomiss %xmm1, %xmm0
+; CHECK-NEXT: je {{LBB.+_1}}
%1 = fcmp one float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -118,16 +75,9 @@ bb1:
}
define i32 @fcmp_ord(float %x, float %y) {
-; CHECK-LABEL: fcmp_ord:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: ucomiss %xmm1, %xmm0
-; CHECK-NEXT: jp LBB6_1
-; CHECK-NEXT: ## %bb.2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB6_1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: fcmp_ord
+; CHECK: ucomiss %xmm1, %xmm0
+; CHECK-NEXT: jp {{LBB.+_1}}
%1 = fcmp ord float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -137,16 +87,9 @@ bb1:
}
define i32 @fcmp_uno(float %x, float %y) {
-; CHECK-LABEL: fcmp_uno:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: ucomiss %xmm1, %xmm0
-; CHECK-NEXT: jp LBB7_2
-; CHECK-NEXT: ## %bb.1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB7_2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: fcmp_uno
+; CHECK: ucomiss %xmm1, %xmm0
+; CHECK-NEXT: jp {{LBB.+_2}}
%1 = fcmp uno float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -156,16 +99,9 @@ bb1:
}
define i32 @fcmp_ueq(float %x, float %y) {
-; CHECK-LABEL: fcmp_ueq:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: ucomiss %xmm1, %xmm0
-; CHECK-NEXT: je LBB8_2
-; CHECK-NEXT: ## %bb.1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB8_2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: fcmp_ueq
+; CHECK: ucomiss %xmm1, %xmm0
+; CHECK-NEXT: je {{LBB.+_2}}
%1 = fcmp ueq float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -175,16 +111,9 @@ bb1:
}
define i32 @fcmp_ugt(float %x, float %y) {
-; CHECK-LABEL: fcmp_ugt:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: ucomiss %xmm0, %xmm1
-; CHECK-NEXT: jae LBB9_1
-; CHECK-NEXT: ## %bb.2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB9_1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: fcmp_ugt
+; CHECK: ucomiss %xmm0, %xmm1
+; CHECK-NEXT: jae {{LBB.+_1}}
%1 = fcmp ugt float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -194,16 +123,9 @@ bb1:
}
define i32 @fcmp_uge(float %x, float %y) {
-; CHECK-LABEL: fcmp_uge:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: ucomiss %xmm0, %xmm1
-; CHECK-NEXT: ja LBB10_1
-; CHECK-NEXT: ## %bb.2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB10_1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: fcmp_uge
+; CHECK: ucomiss %xmm0, %xmm1
+; CHECK-NEXT: ja {{LBB.+_1}}
%1 = fcmp uge float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -213,16 +135,9 @@ bb1:
}
define i32 @fcmp_ult(float %x, float %y) {
-; CHECK-LABEL: fcmp_ult:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: ucomiss %xmm1, %xmm0
-; CHECK-NEXT: jae LBB11_1
-; CHECK-NEXT: ## %bb.2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB11_1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: fcmp_ult
+; CHECK: ucomiss %xmm1, %xmm0
+; CHECK-NEXT: jae {{LBB.+_1}}
%1 = fcmp ult float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -232,16 +147,9 @@ bb1:
}
define i32 @fcmp_ule(float %x, float %y) {
-; CHECK-LABEL: fcmp_ule:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: ucomiss %xmm1, %xmm0
-; CHECK-NEXT: ja LBB12_1
-; CHECK-NEXT: ## %bb.2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB12_1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: fcmp_ule
+; CHECK: ucomiss %xmm1, %xmm0
+; CHECK-NEXT: ja {{LBB.+_1}}
%1 = fcmp ule float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -251,17 +159,10 @@ bb1:
}
define i32 @fcmp_une(float %x, float %y) {
-; CHECK-LABEL: fcmp_une:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: ucomiss %xmm1, %xmm0
-; CHECK-NEXT: jne LBB13_2
-; CHECK-NEXT: jnp LBB13_1
-; CHECK-NEXT: LBB13_2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB13_1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: fcmp_une
+; CHECK: ucomiss %xmm1, %xmm0
+; CHECK-NEXT: jne {{LBB.+_2}}
+; CHECK-NEXT: jnp {{LBB.+_1}}
%1 = fcmp une float %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -271,16 +172,9 @@ bb1:
}
define i32 @icmp_eq(i32 %x, i32 %y) {
-; CHECK-LABEL: icmp_eq:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: cmpl %esi, %edi
-; CHECK-NEXT: je LBB14_2
-; CHECK-NEXT: ## %bb.1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB14_2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: icmp_eq
+; CHECK: cmpl %esi, %edi
+; CHECK-NEXT: jne {{LBB.+_1}}
%1 = icmp eq i32 %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -290,16 +184,9 @@ bb1:
}
define i32 @icmp_ne(i32 %x, i32 %y) {
-; CHECK-LABEL: icmp_ne:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: cmpl %esi, %edi
-; CHECK-NEXT: je LBB15_1
-; CHECK-NEXT: ## %bb.2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB15_1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: icmp_ne
+; CHECK: cmpl %esi, %edi
+; CHECK-NEXT: je {{LBB.+_1}}
%1 = icmp ne i32 %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -309,16 +196,9 @@ bb1:
}
define i32 @icmp_ugt(i32 %x, i32 %y) {
-; CHECK-LABEL: icmp_ugt:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: cmpl %esi, %edi
-; CHECK-NEXT: jbe LBB16_1
-; CHECK-NEXT: ## %bb.2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB16_1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: icmp_ugt
+; CHECK: cmpl %esi, %edi
+; CHECK-NEXT: jbe {{LBB.+_1}}
%1 = icmp ugt i32 %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -328,16 +208,9 @@ bb1:
}
define i32 @icmp_uge(i32 %x, i32 %y) {
-; CHECK-LABEL: icmp_uge:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: cmpl %esi, %edi
-; CHECK-NEXT: jae LBB17_2
-; CHECK-NEXT: ## %bb.1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB17_2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: icmp_uge
+; CHECK: cmpl %esi, %edi
+; CHECK-NEXT: jb {{LBB.+_1}}
%1 = icmp uge i32 %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -347,16 +220,9 @@ bb1:
}
define i32 @icmp_ult(i32 %x, i32 %y) {
-; CHECK-LABEL: icmp_ult:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: cmpl %esi, %edi
-; CHECK-NEXT: jae LBB18_1
-; CHECK-NEXT: ## %bb.2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB18_1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: icmp_ult
+; CHECK: cmpl %esi, %edi
+; CHECK-NEXT: jae {{LBB.+_1}}
%1 = icmp ult i32 %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -366,16 +232,9 @@ bb1:
}
define i32 @icmp_ule(i32 %x, i32 %y) {
-; CHECK-LABEL: icmp_ule:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: cmpl %esi, %edi
-; CHECK-NEXT: jbe LBB19_2
-; CHECK-NEXT: ## %bb.1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB19_2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: icmp_ule
+; CHECK: cmpl %esi, %edi
+; CHECK-NEXT: ja {{LBB.+_1}}
%1 = icmp ule i32 %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -385,16 +244,9 @@ bb1:
}
define i32 @icmp_sgt(i32 %x, i32 %y) {
-; CHECK-LABEL: icmp_sgt:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: cmpl %esi, %edi
-; CHECK-NEXT: jle LBB20_1
-; CHECK-NEXT: ## %bb.2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB20_1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: icmp_sgt
+; CHECK: cmpl %esi, %edi
+; CHECK-NEXT: jle {{LBB.+_1}}
%1 = icmp sgt i32 %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -404,16 +256,9 @@ bb1:
}
define i32 @icmp_sge(i32 %x, i32 %y) {
-; CHECK-LABEL: icmp_sge:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: cmpl %esi, %edi
-; CHECK-NEXT: jge LBB21_2
-; CHECK-NEXT: ## %bb.1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB21_2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: icmp_sge
+; CHECK: cmpl %esi, %edi
+; CHECK-NEXT: jl {{LBB.+_1}}
%1 = icmp sge i32 %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -423,16 +268,9 @@ bb1:
}
define i32 @icmp_slt(i32 %x, i32 %y) {
-; CHECK-LABEL: icmp_slt:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: cmpl %esi, %edi
-; CHECK-NEXT: jge LBB22_1
-; CHECK-NEXT: ## %bb.2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB22_1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: icmp_slt
+; CHECK: cmpl %esi, %edi
+; CHECK-NEXT: jge {{LBB.+_1}}
%1 = icmp slt i32 %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
@@ -442,16 +280,9 @@ bb1:
}
define i32 @icmp_sle(i32 %x, i32 %y) {
-; CHECK-LABEL: icmp_sle:
-; CHECK: ## %bb.0:
-; CHECK-NEXT: cmpl %esi, %edi
-; CHECK-NEXT: jle LBB23_2
-; CHECK-NEXT: ## %bb.1: ## %bb2
-; CHECK-NEXT: movl $1, %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: LBB23_2: ## %bb1
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retq
+; CHECK-LABEL: icmp_sle
+; CHECK: cmpl %esi, %edi
+; CHECK-NEXT: jg {{LBB.+_1}}
%1 = icmp sle i32 %x, %y
br i1 %1, label %bb1, label %bb2
bb2:
diff --git a/llvm/test/CodeGen/X86/funnel-shift.ll b/llvm/test/CodeGen/X86/funnel-shift.ll
index bb778fb61c04..76e45f43342f 100644
--- a/llvm/test/CodeGen/X86/funnel-shift.ll
+++ b/llvm/test/CodeGen/X86/funnel-shift.ll
@@ -941,17 +941,17 @@ define void @PR45265(i32 %0, %struct.S* nocapture readonly %1) nounwind {
; X32-SSE2-NEXT: shldl $24, %ebx, %edi
; X32-SSE2-NEXT: xorl %eax, %edi
; X32-SSE2-NEXT: orl %edi, %ecx
-; X32-SSE2-NEXT: je .LBB44_2
-; X32-SSE2-NEXT: # %bb.1:
+; X32-SSE2-NEXT: jne .LBB44_1
+; X32-SSE2-NEXT: # %bb.2:
; X32-SSE2-NEXT: popl %esi
; X32-SSE2-NEXT: popl %edi
; X32-SSE2-NEXT: popl %ebx
-; X32-SSE2-NEXT: retl
-; X32-SSE2-NEXT: .LBB44_2:
+; X32-SSE2-NEXT: jmp _Z3foov # TAILCALL
+; X32-SSE2-NEXT: .LBB44_1:
; X32-SSE2-NEXT: popl %esi
; X32-SSE2-NEXT: popl %edi
; X32-SSE2-NEXT: popl %ebx
-; X32-SSE2-NEXT: jmp _Z3foov # TAILCALL
+; X32-SSE2-NEXT: retl
;
; X64-AVX2-LABEL: PR45265:
; X64-AVX2: # %bb.0:
@@ -964,11 +964,11 @@ define void @PR45265(i32 %0, %struct.S* nocapture readonly %1) nounwind {
; X64-AVX2-NEXT: movq (%rsi,%rcx,4), %rcx
; X64-AVX2-NEXT: shrdq $40, %rdi, %rcx
; X64-AVX2-NEXT: cmpq %rax, %rcx
-; X64-AVX2-NEXT: je .LBB44_2
-; X64-AVX2-NEXT: # %bb.1:
-; X64-AVX2-NEXT: retq
-; X64-AVX2-NEXT: .LBB44_2:
+; X64-AVX2-NEXT: jne .LBB44_1
+; X64-AVX2-NEXT: # %bb.2:
; X64-AVX2-NEXT: jmp _Z3foov # TAILCALL
+; X64-AVX2-NEXT: .LBB44_1:
+; X64-AVX2-NEXT: retq
%3 = sext i32 %0 to i64
%4 = getelementptr inbounds %struct.S, %struct.S* %1, i64 %3
%5 = bitcast %struct.S* %4 to i88*
diff --git a/llvm/test/CodeGen/X86/indirect-branch-tracking-eh2.ll b/llvm/test/CodeGen/X86/indirect-branch-tracking-eh2.ll
index 8223b15ccafa..6e41c94e979a 100644
--- a/llvm/test/CodeGen/X86/indirect-branch-tracking-eh2.ll
+++ b/llvm/test/CodeGen/X86/indirect-branch-tracking-eh2.ll
@@ -1,183 +1,59 @@
-; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc -mtriple x86_64-unknown-unknown -exception-model sjlj -verify-machineinstrs=0 < %s | FileCheck %s --check-prefix=NUM
; RUN: llc -mtriple x86_64-unknown-unknown -exception-model sjlj -verify-machineinstrs=0 < %s | FileCheck %s --check-prefix=SJLJ
+; NUM-COUNT-3: endbr64
+
+;SJLJ: main: # @main
+;SJLJ-NEXT: .Lfunc_begin0:
+;SJLJ-NEXT: # %bb.0: # %entry
+;SJLJ-NEXT: endbr64
+;SJLJ-NEXT: pushq %rbp
+;SJLJ: callq _Unwind_SjLj_Register
+;SJLJ-NEXT: .Ltmp0:
+;SJLJ-NEXT: callq _Z3foov
+;SJLJ-NEXT: .Ltmp1:
+;SJLJ-NEXT: # %bb.1: # %invoke.cont
+;SJLJ-NEXT: movl
+;SJLJ-NEXT: .LBB0_7: # %return
+;SJLJ: callq _Unwind_SjLj_Unregister
+;SJLJ: retq
+;SJLJ-NEXT: .LBB0_9:
+;SJLJ-NEXT: endbr64
+;SJLJ-NEXT: movl
+;SJLJ-NEXT: cmpl
+;SJLJ-NEXT: jb .LBB0_10
+;SJLJ-NEXT: # %bb.11:
+;SJLJ-NEXT: ud2
+;SJLJ-NEXT: .LBB0_10:
+;SJLJ-NEXT: leaq .LJTI0_0(%rip), %rcx
+;SJLJ-NEXT: jmpq *(%rcx,%rax,8)
+;SJLJ-NEXT: .LBB0_2: # %lpad
+;SJLJ-NEXT: .Ltmp2:
+;SJLJ-NEXT: endbr64
+;SJLJ: jne .LBB0_4
+;SJLJ-NEXT: # %bb.3: # %catch3
+;SJLJ: callq __cxa_begin_catch
+;SJLJ: jmp .LBB0_6
+;SJLJ-NEXT: .LBB0_4: # %catch.fallthrough
+;SJLJ-NEXT: cmpl
+;SJLJ-NEXT: jne .LBB0_8
+;SJLJ-NEXT: # %bb.5: # %catch
+;SJLJ: callq __cxa_begin_catch
+;SJLJ: cmpb
+;SJLJ-NEXT: .LBB0_6: # %return
+;SJLJ: callq __cxa_end_catch
+;SJLJ-NEXT: jmp .LBB0_7
+;SJLJ-NEXT: .LBB0_8: # %eh.resume
+;SJLJ-NEXT: movl
+;SJLJ-NEXT: .Lfunc_end0:
+;SJLJ: .LJTI0_0:
+;SJLJ-NEXT: .quad .LBB0_2
+
@_ZTIi = external dso_local constant i8*
@_ZTIc = external dso_local constant i8*
; Function Attrs: noinline norecurse optnone uwtable
define dso_local i32 @main() #0 personality i8* bitcast (i32 (...)* @__gxx_personality_sj0 to i8*) {
-; NUM-LABEL: main:
-; NUM: # %bb.0: # %entry
-; NUM-NEXT: endbr64
-; NUM-NEXT: pushq %rbp
-; NUM-NEXT: movq %rsp, %rbp
-; NUM-NEXT: pushq %r15
-; NUM-NEXT: pushq %r14
-; NUM-NEXT: pushq %r13
-; NUM-NEXT: pushq %r12
-; NUM-NEXT: pushq %rbx
-; NUM-NEXT: subq $120, %rsp
-; NUM-NEXT: movl $0, -44(%rbp)
-; NUM-NEXT: movq $__gxx_personality_sj0, -120(%rbp)
-; NUM-NEXT: movq $GCC_except_table0, -112(%rbp)
-; NUM-NEXT: movq %rbp, -104(%rbp)
-; NUM-NEXT: movq %rsp, -88(%rbp)
-; NUM-NEXT: movq $.LBB0_9, -96(%rbp)
-; NUM-NEXT: movl $1, -144(%rbp)
-; NUM-NEXT: leaq -152(%rbp), %rdi
-; NUM-NEXT: callq _Unwind_SjLj_Register
-; NUM-NEXT: .Ltmp0:
-; NUM-NEXT: callq _Z3foov
-; NUM-NEXT: .Ltmp1:
-; NUM-NEXT: # %bb.1: # %invoke.cont
-; NUM-NEXT: movl $1, -44(%rbp)
-; NUM-NEXT: .LBB0_7: # %return
-; NUM-NEXT: movl -44(%rbp), %ebx
-; NUM-NEXT: leaq -152(%rbp), %rdi
-; NUM-NEXT: callq _Unwind_SjLj_Unregister
-; NUM-NEXT: movl %ebx, %eax
-; NUM-NEXT: addq $120, %rsp
-; NUM-NEXT: popq %rbx
-; NUM-NEXT: popq %r12
-; NUM-NEXT: popq %r13
-; NUM-NEXT: popq %r14
-; NUM-NEXT: popq %r15
-; NUM-NEXT: popq %rbp
-; NUM-NEXT: retq
-; NUM-NEXT: .LBB0_9:
-; NUM-NEXT: endbr64
-; NUM-NEXT: movl -144(%rbp), %eax
-; NUM-NEXT: cmpl $1, %eax
-; NUM-NEXT: jb .LBB0_10
-; NUM-NEXT: # %bb.11:
-; NUM-NEXT: ud2
-; NUM-NEXT: .LBB0_10:
-; NUM-NEXT: leaq {{.*}}(%rip), %rcx
-; NUM-NEXT: jmpq *(%rcx,%rax,8)
-; NUM-NEXT: .LBB0_2: # %lpad
-; NUM-NEXT: .Ltmp2:
-; NUM-NEXT: endbr64
-; NUM-NEXT: movl -140(%rbp), %ecx
-; NUM-NEXT: movl -136(%rbp), %eax
-; NUM-NEXT: movq %rcx, -56(%rbp)
-; NUM-NEXT: movl %eax, -64(%rbp)
-; NUM-NEXT: cmpl $2, %eax
-; NUM-NEXT: je .LBB0_3
-; NUM-NEXT: # %bb.4: # %catch.fallthrough
-; NUM-NEXT: cmpl $1, %eax
-; NUM-NEXT: jne .LBB0_8
-; NUM-NEXT: # %bb.5: # %catch
-; NUM-NEXT: movq -56(%rbp), %rdi
-; NUM-NEXT: movl $-1, -144(%rbp)
-; NUM-NEXT: callq __cxa_begin_catch
-; NUM-NEXT: movb (%rax), %al
-; NUM-NEXT: movb %al, -45(%rbp)
-; NUM-NEXT: xorl %ecx, %ecx
-; NUM-NEXT: cmpb $3, %al
-; NUM-NEXT: jmp .LBB0_6
-; NUM-NEXT: .LBB0_3: # %catch3
-; NUM-NEXT: movq -56(%rbp), %rdi
-; NUM-NEXT: movl $-1, -144(%rbp)
-; NUM-NEXT: callq __cxa_begin_catch
-; NUM-NEXT: movl (%rax), %eax
-; NUM-NEXT: movl %eax, -60(%rbp)
-; NUM-NEXT: xorl %ecx, %ecx
-; NUM-NEXT: cmpl $5, %eax
-; NUM-NEXT: .LBB0_6: # %return
-; NUM-NEXT: setne %cl
-; NUM-NEXT: movl %ecx, -44(%rbp)
-; NUM-NEXT: movl $-1, -144(%rbp)
-; NUM-NEXT: callq __cxa_end_catch
-; NUM-NEXT: jmp .LBB0_7
-; NUM-NEXT: .LBB0_8: # %eh.resume
-; NUM-NEXT: movl $-1, -144(%rbp)
-;
-; SJLJ-LABEL: main:
-; SJLJ: # %bb.0: # %entry
-; SJLJ-NEXT: endbr64
-; SJLJ-NEXT: pushq %rbp
-; SJLJ-NEXT: movq %rsp, %rbp
-; SJLJ-NEXT: pushq %r15
-; SJLJ-NEXT: pushq %r14
-; SJLJ-NEXT: pushq %r13
-; SJLJ-NEXT: pushq %r12
-; SJLJ-NEXT: pushq %rbx
-; SJLJ-NEXT: subq $120, %rsp
-; SJLJ-NEXT: movl $0, -44(%rbp)
-; SJLJ-NEXT: movq $__gxx_personality_sj0, -120(%rbp)
-; SJLJ-NEXT: movq $GCC_except_table0, -112(%rbp)
-; SJLJ-NEXT: movq %rbp, -104(%rbp)
-; SJLJ-NEXT: movq %rsp, -88(%rbp)
-; SJLJ-NEXT: movq $.LBB0_9, -96(%rbp)
-; SJLJ-NEXT: movl $1, -144(%rbp)
-; SJLJ-NEXT: leaq -152(%rbp), %rdi
-; SJLJ-NEXT: callq _Unwind_SjLj_Register
-; SJLJ-NEXT: .Ltmp0:
-; SJLJ-NEXT: callq _Z3foov
-; SJLJ-NEXT: .Ltmp1:
-; SJLJ-NEXT: # %bb.1: # %invoke.cont
-; SJLJ-NEXT: movl $1, -44(%rbp)
-; SJLJ-NEXT: .LBB0_7: # %return
-; SJLJ-NEXT: movl -44(%rbp), %ebx
-; SJLJ-NEXT: leaq -152(%rbp), %rdi
-; SJLJ-NEXT: callq _Unwind_SjLj_Unregister
-; SJLJ-NEXT: movl %ebx, %eax
-; SJLJ-NEXT: addq $120, %rsp
-; SJLJ-NEXT: popq %rbx
-; SJLJ-NEXT: popq %r12
-; SJLJ-NEXT: popq %r13
-; SJLJ-NEXT: popq %r14
-; SJLJ-NEXT: popq %r15
-; SJLJ-NEXT: popq %rbp
-; SJLJ-NEXT: retq
-; SJLJ-NEXT: .LBB0_9:
-; SJLJ-NEXT: endbr64
-; SJLJ-NEXT: movl -144(%rbp), %eax
-; SJLJ-NEXT: cmpl $1, %eax
-; SJLJ-NEXT: jb .LBB0_10
-; SJLJ-NEXT: # %bb.11:
-; SJLJ-NEXT: ud2
-; SJLJ-NEXT: .LBB0_10:
-; SJLJ-NEXT: leaq {{.*}}(%rip), %rcx
-; SJLJ-NEXT: jmpq *(%rcx,%rax,8)
-; SJLJ-NEXT: .LBB0_2: # %lpad
-; SJLJ-NEXT: .Ltmp2:
-; SJLJ-NEXT: endbr64
-; SJLJ-NEXT: movl -140(%rbp), %ecx
-; SJLJ-NEXT: movl -136(%rbp), %eax
-; SJLJ-NEXT: movq %rcx, -56(%rbp)
-; SJLJ-NEXT: movl %eax, -64(%rbp)
-; SJLJ-NEXT: cmpl $2, %eax
-; SJLJ-NEXT: je .LBB0_3
-; SJLJ-NEXT: # %bb.4: # %catch.fallthrough
-; SJLJ-NEXT: cmpl $1, %eax
-; SJLJ-NEXT: jne .LBB0_8
-; SJLJ-NEXT: # %bb.5: # %catch
-; SJLJ-NEXT: movq -56(%rbp), %rdi
-; SJLJ-NEXT: movl $-1, -144(%rbp)
-; SJLJ-NEXT: callq __cxa_begin_catch
-; SJLJ-NEXT: movb (%rax), %al
-; SJLJ-NEXT: movb %al, -45(%rbp)
-; SJLJ-NEXT: xorl %ecx, %ecx
-; SJLJ-NEXT: cmpb $3, %al
-; SJLJ-NEXT: jmp .LBB0_6
-; SJLJ-NEXT: .LBB0_3: # %catch3
-; SJLJ-NEXT: movq -56(%rbp), %rdi
-; SJLJ-NEXT: movl $-1, -144(%rbp)
-; SJLJ-NEXT: callq __cxa_begin_catch
-; SJLJ-NEXT: movl (%rax), %eax
-; SJLJ-NEXT: movl %eax, -60(%rbp)
-; SJLJ-NEXT: xorl %ecx, %ecx
-; SJLJ-NEXT: cmpl $5, %eax
-; SJLJ-NEXT: .LBB0_6: # %return
-; SJLJ-NEXT: setne %cl
-; SJLJ-NEXT: movl %ecx, -44(%rbp)
-; SJLJ-NEXT: movl $-1, -144(%rbp)
-; SJLJ-NEXT: callq __cxa_end_catch
-; SJLJ-NEXT: jmp .LBB0_7
-; SJLJ-NEXT: .LBB0_8: # %eh.resume
-; SJLJ-NEXT: movl $-1, -144(%rbp)
entry:
%retval = alloca i32, align 4
%exn.slot = alloca i8*
diff --git a/llvm/test/CodeGen/X86/jump_sign.ll b/llvm/test/CodeGen/X86/jump_sign.ll
index 6f09506f6666..78a322004380 100644
--- a/llvm/test/CodeGen/X86/jump_sign.ll
+++ b/llvm/test/CodeGen/X86/jump_sign.ll
@@ -139,12 +139,11 @@ define i32 @func_l2(i32 %a, i32 %b) nounwind {
; CHECK-NEXT: movl {{[0-9]+}}(%esp), %ecx
; CHECK-NEXT: movl %ecx, %eax
; CHECK-NEXT: subl %edx, %eax
-; CHECK-NEXT: je .LBB8_1
-; CHECK-NEXT: # %bb.2: # %if.else
-; CHECK-NEXT: retl
-; CHECK-NEXT: .LBB8_1: # %if.then
+; CHECK-NEXT: jne .LBB8_2
+; CHECK-NEXT: # %bb.1: # %if.then
; CHECK-NEXT: cmpl %ecx, %edx
; CHECK-NEXT: cmovlel %ecx, %eax
+; CHECK-NEXT: .LBB8_2: # %if.else
; CHECK-NEXT: retl
%cmp = icmp eq i32 %b, %a
%sub = sub nsw i32 %a, %b
@@ -330,13 +329,12 @@ define i8* @func_r(i8* %base, i32* nocapture %offset, i32 %size) nounwind {
; CHECK-NEXT: movl (%edx), %ecx
; CHECK-NEXT: xorl %eax, %eax
; CHECK-NEXT: subl {{[0-9]+}}(%esp), %ecx
-; CHECK-NEXT: jge .LBB15_1
-; CHECK-NEXT: # %bb.2: # %return
-; CHECK-NEXT: retl
-; CHECK-NEXT: .LBB15_1: # %if.end
+; CHECK-NEXT: jl .LBB15_2
+; CHECK-NEXT: # %bb.1: # %if.end
; CHECK-NEXT: movl {{[0-9]+}}(%esp), %eax
; CHECK-NEXT: movl %ecx, (%edx)
; CHECK-NEXT: addl %ecx, %eax
+; CHECK-NEXT: .LBB15_2: # %return
; CHECK-NEXT: retl
entry:
%0 = load i32, i32* %offset, align 8
diff --git a/llvm/test/CodeGen/X86/lsr-negative-stride.ll b/llvm/test/CodeGen/X86/lsr-negative-stride.ll
index 0d25e1414396..26c6128ab48d 100644
--- a/llvm/test/CodeGen/X86/lsr-negative-stride.ll
+++ b/llvm/test/CodeGen/X86/lsr-negative-stride.ll
@@ -19,7 +19,11 @@ define i32 @t(i32 %a, i32 %b) nounwind {
; CHECK-NEXT: movl {{[0-9]+}}(%esp), %ecx
; CHECK-NEXT: movl {{[0-9]+}}(%esp), %edx
; CHECK-NEXT: cmpl %ecx, %edx
-; CHECK-NEXT: je .LBB0_1
+; CHECK-NEXT: jne .LBB0_2
+; CHECK-NEXT: # %bb.1:
+; CHECK-NEXT: movl %edx, %eax
+; CHECK-NEXT: retl
+; CHECK-NEXT: .p2align 4, 0x90
; CHECK-NEXT: .LBB0_2: # %bb.outer
; CHECK-NEXT: # =>This Loop Header: Depth=1
; CHECK-NEXT: # Child Loop BB0_3 Depth 2
@@ -45,9 +49,6 @@ define i32 @t(i32 %a, i32 %b) nounwind {
; CHECK-NEXT: jne .LBB0_2
; CHECK-NEXT: .LBB0_6: # %bb17
; CHECK-NEXT: retl
-; CHECK-NEXT: .LBB0_1:
-; CHECK-NEXT: movl %edx, %eax
-; CHECK-NEXT: retl
entry:
%tmp1434 = icmp eq i32 %a, %b ; <i1> [#uses=1]
br i1 %tmp1434, label %bb17, label %bb.outer
diff --git a/llvm/test/CodeGen/X86/machine-cse.ll b/llvm/test/CodeGen/X86/machine-cse.ll
index 7ec9cafb5465..281aaca7c8d1 100644
--- a/llvm/test/CodeGen/X86/machine-cse.ll
+++ b/llvm/test/CodeGen/X86/machine-cse.ll
@@ -110,11 +110,10 @@ define i32 @cross_mbb_phys_cse(i32 %a, i32 %b) nounwind ssp {
; CHECK: # %bb.0: # %entry
; CHECK-NEXT: movl $1, %eax
; CHECK-NEXT: cmpl %esi, %edi
-; CHECK-NEXT: jbe .LBB2_1
-; CHECK-NEXT: # %bb.2: # %return
-; CHECK-NEXT: retq
-; CHECK-NEXT: .LBB2_1: # %if.end
+; CHECK-NEXT: ja .LBB2_2
+; CHECK-NEXT: # %bb.1: # %if.end
; CHECK-NEXT: sbbl %eax, %eax
+; CHECK-NEXT: .LBB2_2: # %return
; CHECK-NEXT: retq
entry:
%cmp = icmp ugt i32 %a, %b
diff --git a/llvm/test/CodeGen/X86/memcmp-more-load-pairs.ll b/llvm/test/CodeGen/X86/memcmp-more-load-pairs.ll
index 6c7076757d0e..374f573eed7b 100644
--- a/llvm/test/CodeGen/X86/memcmp-more-load-pairs.ll
+++ b/llvm/test/CodeGen/X86/memcmp-more-load-pairs.ll
@@ -239,19 +239,19 @@ define i32 @length3(i8* %X, i8* %Y) nounwind {
; X86-NEXT: rolw $8, %dx
; X86-NEXT: rolw $8, %si
; X86-NEXT: cmpw %si, %dx
-; X86-NEXT: je .LBB9_1
-; X86-NEXT: # %bb.3: # %res_block
-; X86-NEXT: setae %al
-; X86-NEXT: movzbl %al, %eax
-; X86-NEXT: leal -1(%eax,%eax), %eax
-; X86-NEXT: popl %esi
-; X86-NEXT: retl
-; X86-NEXT: .LBB9_1: # %loadbb1
+; X86-NEXT: jne .LBB9_3
+; X86-NEXT: # %bb.1: # %loadbb1
; X86-NEXT: movzbl 2(%eax), %eax
; X86-NEXT: movzbl 2(%ecx), %ecx
; X86-NEXT: subl %ecx, %eax
; X86-NEXT: popl %esi
; X86-NEXT: retl
+; X86-NEXT: .LBB9_3: # %res_block
+; X86-NEXT: setae %al
+; X86-NEXT: movzbl %al, %eax
+; X86-NEXT: leal -1(%eax,%eax), %eax
+; X86-NEXT: popl %esi
+; X86-NEXT: retl
;
; X64-LABEL: length3:
; X64: # %bb.0:
@@ -260,16 +260,16 @@ define i32 @length3(i8* %X, i8* %Y) nounwind {
; X64-NEXT: rolw $8, %ax
; X64-NEXT: rolw $8, %cx
; X64-NEXT: cmpw %cx, %ax
-; X64-NEXT: je .LBB9_1
-; X64-NEXT: # %bb.3: # %res_block
-; X64-NEXT: setae %al
-; X64-NEXT: movzbl %al, %eax
-; X64-NEXT: leal -1(%rax,%rax), %eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB9_1: # %loadbb1
+; X64-NEXT: jne .LBB9_3
+; X64-NEXT: # %bb.1: # %loadbb1
; X64-NEXT: movzbl 2(%rdi), %eax
; X64-NEXT: movzbl 2(%rsi), %ecx
; X64-NEXT: subl %ecx, %eax
+; X64-NEXT: retq
+; X64-NEXT: .LBB9_3: # %res_block
+; X64-NEXT: setae %al
+; X64-NEXT: movzbl %al, %eax
+; X64-NEXT: leal -1(%rax,%rax), %eax
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 3) nounwind
ret i32 %m
@@ -454,19 +454,19 @@ define i32 @length5(i8* %X, i8* %Y) nounwind {
; X86-NEXT: bswapl %edx
; X86-NEXT: bswapl %esi
; X86-NEXT: cmpl %esi, %edx
-; X86-NEXT: je .LBB16_1
-; X86-NEXT: # %bb.3: # %res_block
-; X86-NEXT: setae %al
-; X86-NEXT: movzbl %al, %eax
-; X86-NEXT: leal -1(%eax,%eax), %eax
-; X86-NEXT: popl %esi
-; X86-NEXT: retl
-; X86-NEXT: .LBB16_1: # %loadbb1
+; X86-NEXT: jne .LBB16_3
+; X86-NEXT: # %bb.1: # %loadbb1
; X86-NEXT: movzbl 4(%eax), %eax
; X86-NEXT: movzbl 4(%ecx), %ecx
; X86-NEXT: subl %ecx, %eax
; X86-NEXT: popl %esi
; X86-NEXT: retl
+; X86-NEXT: .LBB16_3: # %res_block
+; X86-NEXT: setae %al
+; X86-NEXT: movzbl %al, %eax
+; X86-NEXT: leal -1(%eax,%eax), %eax
+; X86-NEXT: popl %esi
+; X86-NEXT: retl
;
; X64-LABEL: length5:
; X64: # %bb.0:
@@ -475,16 +475,16 @@ define i32 @length5(i8* %X, i8* %Y) nounwind {
; X64-NEXT: bswapl %eax
; X64-NEXT: bswapl %ecx
; X64-NEXT: cmpl %ecx, %eax
-; X64-NEXT: je .LBB16_1
-; X64-NEXT: # %bb.3: # %res_block
-; X64-NEXT: setae %al
-; X64-NEXT: movzbl %al, %eax
-; X64-NEXT: leal -1(%rax,%rax), %eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB16_1: # %loadbb1
+; X64-NEXT: jne .LBB16_3
+; X64-NEXT: # %bb.1: # %loadbb1
; X64-NEXT: movzbl 4(%rdi), %eax
; X64-NEXT: movzbl 4(%rsi), %ecx
; X64-NEXT: subl %ecx, %eax
+; X64-NEXT: retq
+; X64-NEXT: .LBB16_3: # %res_block
+; X64-NEXT: setae %al
+; X64-NEXT: movzbl %al, %eax
+; X64-NEXT: leal -1(%rax,%rax), %eax
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 5) nounwind
ret i32 %m
@@ -530,16 +530,16 @@ define i1 @length5_lt(i8* %X, i8* %Y) nounwind {
; X86-NEXT: bswapl %edx
; X86-NEXT: bswapl %esi
; X86-NEXT: cmpl %esi, %edx
-; X86-NEXT: je .LBB18_1
-; X86-NEXT: # %bb.3: # %res_block
-; X86-NEXT: setae %al
-; X86-NEXT: movzbl %al, %eax
-; X86-NEXT: leal -1(%eax,%eax), %eax
-; X86-NEXT: jmp .LBB18_2
-; X86-NEXT: .LBB18_1: # %loadbb1
+; X86-NEXT: jne .LBB18_3
+; X86-NEXT: # %bb.1: # %loadbb1
; X86-NEXT: movzbl 4(%eax), %eax
; X86-NEXT: movzbl 4(%ecx), %ecx
; X86-NEXT: subl %ecx, %eax
+; X86-NEXT: jmp .LBB18_2
+; X86-NEXT: .LBB18_3: # %res_block
+; X86-NEXT: setae %al
+; X86-NEXT: movzbl %al, %eax
+; X86-NEXT: leal -1(%eax,%eax), %eax
; X86-NEXT: .LBB18_2: # %endblock
; X86-NEXT: shrl $31, %eax
; X86-NEXT: # kill: def $al killed $al killed $eax
@@ -553,20 +553,20 @@ define i1 @length5_lt(i8* %X, i8* %Y) nounwind {
; X64-NEXT: bswapl %eax
; X64-NEXT: bswapl %ecx
; X64-NEXT: cmpl %ecx, %eax
-; X64-NEXT: je .LBB18_1
-; X64-NEXT: # %bb.3: # %res_block
-; X64-NEXT: setae %al
-; X64-NEXT: movzbl %al, %eax
-; X64-NEXT: leal -1(%rax,%rax), %eax
-; X64-NEXT: shrl $31, %eax
-; X64-NEXT: # kill: def $al killed $al killed $eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB18_1: # %loadbb1
+; X64-NEXT: jne .LBB18_3
+; X64-NEXT: # %bb.1: # %loadbb1
; X64-NEXT: movzbl 4(%rdi), %eax
; X64-NEXT: movzbl 4(%rsi), %ecx
; X64-NEXT: subl %ecx, %eax
; X64-NEXT: shrl $31, %eax
; X64-NEXT: # kill: def $al killed $al killed $eax
+; X64-NEXT: retq
+; X64-NEXT: .LBB18_3: # %res_block
+; X64-NEXT: setae %al
+; X64-NEXT: movzbl %al, %eax
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: shrl $31, %eax
+; X64-NEXT: # kill: def $al killed $al killed $eax
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 5) nounwind
%c = icmp slt i32 %m, 0
@@ -577,56 +577,53 @@ define i32 @length7(i8* %X, i8* %Y) nounwind {
; X86-LABEL: length7:
; X86: # %bb.0:
; X86-NEXT: pushl %esi
-; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
+; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl {{[0-9]+}}(%esp), %esi
-; X86-NEXT: movl (%esi), %eax
-; X86-NEXT: movl (%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl (%esi), %ecx
+; X86-NEXT: movl (%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: jne .LBB19_4
+; X86-NEXT: bswapl %edx
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: jne .LBB19_2
; X86-NEXT: # %bb.1: # %loadbb1
-; X86-NEXT: movl 3(%esi), %eax
-; X86-NEXT: movl 3(%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl 3(%esi), %ecx
+; X86-NEXT: movl 3(%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: je .LBB19_2
-; X86-NEXT: .LBB19_4: # %res_block
-; X86-NEXT: xorl %edx, %edx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: setae %dl
-; X86-NEXT: leal -1(%edx,%edx), %eax
-; X86-NEXT: popl %esi
-; X86-NEXT: retl
-; X86-NEXT: .LBB19_2:
+; X86-NEXT: bswapl %edx
; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: je .LBB19_3
+; X86-NEXT: .LBB19_2: # %res_block
+; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: setae %al
+; X86-NEXT: leal -1(%eax,%eax), %eax
+; X86-NEXT: .LBB19_3: # %endblock
; X86-NEXT: popl %esi
; X86-NEXT: retl
;
; X64-LABEL: length7:
; X64: # %bb.0:
-; X64-NEXT: movl (%rdi), %eax
-; X64-NEXT: movl (%rsi), %ecx
-; X64-NEXT: bswapl %eax
+; X64-NEXT: movl (%rdi), %ecx
+; X64-NEXT: movl (%rsi), %edx
; X64-NEXT: bswapl %ecx
-; X64-NEXT: cmpl %ecx, %eax
-; X64-NEXT: jne .LBB19_4
+; X64-NEXT: bswapl %edx
+; X64-NEXT: cmpl %edx, %ecx
+; X64-NEXT: jne .LBB19_2
; X64-NEXT: # %bb.1: # %loadbb1
-; X64-NEXT: movl 3(%rdi), %eax
-; X64-NEXT: movl 3(%rsi), %ecx
-; X64-NEXT: bswapl %eax
+; X64-NEXT: movl 3(%rdi), %ecx
+; X64-NEXT: movl 3(%rsi), %edx
; X64-NEXT: bswapl %ecx
-; X64-NEXT: cmpl %ecx, %eax
-; X64-NEXT: je .LBB19_2
-; X64-NEXT: .LBB19_4: # %res_block
-; X64-NEXT: xorl %edx, %edx
-; X64-NEXT: cmpl %ecx, %eax
-; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB19_2:
+; X64-NEXT: bswapl %edx
+; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpl %edx, %ecx
+; X64-NEXT: je .LBB19_3
+; X64-NEXT: .LBB19_2: # %res_block
; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpl %edx, %ecx
+; X64-NEXT: setae %al
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: .LBB19_3: # %endblock
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 7) nounwind
ret i32 %m
@@ -663,60 +660,55 @@ define i1 @length7_lt(i8* %X, i8* %Y) nounwind {
; X86-LABEL: length7_lt:
; X86: # %bb.0:
; X86-NEXT: pushl %esi
-; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
+; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl {{[0-9]+}}(%esp), %esi
-; X86-NEXT: movl (%esi), %eax
-; X86-NEXT: movl (%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl (%esi), %ecx
+; X86-NEXT: movl (%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: jne .LBB21_4
+; X86-NEXT: bswapl %edx
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: jne .LBB21_2
; X86-NEXT: # %bb.1: # %loadbb1
-; X86-NEXT: movl 3(%esi), %eax
-; X86-NEXT: movl 3(%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl 3(%esi), %ecx
+; X86-NEXT: movl 3(%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: je .LBB21_2
-; X86-NEXT: .LBB21_4: # %res_block
-; X86-NEXT: xorl %edx, %edx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: setae %dl
-; X86-NEXT: leal -1(%edx,%edx), %eax
+; X86-NEXT: bswapl %edx
+; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: je .LBB21_3
+; X86-NEXT: .LBB21_2: # %res_block
+; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: setae %al
+; X86-NEXT: leal -1(%eax,%eax), %eax
; X86-NEXT: .LBB21_3: # %endblock
; X86-NEXT: shrl $31, %eax
; X86-NEXT: # kill: def $al killed $al killed $eax
; X86-NEXT: popl %esi
; X86-NEXT: retl
-; X86-NEXT: .LBB21_2:
-; X86-NEXT: xorl %eax, %eax
-; X86-NEXT: jmp .LBB21_3
;
; X64-LABEL: length7_lt:
; X64: # %bb.0:
-; X64-NEXT: movl (%rdi), %eax
-; X64-NEXT: movl (%rsi), %ecx
-; X64-NEXT: bswapl %eax
+; X64-NEXT: movl (%rdi), %ecx
+; X64-NEXT: movl (%rsi), %edx
; X64-NEXT: bswapl %ecx
-; X64-NEXT: cmpl %ecx, %eax
+; X64-NEXT: bswapl %edx
+; X64-NEXT: cmpl %edx, %ecx
; X64-NEXT: jne .LBB21_2
; X64-NEXT: # %bb.1: # %loadbb1
-; X64-NEXT: movl 3(%rdi), %eax
-; X64-NEXT: movl 3(%rsi), %ecx
-; X64-NEXT: bswapl %eax
+; X64-NEXT: movl 3(%rdi), %ecx
+; X64-NEXT: movl 3(%rsi), %edx
; X64-NEXT: bswapl %ecx
-; X64-NEXT: cmpl %ecx, %eax
+; X64-NEXT: bswapl %edx
+; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpl %edx, %ecx
; X64-NEXT: je .LBB21_3
; X64-NEXT: .LBB21_2: # %res_block
-; X64-NEXT: xorl %edx, %edx
-; X64-NEXT: cmpl %ecx, %eax
-; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
-; X64-NEXT: shrl $31, %eax
-; X64-NEXT: # kill: def $al killed $al killed $eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB21_3:
; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpl %edx, %ecx
+; X64-NEXT: setae %al
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: .LBB21_3: # %endblock
; X64-NEXT: shrl $31, %eax
; X64-NEXT: # kill: def $al killed $al killed $eax
; X64-NEXT: retq
@@ -729,30 +721,28 @@ define i32 @length8(i8* %X, i8* %Y) nounwind {
; X86-LABEL: length8:
; X86: # %bb.0:
; X86-NEXT: pushl %esi
-; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
+; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl {{[0-9]+}}(%esp), %esi
-; X86-NEXT: movl (%esi), %eax
-; X86-NEXT: movl (%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl (%esi), %ecx
+; X86-NEXT: movl (%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: jne .LBB22_4
+; X86-NEXT: bswapl %edx
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: jne .LBB22_2
; X86-NEXT: # %bb.1: # %loadbb1
-; X86-NEXT: movl 4(%esi), %eax
-; X86-NEXT: movl 4(%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl 4(%esi), %ecx
+; X86-NEXT: movl 4(%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: je .LBB22_2
-; X86-NEXT: .LBB22_4: # %res_block
-; X86-NEXT: xorl %edx, %edx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: setae %dl
-; X86-NEXT: leal -1(%edx,%edx), %eax
-; X86-NEXT: popl %esi
-; X86-NEXT: retl
-; X86-NEXT: .LBB22_2:
+; X86-NEXT: bswapl %edx
; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: je .LBB22_3
+; X86-NEXT: .LBB22_2: # %res_block
+; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: setae %al
+; X86-NEXT: leal -1(%eax,%eax), %eax
+; X86-NEXT: .LBB22_3: # %endblock
; X86-NEXT: popl %esi
; X86-NEXT: retl
;
@@ -954,63 +944,60 @@ define i32 @length12(i8* %X, i8* %Y) nounwind {
; X86-LABEL: length12:
; X86: # %bb.0:
; X86-NEXT: pushl %esi
-; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
+; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl {{[0-9]+}}(%esp), %esi
-; X86-NEXT: movl (%esi), %eax
-; X86-NEXT: movl (%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl (%esi), %ecx
+; X86-NEXT: movl (%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: jne .LBB29_5
+; X86-NEXT: bswapl %edx
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: jne .LBB29_3
; X86-NEXT: # %bb.1: # %loadbb1
-; X86-NEXT: movl 4(%esi), %eax
-; X86-NEXT: movl 4(%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl 4(%esi), %ecx
+; X86-NEXT: movl 4(%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: jne .LBB29_5
+; X86-NEXT: bswapl %edx
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: jne .LBB29_3
; X86-NEXT: # %bb.2: # %loadbb2
-; X86-NEXT: movl 8(%esi), %eax
-; X86-NEXT: movl 8(%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl 8(%esi), %ecx
+; X86-NEXT: movl 8(%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: je .LBB29_3
-; X86-NEXT: .LBB29_5: # %res_block
-; X86-NEXT: xorl %edx, %edx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: setae %dl
-; X86-NEXT: leal -1(%edx,%edx), %eax
-; X86-NEXT: popl %esi
-; X86-NEXT: retl
-; X86-NEXT: .LBB29_3:
+; X86-NEXT: bswapl %edx
+; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: je .LBB29_4
+; X86-NEXT: .LBB29_3: # %res_block
; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: setae %al
+; X86-NEXT: leal -1(%eax,%eax), %eax
+; X86-NEXT: .LBB29_4: # %endblock
; X86-NEXT: popl %esi
; X86-NEXT: retl
;
; X64-LABEL: length12:
; X64: # %bb.0:
-; X64-NEXT: movq (%rdi), %rax
-; X64-NEXT: movq (%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq (%rdi), %rcx
+; X64-NEXT: movq (%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB29_4
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: jne .LBB29_2
; X64-NEXT: # %bb.1: # %loadbb1
-; X64-NEXT: movl 8(%rdi), %eax
-; X64-NEXT: movl 8(%rsi), %ecx
-; X64-NEXT: bswapl %eax
+; X64-NEXT: movl 8(%rdi), %ecx
+; X64-NEXT: movl 8(%rsi), %edx
; X64-NEXT: bswapl %ecx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: je .LBB29_2
-; X64-NEXT: .LBB29_4: # %res_block
-; X64-NEXT: xorl %edx, %edx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB29_2:
+; X64-NEXT: bswapl %edx
+; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: je .LBB29_3
+; X64-NEXT: .LBB29_2: # %res_block
; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: setae %al
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: .LBB29_3: # %endblock
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 12) nounwind
ret i32 %m
@@ -1129,70 +1116,67 @@ define i32 @length16(i8* %X, i8* %Y) nounwind {
; X86-LABEL: length16:
; X86: # %bb.0:
; X86-NEXT: pushl %esi
-; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
+; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl {{[0-9]+}}(%esp), %esi
-; X86-NEXT: movl (%esi), %eax
-; X86-NEXT: movl (%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl (%esi), %ecx
+; X86-NEXT: movl (%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: jne .LBB33_6
+; X86-NEXT: bswapl %edx
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: jne .LBB33_4
; X86-NEXT: # %bb.1: # %loadbb1
-; X86-NEXT: movl 4(%esi), %eax
-; X86-NEXT: movl 4(%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl 4(%esi), %ecx
+; X86-NEXT: movl 4(%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: jne .LBB33_6
+; X86-NEXT: bswapl %edx
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: jne .LBB33_4
; X86-NEXT: # %bb.2: # %loadbb2
-; X86-NEXT: movl 8(%esi), %eax
-; X86-NEXT: movl 8(%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl 8(%esi), %ecx
+; X86-NEXT: movl 8(%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: jne .LBB33_6
+; X86-NEXT: bswapl %edx
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: jne .LBB33_4
; X86-NEXT: # %bb.3: # %loadbb3
-; X86-NEXT: movl 12(%esi), %eax
-; X86-NEXT: movl 12(%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl 12(%esi), %ecx
+; X86-NEXT: movl 12(%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: je .LBB33_4
-; X86-NEXT: .LBB33_6: # %res_block
-; X86-NEXT: xorl %edx, %edx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: setae %dl
-; X86-NEXT: leal -1(%edx,%edx), %eax
-; X86-NEXT: popl %esi
-; X86-NEXT: retl
-; X86-NEXT: .LBB33_4:
+; X86-NEXT: bswapl %edx
+; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: je .LBB33_5
+; X86-NEXT: .LBB33_4: # %res_block
; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: setae %al
+; X86-NEXT: leal -1(%eax,%eax), %eax
+; X86-NEXT: .LBB33_5: # %endblock
; X86-NEXT: popl %esi
; X86-NEXT: retl
;
; X64-LABEL: length16:
; X64: # %bb.0:
-; X64-NEXT: movq (%rdi), %rax
-; X64-NEXT: movq (%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq (%rdi), %rcx
+; X64-NEXT: movq (%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB33_4
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: jne .LBB33_2
; X64-NEXT: # %bb.1: # %loadbb1
-; X64-NEXT: movq 8(%rdi), %rax
-; X64-NEXT: movq 8(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq 8(%rdi), %rcx
+; X64-NEXT: movq 8(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: je .LBB33_2
-; X64-NEXT: .LBB33_4: # %res_block
-; X64-NEXT: xorl %edx, %edx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB33_2:
+; X64-NEXT: bswapq %rdx
; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: je .LBB33_3
+; X64-NEXT: .LBB33_2: # %res_block
+; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: setae %al
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: .LBB33_3: # %endblock
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 16) nounwind
ret i32 %m
@@ -1307,74 +1291,69 @@ define i1 @length16_lt(i8* %x, i8* %y) nounwind {
; X86-LABEL: length16_lt:
; X86: # %bb.0:
; X86-NEXT: pushl %esi
-; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
+; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl {{[0-9]+}}(%esp), %esi
-; X86-NEXT: movl (%esi), %eax
-; X86-NEXT: movl (%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl (%esi), %ecx
+; X86-NEXT: movl (%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: jne .LBB35_6
+; X86-NEXT: bswapl %edx
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: jne .LBB35_4
; X86-NEXT: # %bb.1: # %loadbb1
-; X86-NEXT: movl 4(%esi), %eax
-; X86-NEXT: movl 4(%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl 4(%esi), %ecx
+; X86-NEXT: movl 4(%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: jne .LBB35_6
+; X86-NEXT: bswapl %edx
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: jne .LBB35_4
; X86-NEXT: # %bb.2: # %loadbb2
-; X86-NEXT: movl 8(%esi), %eax
-; X86-NEXT: movl 8(%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl 8(%esi), %ecx
+; X86-NEXT: movl 8(%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: jne .LBB35_6
-; X86-NEXT: # %bb.3: # %loadbb3
-; X86-NEXT: movl 12(%esi), %eax
-; X86-NEXT: movl 12(%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: bswapl %edx
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: jne .LBB35_4
+; X86-NEXT: # %bb.3: # %loadbb3
+; X86-NEXT: movl 12(%esi), %ecx
+; X86-NEXT: movl 12(%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: je .LBB35_4
-; X86-NEXT: .LBB35_6: # %res_block
-; X86-NEXT: xorl %edx, %edx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: setae %dl
-; X86-NEXT: leal -1(%edx,%edx), %eax
+; X86-NEXT: bswapl %edx
+; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: je .LBB35_5
+; X86-NEXT: .LBB35_4: # %res_block
+; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: setae %al
+; X86-NEXT: leal -1(%eax,%eax), %eax
; X86-NEXT: .LBB35_5: # %endblock
; X86-NEXT: shrl $31, %eax
; X86-NEXT: # kill: def $al killed $al killed $eax
; X86-NEXT: popl %esi
; X86-NEXT: retl
-; X86-NEXT: .LBB35_4:
-; X86-NEXT: xorl %eax, %eax
-; X86-NEXT: jmp .LBB35_5
;
; X64-LABEL: length16_lt:
; X64: # %bb.0:
-; X64-NEXT: movq (%rdi), %rax
-; X64-NEXT: movq (%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq (%rdi), %rcx
+; X64-NEXT: movq (%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
; X64-NEXT: jne .LBB35_2
; X64-NEXT: # %bb.1: # %loadbb1
-; X64-NEXT: movq 8(%rdi), %rax
-; X64-NEXT: movq 8(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq 8(%rdi), %rcx
+; X64-NEXT: movq 8(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
; X64-NEXT: je .LBB35_3
; X64-NEXT: .LBB35_2: # %res_block
-; X64-NEXT: xorl %edx, %edx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
-; X64-NEXT: shrl $31, %eax
-; X64-NEXT: # kill: def $al killed $al killed $eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB35_3:
; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: setae %al
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: .LBB35_3: # %endblock
; X64-NEXT: shrl $31, %eax
; X64-NEXT: # kill: def $al killed $al killed $eax
; X64-NEXT: retq
@@ -1394,41 +1373,39 @@ define i1 @length16_gt(i8* %x, i8* %y) nounwind {
; X86-NEXT: bswapl %eax
; X86-NEXT: bswapl %ecx
; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: jne .LBB36_6
+; X86-NEXT: jne .LBB36_4
; X86-NEXT: # %bb.1: # %loadbb1
; X86-NEXT: movl 4(%esi), %eax
; X86-NEXT: movl 4(%edx), %ecx
; X86-NEXT: bswapl %eax
; X86-NEXT: bswapl %ecx
; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: jne .LBB36_6
+; X86-NEXT: jne .LBB36_4
; X86-NEXT: # %bb.2: # %loadbb2
; X86-NEXT: movl 8(%esi), %eax
; X86-NEXT: movl 8(%edx), %ecx
; X86-NEXT: bswapl %eax
; X86-NEXT: bswapl %ecx
; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: jne .LBB36_6
+; X86-NEXT: jne .LBB36_4
; X86-NEXT: # %bb.3: # %loadbb3
; X86-NEXT: movl 12(%esi), %eax
; X86-NEXT: movl 12(%edx), %ecx
; X86-NEXT: bswapl %eax
; X86-NEXT: bswapl %ecx
+; X86-NEXT: xorl %edx, %edx
; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: je .LBB36_4
-; X86-NEXT: .LBB36_6: # %res_block
+; X86-NEXT: je .LBB36_5
+; X86-NEXT: .LBB36_4: # %res_block
; X86-NEXT: xorl %edx, %edx
; X86-NEXT: cmpl %ecx, %eax
; X86-NEXT: setae %dl
-; X86-NEXT: leal -1(%edx,%edx), %eax
+; X86-NEXT: leal -1(%edx,%edx), %edx
; X86-NEXT: .LBB36_5: # %endblock
-; X86-NEXT: testl %eax, %eax
+; X86-NEXT: testl %edx, %edx
; X86-NEXT: setg %al
; X86-NEXT: popl %esi
; X86-NEXT: retl
-; X86-NEXT: .LBB36_4:
-; X86-NEXT: xorl %eax, %eax
-; X86-NEXT: jmp .LBB36_5
;
; X64-LABEL: length16_gt:
; X64: # %bb.0:
@@ -1437,26 +1414,24 @@ define i1 @length16_gt(i8* %x, i8* %y) nounwind {
; X64-NEXT: bswapq %rax
; X64-NEXT: bswapq %rcx
; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB36_4
+; X64-NEXT: jne .LBB36_2
; X64-NEXT: # %bb.1: # %loadbb1
; X64-NEXT: movq 8(%rdi), %rax
; X64-NEXT: movq 8(%rsi), %rcx
; X64-NEXT: bswapq %rax
; X64-NEXT: bswapq %rcx
+; X64-NEXT: xorl %edx, %edx
; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: je .LBB36_2
-; X64-NEXT: .LBB36_4: # %res_block
+; X64-NEXT: je .LBB36_3
+; X64-NEXT: .LBB36_2: # %res_block
; X64-NEXT: xorl %edx, %edx
; X64-NEXT: cmpq %rcx, %rax
; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
+; X64-NEXT: leal -1(%rdx,%rdx), %edx
; X64-NEXT: .LBB36_3: # %endblock
-; X64-NEXT: testl %eax, %eax
+; X64-NEXT: testl %edx, %edx
; X64-NEXT: setg %al
; X64-NEXT: retq
-; X64-NEXT: .LBB36_2:
-; X64-NEXT: xorl %eax, %eax
-; X64-NEXT: jmp .LBB36_3
%call = tail call i32 @memcmp(i8* %x, i8* %y, i64 16) nounwind
%cmp = icmp sgt i32 %call, 0
ret i1 %cmp
@@ -1574,34 +1549,33 @@ define i32 @length24(i8* %X, i8* %Y) nounwind {
;
; X64-LABEL: length24:
; X64: # %bb.0:
-; X64-NEXT: movq (%rdi), %rax
-; X64-NEXT: movq (%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq (%rdi), %rcx
+; X64-NEXT: movq (%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB38_5
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: jne .LBB38_3
; X64-NEXT: # %bb.1: # %loadbb1
-; X64-NEXT: movq 8(%rdi), %rax
-; X64-NEXT: movq 8(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq 8(%rdi), %rcx
+; X64-NEXT: movq 8(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB38_5
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: jne .LBB38_3
; X64-NEXT: # %bb.2: # %loadbb2
-; X64-NEXT: movq 16(%rdi), %rax
-; X64-NEXT: movq 16(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq 16(%rdi), %rcx
+; X64-NEXT: movq 16(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: je .LBB38_3
-; X64-NEXT: .LBB38_5: # %res_block
-; X64-NEXT: xorl %edx, %edx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB38_3:
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: je .LBB38_4
+; X64-NEXT: .LBB38_3: # %res_block
; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: setae %al
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: .LBB38_4: # %endblock
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 24) nounwind
ret i32 %m
@@ -1734,36 +1708,33 @@ define i1 @length24_lt(i8* %x, i8* %y) nounwind {
;
; X64-LABEL: length24_lt:
; X64: # %bb.0:
-; X64-NEXT: movq (%rdi), %rax
-; X64-NEXT: movq (%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq (%rdi), %rcx
+; X64-NEXT: movq (%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
; X64-NEXT: jne .LBB40_3
; X64-NEXT: # %bb.1: # %loadbb1
-; X64-NEXT: movq 8(%rdi), %rax
-; X64-NEXT: movq 8(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq 8(%rdi), %rcx
+; X64-NEXT: movq 8(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
; X64-NEXT: jne .LBB40_3
; X64-NEXT: # %bb.2: # %loadbb2
-; X64-NEXT: movq 16(%rdi), %rax
-; X64-NEXT: movq 16(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq 16(%rdi), %rcx
+; X64-NEXT: movq 16(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
; X64-NEXT: je .LBB40_4
; X64-NEXT: .LBB40_3: # %res_block
-; X64-NEXT: xorl %edx, %edx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
-; X64-NEXT: shrl $31, %eax
-; X64-NEXT: # kill: def $al killed $al killed $eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB40_4:
; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: setae %al
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: .LBB40_4: # %endblock
; X64-NEXT: shrl $31, %eax
; X64-NEXT: # kill: def $al killed $al killed $eax
; X64-NEXT: retq
@@ -1792,33 +1763,31 @@ define i1 @length24_gt(i8* %x, i8* %y) nounwind {
; X64-NEXT: bswapq %rax
; X64-NEXT: bswapq %rcx
; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB41_5
+; X64-NEXT: jne .LBB41_3
; X64-NEXT: # %bb.1: # %loadbb1
; X64-NEXT: movq 8(%rdi), %rax
; X64-NEXT: movq 8(%rsi), %rcx
; X64-NEXT: bswapq %rax
; X64-NEXT: bswapq %rcx
; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB41_5
+; X64-NEXT: jne .LBB41_3
; X64-NEXT: # %bb.2: # %loadbb2
; X64-NEXT: movq 16(%rdi), %rax
; X64-NEXT: movq 16(%rsi), %rcx
; X64-NEXT: bswapq %rax
; X64-NEXT: bswapq %rcx
+; X64-NEXT: xorl %edx, %edx
; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: je .LBB41_3
-; X64-NEXT: .LBB41_5: # %res_block
+; X64-NEXT: je .LBB41_4
+; X64-NEXT: .LBB41_3: # %res_block
; X64-NEXT: xorl %edx, %edx
; X64-NEXT: cmpq %rcx, %rax
; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
+; X64-NEXT: leal -1(%rdx,%rdx), %edx
; X64-NEXT: .LBB41_4: # %endblock
-; X64-NEXT: testl %eax, %eax
+; X64-NEXT: testl %edx, %edx
; X64-NEXT: setg %al
; X64-NEXT: retq
-; X64-NEXT: .LBB41_3:
-; X64-NEXT: xorl %eax, %eax
-; X64-NEXT: jmp .LBB41_4
%call = tail call i32 @memcmp(i8* %x, i8* %y, i64 24) nounwind
%cmp = icmp sgt i32 %call, 0
ret i1 %cmp
@@ -1938,41 +1907,40 @@ define i32 @length31(i8* %X, i8* %Y) nounwind {
;
; X64-LABEL: length31:
; X64: # %bb.0:
-; X64-NEXT: movq (%rdi), %rax
-; X64-NEXT: movq (%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq (%rdi), %rcx
+; X64-NEXT: movq (%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB43_6
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: jne .LBB43_4
; X64-NEXT: # %bb.1: # %loadbb1
-; X64-NEXT: movq 8(%rdi), %rax
-; X64-NEXT: movq 8(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq 8(%rdi), %rcx
+; X64-NEXT: movq 8(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB43_6
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: jne .LBB43_4
; X64-NEXT: # %bb.2: # %loadbb2
-; X64-NEXT: movq 16(%rdi), %rax
-; X64-NEXT: movq 16(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq 16(%rdi), %rcx
+; X64-NEXT: movq 16(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB43_6
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: jne .LBB43_4
; X64-NEXT: # %bb.3: # %loadbb3
-; X64-NEXT: movq 23(%rdi), %rax
-; X64-NEXT: movq 23(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq 23(%rdi), %rcx
+; X64-NEXT: movq 23(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: je .LBB43_4
-; X64-NEXT: .LBB43_6: # %res_block
-; X64-NEXT: xorl %edx, %edx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB43_4:
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: je .LBB43_5
+; X64-NEXT: .LBB43_4: # %res_block
; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: setae %al
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: .LBB43_5: # %endblock
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 31) nounwind
ret i32 %m
@@ -2104,43 +2072,40 @@ define i1 @length31_lt(i8* %x, i8* %y) nounwind {
;
; X64-LABEL: length31_lt:
; X64: # %bb.0:
-; X64-NEXT: movq (%rdi), %rax
-; X64-NEXT: movq (%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq (%rdi), %rcx
+; X64-NEXT: movq (%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
; X64-NEXT: jne .LBB45_4
; X64-NEXT: # %bb.1: # %loadbb1
-; X64-NEXT: movq 8(%rdi), %rax
-; X64-NEXT: movq 8(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq 8(%rdi), %rcx
+; X64-NEXT: movq 8(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
; X64-NEXT: jne .LBB45_4
; X64-NEXT: # %bb.2: # %loadbb2
-; X64-NEXT: movq 16(%rdi), %rax
-; X64-NEXT: movq 16(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq 16(%rdi), %rcx
+; X64-NEXT: movq 16(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
; X64-NEXT: jne .LBB45_4
; X64-NEXT: # %bb.3: # %loadbb3
-; X64-NEXT: movq 23(%rdi), %rax
-; X64-NEXT: movq 23(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq 23(%rdi), %rcx
+; X64-NEXT: movq 23(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
; X64-NEXT: je .LBB45_5
; X64-NEXT: .LBB45_4: # %res_block
-; X64-NEXT: xorl %edx, %edx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
-; X64-NEXT: shrl $31, %eax
-; X64-NEXT: # kill: def $al killed $al killed $eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB45_5:
; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: setae %al
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: .LBB45_5: # %endblock
; X64-NEXT: shrl $31, %eax
; X64-NEXT: # kill: def $al killed $al killed $eax
; X64-NEXT: retq
@@ -2169,40 +2134,38 @@ define i1 @length31_gt(i8* %x, i8* %y) nounwind {
; X64-NEXT: bswapq %rax
; X64-NEXT: bswapq %rcx
; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB46_6
+; X64-NEXT: jne .LBB46_4
; X64-NEXT: # %bb.1: # %loadbb1
; X64-NEXT: movq 8(%rdi), %rax
; X64-NEXT: movq 8(%rsi), %rcx
; X64-NEXT: bswapq %rax
; X64-NEXT: bswapq %rcx
; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB46_6
+; X64-NEXT: jne .LBB46_4
; X64-NEXT: # %bb.2: # %loadbb2
; X64-NEXT: movq 16(%rdi), %rax
; X64-NEXT: movq 16(%rsi), %rcx
; X64-NEXT: bswapq %rax
; X64-NEXT: bswapq %rcx
; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB46_6
+; X64-NEXT: jne .LBB46_4
; X64-NEXT: # %bb.3: # %loadbb3
; X64-NEXT: movq 23(%rdi), %rax
; X64-NEXT: movq 23(%rsi), %rcx
; X64-NEXT: bswapq %rax
; X64-NEXT: bswapq %rcx
+; X64-NEXT: xorl %edx, %edx
; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: je .LBB46_4
-; X64-NEXT: .LBB46_6: # %res_block
+; X64-NEXT: je .LBB46_5
+; X64-NEXT: .LBB46_4: # %res_block
; X64-NEXT: xorl %edx, %edx
; X64-NEXT: cmpq %rcx, %rax
; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
+; X64-NEXT: leal -1(%rdx,%rdx), %edx
; X64-NEXT: .LBB46_5: # %endblock
-; X64-NEXT: testl %eax, %eax
+; X64-NEXT: testl %edx, %edx
; X64-NEXT: setg %al
; X64-NEXT: retq
-; X64-NEXT: .LBB46_4:
-; X64-NEXT: xorl %eax, %eax
-; X64-NEXT: jmp .LBB46_5
%call = tail call i32 @memcmp(i8* %x, i8* %y, i64 31) nounwind
%cmp = icmp sgt i32 %call, 0
ret i1 %cmp
@@ -2433,41 +2396,40 @@ define i32 @length32(i8* %X, i8* %Y) nounwind {
;
; X64-LABEL: length32:
; X64: # %bb.0:
-; X64-NEXT: movq (%rdi), %rax
-; X64-NEXT: movq (%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq (%rdi), %rcx
+; X64-NEXT: movq (%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB49_6
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: jne .LBB49_4
; X64-NEXT: # %bb.1: # %loadbb1
-; X64-NEXT: movq 8(%rdi), %rax
-; X64-NEXT: movq 8(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq 8(%rdi), %rcx
+; X64-NEXT: movq 8(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB49_6
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: jne .LBB49_4
; X64-NEXT: # %bb.2: # %loadbb2
-; X64-NEXT: movq 16(%rdi), %rax
-; X64-NEXT: movq 16(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq 16(%rdi), %rcx
+; X64-NEXT: movq 16(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB49_6
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: jne .LBB49_4
; X64-NEXT: # %bb.3: # %loadbb3
-; X64-NEXT: movq 24(%rdi), %rax
-; X64-NEXT: movq 24(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq 24(%rdi), %rcx
+; X64-NEXT: movq 24(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: je .LBB49_4
-; X64-NEXT: .LBB49_6: # %res_block
-; X64-NEXT: xorl %edx, %edx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB49_4:
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: je .LBB49_5
+; X64-NEXT: .LBB49_4: # %res_block
; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: setae %al
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: .LBB49_5: # %endblock
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 32) nounwind
ret i32 %m
@@ -2614,43 +2576,40 @@ define i1 @length32_lt(i8* %x, i8* %y) nounwind {
;
; X64-LABEL: length32_lt:
; X64: # %bb.0:
-; X64-NEXT: movq (%rdi), %rax
-; X64-NEXT: movq (%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq (%rdi), %rcx
+; X64-NEXT: movq (%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
; X64-NEXT: jne .LBB51_4
; X64-NEXT: # %bb.1: # %loadbb1
-; X64-NEXT: movq 8(%rdi), %rax
-; X64-NEXT: movq 8(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq 8(%rdi), %rcx
+; X64-NEXT: movq 8(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
; X64-NEXT: jne .LBB51_4
; X64-NEXT: # %bb.2: # %loadbb2
-; X64-NEXT: movq 16(%rdi), %rax
-; X64-NEXT: movq 16(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq 16(%rdi), %rcx
+; X64-NEXT: movq 16(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
; X64-NEXT: jne .LBB51_4
; X64-NEXT: # %bb.3: # %loadbb3
-; X64-NEXT: movq 24(%rdi), %rax
-; X64-NEXT: movq 24(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq 24(%rdi), %rcx
+; X64-NEXT: movq 24(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
; X64-NEXT: je .LBB51_5
; X64-NEXT: .LBB51_4: # %res_block
-; X64-NEXT: xorl %edx, %edx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
-; X64-NEXT: shrl $31, %eax
-; X64-NEXT: # kill: def $al killed $al killed $eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB51_5:
; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: setae %al
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: .LBB51_5: # %endblock
; X64-NEXT: shrl $31, %eax
; X64-NEXT: # kill: def $al killed $al killed $eax
; X64-NEXT: retq
@@ -2679,40 +2638,38 @@ define i1 @length32_gt(i8* %x, i8* %y) nounwind {
; X64-NEXT: bswapq %rax
; X64-NEXT: bswapq %rcx
; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB52_6
+; X64-NEXT: jne .LBB52_4
; X64-NEXT: # %bb.1: # %loadbb1
; X64-NEXT: movq 8(%rdi), %rax
; X64-NEXT: movq 8(%rsi), %rcx
; X64-NEXT: bswapq %rax
; X64-NEXT: bswapq %rcx
; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB52_6
+; X64-NEXT: jne .LBB52_4
; X64-NEXT: # %bb.2: # %loadbb2
; X64-NEXT: movq 16(%rdi), %rax
; X64-NEXT: movq 16(%rsi), %rcx
; X64-NEXT: bswapq %rax
; X64-NEXT: bswapq %rcx
; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB52_6
+; X64-NEXT: jne .LBB52_4
; X64-NEXT: # %bb.3: # %loadbb3
; X64-NEXT: movq 24(%rdi), %rax
; X64-NEXT: movq 24(%rsi), %rcx
; X64-NEXT: bswapq %rax
; X64-NEXT: bswapq %rcx
+; X64-NEXT: xorl %edx, %edx
; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: je .LBB52_4
-; X64-NEXT: .LBB52_6: # %res_block
+; X64-NEXT: je .LBB52_5
+; X64-NEXT: .LBB52_4: # %res_block
; X64-NEXT: xorl %edx, %edx
; X64-NEXT: cmpq %rcx, %rax
; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
+; X64-NEXT: leal -1(%rdx,%rdx), %edx
; X64-NEXT: .LBB52_5: # %endblock
-; X64-NEXT: testl %eax, %eax
+; X64-NEXT: testl %edx, %edx
; X64-NEXT: setg %al
; X64-NEXT: retq
-; X64-NEXT: .LBB52_4:
-; X64-NEXT: xorl %eax, %eax
-; X64-NEXT: jmp .LBB52_5
%call = tail call i32 @memcmp(i8* %x, i8* %y, i64 32) nounwind
%cmp = icmp sgt i32 %call, 0
ret i1 %cmp
diff --git a/llvm/test/CodeGen/X86/memcmp-optsize.ll b/llvm/test/CodeGen/X86/memcmp-optsize.ll
index 03e70e672097..594a4a68dac0 100644
--- a/llvm/test/CodeGen/X86/memcmp-optsize.ll
+++ b/llvm/test/CodeGen/X86/memcmp-optsize.ll
@@ -120,16 +120,16 @@ define i32 @length3(i8* %X, i8* %Y) nounwind optsize {
; X86-NEXT: rolw $8, %dx
; X86-NEXT: rolw $8, %si
; X86-NEXT: cmpw %si, %dx
-; X86-NEXT: je .LBB4_1
-; X86-NEXT: # %bb.3: # %res_block
-; X86-NEXT: setae %al
-; X86-NEXT: movzbl %al, %eax
-; X86-NEXT: leal -1(%eax,%eax), %eax
-; X86-NEXT: jmp .LBB4_2
-; X86-NEXT: .LBB4_1: # %loadbb1
+; X86-NEXT: jne .LBB4_3
+; X86-NEXT: # %bb.1: # %loadbb1
; X86-NEXT: movzbl 2(%eax), %eax
; X86-NEXT: movzbl 2(%ecx), %ecx
; X86-NEXT: subl %ecx, %eax
+; X86-NEXT: jmp .LBB4_2
+; X86-NEXT: .LBB4_3: # %res_block
+; X86-NEXT: setae %al
+; X86-NEXT: movzbl %al, %eax
+; X86-NEXT: leal -1(%eax,%eax), %eax
; X86-NEXT: .LBB4_2: # %endblock
; X86-NEXT: popl %esi
; X86-NEXT: retl
@@ -141,16 +141,16 @@ define i32 @length3(i8* %X, i8* %Y) nounwind optsize {
; X64-NEXT: rolw $8, %ax
; X64-NEXT: rolw $8, %cx
; X64-NEXT: cmpw %cx, %ax
-; X64-NEXT: je .LBB4_1
-; X64-NEXT: # %bb.3: # %res_block
-; X64-NEXT: setae %al
-; X64-NEXT: movzbl %al, %eax
-; X64-NEXT: leal -1(%rax,%rax), %eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB4_1: # %loadbb1
+; X64-NEXT: jne .LBB4_3
+; X64-NEXT: # %bb.1: # %loadbb1
; X64-NEXT: movzbl 2(%rdi), %eax
; X64-NEXT: movzbl 2(%rsi), %ecx
; X64-NEXT: subl %ecx, %eax
+; X64-NEXT: retq
+; X64-NEXT: .LBB4_3: # %res_block
+; X64-NEXT: setae %al
+; X64-NEXT: movzbl %al, %eax
+; X64-NEXT: leal -1(%rax,%rax), %eax
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 3) nounwind
ret i32 %m
@@ -265,16 +265,16 @@ define i32 @length5(i8* %X, i8* %Y) nounwind optsize {
; X86-NEXT: bswapl %edx
; X86-NEXT: bswapl %esi
; X86-NEXT: cmpl %esi, %edx
-; X86-NEXT: je .LBB9_1
-; X86-NEXT: # %bb.3: # %res_block
-; X86-NEXT: setae %al
-; X86-NEXT: movzbl %al, %eax
-; X86-NEXT: leal -1(%eax,%eax), %eax
-; X86-NEXT: jmp .LBB9_2
-; X86-NEXT: .LBB9_1: # %loadbb1
+; X86-NEXT: jne .LBB9_3
+; X86-NEXT: # %bb.1: # %loadbb1
; X86-NEXT: movzbl 4(%eax), %eax
; X86-NEXT: movzbl 4(%ecx), %ecx
; X86-NEXT: subl %ecx, %eax
+; X86-NEXT: jmp .LBB9_2
+; X86-NEXT: .LBB9_3: # %res_block
+; X86-NEXT: setae %al
+; X86-NEXT: movzbl %al, %eax
+; X86-NEXT: leal -1(%eax,%eax), %eax
; X86-NEXT: .LBB9_2: # %endblock
; X86-NEXT: popl %esi
; X86-NEXT: retl
@@ -286,16 +286,16 @@ define i32 @length5(i8* %X, i8* %Y) nounwind optsize {
; X64-NEXT: bswapl %eax
; X64-NEXT: bswapl %ecx
; X64-NEXT: cmpl %ecx, %eax
-; X64-NEXT: je .LBB9_1
-; X64-NEXT: # %bb.3: # %res_block
-; X64-NEXT: setae %al
-; X64-NEXT: movzbl %al, %eax
-; X64-NEXT: leal -1(%rax,%rax), %eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB9_1: # %loadbb1
+; X64-NEXT: jne .LBB9_3
+; X64-NEXT: # %bb.1: # %loadbb1
; X64-NEXT: movzbl 4(%rdi), %eax
; X64-NEXT: movzbl 4(%rsi), %ecx
; X64-NEXT: subl %ecx, %eax
+; X64-NEXT: retq
+; X64-NEXT: .LBB9_3: # %res_block
+; X64-NEXT: setae %al
+; X64-NEXT: movzbl %al, %eax
+; X64-NEXT: leal -1(%rax,%rax), %eax
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 5) nounwind
ret i32 %m
@@ -334,32 +334,30 @@ define i32 @length8(i8* %X, i8* %Y) nounwind optsize {
; X86-LABEL: length8:
; X86: # %bb.0:
; X86-NEXT: pushl %esi
-; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
+; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl {{[0-9]+}}(%esp), %esi
-; X86-NEXT: movl (%esi), %eax
-; X86-NEXT: movl (%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl (%esi), %ecx
+; X86-NEXT: movl (%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: jne .LBB11_4
+; X86-NEXT: bswapl %edx
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: jne .LBB11_2
; X86-NEXT: # %bb.1: # %loadbb1
-; X86-NEXT: movl 4(%esi), %eax
-; X86-NEXT: movl 4(%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl 4(%esi), %ecx
+; X86-NEXT: movl 4(%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: je .LBB11_2
-; X86-NEXT: .LBB11_4: # %res_block
-; X86-NEXT: xorl %edx, %edx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: setae %dl
-; X86-NEXT: leal -1(%edx,%edx), %eax
+; X86-NEXT: bswapl %edx
+; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: je .LBB11_3
+; X86-NEXT: .LBB11_2: # %res_block
+; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: setae %al
+; X86-NEXT: leal -1(%eax,%eax), %eax
; X86-NEXT: .LBB11_3: # %endblock
; X86-NEXT: popl %esi
; X86-NEXT: retl
-; X86-NEXT: .LBB11_2:
-; X86-NEXT: xorl %eax, %eax
-; X86-NEXT: jmp .LBB11_3
;
; X64-LABEL: length8:
; X64: # %bb.0:
@@ -463,27 +461,26 @@ define i32 @length12(i8* %X, i8* %Y) nounwind optsize {
;
; X64-LABEL: length12:
; X64: # %bb.0:
-; X64-NEXT: movq (%rdi), %rax
-; X64-NEXT: movq (%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq (%rdi), %rcx
+; X64-NEXT: movq (%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB15_4
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: jne .LBB15_2
; X64-NEXT: # %bb.1: # %loadbb1
-; X64-NEXT: movl 8(%rdi), %eax
-; X64-NEXT: movl 8(%rsi), %ecx
-; X64-NEXT: bswapl %eax
+; X64-NEXT: movl 8(%rdi), %ecx
+; X64-NEXT: movl 8(%rsi), %edx
; X64-NEXT: bswapl %ecx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: je .LBB15_2
-; X64-NEXT: .LBB15_4: # %res_block
-; X64-NEXT: xorl %edx, %edx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB15_2:
+; X64-NEXT: bswapl %edx
+; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: je .LBB15_3
+; X64-NEXT: .LBB15_2: # %res_block
; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: setae %al
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: .LBB15_3: # %endblock
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 12) nounwind
ret i32 %m
@@ -504,27 +501,26 @@ define i32 @length16(i8* %X, i8* %Y) nounwind optsize {
;
; X64-LABEL: length16:
; X64: # %bb.0:
-; X64-NEXT: movq (%rdi), %rax
-; X64-NEXT: movq (%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq (%rdi), %rcx
+; X64-NEXT: movq (%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB16_4
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: jne .LBB16_2
; X64-NEXT: # %bb.1: # %loadbb1
-; X64-NEXT: movq 8(%rdi), %rax
-; X64-NEXT: movq 8(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq 8(%rdi), %rcx
+; X64-NEXT: movq 8(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: je .LBB16_2
-; X64-NEXT: .LBB16_4: # %res_block
-; X64-NEXT: xorl %edx, %edx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB16_2:
+; X64-NEXT: bswapq %rdx
; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: je .LBB16_3
+; X64-NEXT: .LBB16_2: # %res_block
+; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: setae %al
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: .LBB16_3: # %endblock
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 16) nounwind
ret i32 %m
diff --git a/llvm/test/CodeGen/X86/memcmp-pgso.ll b/llvm/test/CodeGen/X86/memcmp-pgso.ll
index d9273a52c523..75e9f5975d95 100644
--- a/llvm/test/CodeGen/X86/memcmp-pgso.ll
+++ b/llvm/test/CodeGen/X86/memcmp-pgso.ll
@@ -120,19 +120,19 @@ define i32 @length3(i8* %X, i8* %Y) nounwind !prof !14 {
; X86-NEXT: rolw $8, %dx
; X86-NEXT: rolw $8, %si
; X86-NEXT: cmpw %si, %dx
-; X86-NEXT: je .LBB4_1
-; X86-NEXT: # %bb.3: # %res_block
+; X86-NEXT: jne .LBB4_3
+; X86-NEXT: # %bb.1: # %loadbb1
+; X86-NEXT: movzbl 2(%eax), %eax
+; X86-NEXT: movzbl 2(%ecx), %ecx
+; X86-NEXT: subl %ecx, %eax
+; X86-NEXT: jmp .LBB4_2
+; X86-NEXT: .LBB4_3: # %res_block
; X86-NEXT: setae %al
; X86-NEXT: movzbl %al, %eax
; X86-NEXT: leal -1(%eax,%eax), %eax
; X86-NEXT: .LBB4_2: # %endblock
; X86-NEXT: popl %esi
; X86-NEXT: retl
-; X86-NEXT: .LBB4_1: # %loadbb1
-; X86-NEXT: movzbl 2(%eax), %eax
-; X86-NEXT: movzbl 2(%ecx), %ecx
-; X86-NEXT: subl %ecx, %eax
-; X86-NEXT: jmp .LBB4_2
;
; X64-LABEL: length3:
; X64: # %bb.0:
@@ -141,16 +141,16 @@ define i32 @length3(i8* %X, i8* %Y) nounwind !prof !14 {
; X64-NEXT: rolw $8, %ax
; X64-NEXT: rolw $8, %cx
; X64-NEXT: cmpw %cx, %ax
-; X64-NEXT: je .LBB4_1
-; X64-NEXT: # %bb.3: # %res_block
-; X64-NEXT: setae %al
-; X64-NEXT: movzbl %al, %eax
-; X64-NEXT: leal -1(%rax,%rax), %eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB4_1: # %loadbb1
+; X64-NEXT: jne .LBB4_3
+; X64-NEXT: # %bb.1: # %loadbb1
; X64-NEXT: movzbl 2(%rdi), %eax
; X64-NEXT: movzbl 2(%rsi), %ecx
; X64-NEXT: subl %ecx, %eax
+; X64-NEXT: retq
+; X64-NEXT: .LBB4_3: # %res_block
+; X64-NEXT: setae %al
+; X64-NEXT: movzbl %al, %eax
+; X64-NEXT: leal -1(%rax,%rax), %eax
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 3) nounwind
ret i32 %m
@@ -265,19 +265,19 @@ define i32 @length5(i8* %X, i8* %Y) nounwind !prof !14 {
; X86-NEXT: bswapl %edx
; X86-NEXT: bswapl %esi
; X86-NEXT: cmpl %esi, %edx
-; X86-NEXT: je .LBB9_1
-; X86-NEXT: # %bb.3: # %res_block
+; X86-NEXT: jne .LBB9_3
+; X86-NEXT: # %bb.1: # %loadbb1
+; X86-NEXT: movzbl 4(%eax), %eax
+; X86-NEXT: movzbl 4(%ecx), %ecx
+; X86-NEXT: subl %ecx, %eax
+; X86-NEXT: jmp .LBB9_2
+; X86-NEXT: .LBB9_3: # %res_block
; X86-NEXT: setae %al
; X86-NEXT: movzbl %al, %eax
; X86-NEXT: leal -1(%eax,%eax), %eax
; X86-NEXT: .LBB9_2: # %endblock
; X86-NEXT: popl %esi
; X86-NEXT: retl
-; X86-NEXT: .LBB9_1: # %loadbb1
-; X86-NEXT: movzbl 4(%eax), %eax
-; X86-NEXT: movzbl 4(%ecx), %ecx
-; X86-NEXT: subl %ecx, %eax
-; X86-NEXT: jmp .LBB9_2
;
; X64-LABEL: length5:
; X64: # %bb.0:
@@ -286,16 +286,16 @@ define i32 @length5(i8* %X, i8* %Y) nounwind !prof !14 {
; X64-NEXT: bswapl %eax
; X64-NEXT: bswapl %ecx
; X64-NEXT: cmpl %ecx, %eax
-; X64-NEXT: je .LBB9_1
-; X64-NEXT: # %bb.3: # %res_block
-; X64-NEXT: setae %al
-; X64-NEXT: movzbl %al, %eax
-; X64-NEXT: leal -1(%rax,%rax), %eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB9_1: # %loadbb1
+; X64-NEXT: jne .LBB9_3
+; X64-NEXT: # %bb.1: # %loadbb1
; X64-NEXT: movzbl 4(%rdi), %eax
; X64-NEXT: movzbl 4(%rsi), %ecx
; X64-NEXT: subl %ecx, %eax
+; X64-NEXT: retq
+; X64-NEXT: .LBB9_3: # %res_block
+; X64-NEXT: setae %al
+; X64-NEXT: movzbl %al, %eax
+; X64-NEXT: leal -1(%rax,%rax), %eax
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 5) nounwind
ret i32 %m
@@ -334,32 +334,30 @@ define i32 @length8(i8* %X, i8* %Y) nounwind !prof !14 {
; X86-LABEL: length8:
; X86: # %bb.0:
; X86-NEXT: pushl %esi
-; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
+; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl {{[0-9]+}}(%esp), %esi
-; X86-NEXT: movl (%esi), %eax
-; X86-NEXT: movl (%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl (%esi), %ecx
+; X86-NEXT: movl (%eax), %edx
+; X86-NEXT: bswapl %ecx
+; X86-NEXT: bswapl %edx
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: jne .LBB11_2
+; X86-NEXT: # %bb.1: # %loadbb1
+; X86-NEXT: movl 4(%esi), %ecx
+; X86-NEXT: movl 4(%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: je .LBB11_1
-; X86-NEXT: .LBB11_4: # %res_block
-; X86-NEXT: xorl %edx, %edx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: setae %dl
-; X86-NEXT: leal -1(%edx,%edx), %eax
+; X86-NEXT: bswapl %edx
+; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: je .LBB11_3
+; X86-NEXT: .LBB11_2: # %res_block
+; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: setae %al
+; X86-NEXT: leal -1(%eax,%eax), %eax
; X86-NEXT: .LBB11_3: # %endblock
; X86-NEXT: popl %esi
; X86-NEXT: retl
-; X86-NEXT: .LBB11_1: # %loadbb1
-; X86-NEXT: movl 4(%esi), %eax
-; X86-NEXT: movl 4(%edx), %ecx
-; X86-NEXT: bswapl %eax
-; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: jne .LBB11_4
-; X86-NEXT: # %bb.2:
-; X86-NEXT: xorl %eax, %eax
-; X86-NEXT: jmp .LBB11_3
;
; X64-LABEL: length8:
; X64: # %bb.0:
@@ -463,27 +461,26 @@ define i32 @length12(i8* %X, i8* %Y) nounwind !prof !14 {
;
; X64-LABEL: length12:
; X64: # %bb.0:
-; X64-NEXT: movq (%rdi), %rax
-; X64-NEXT: movq (%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq (%rdi), %rcx
+; X64-NEXT: movq (%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: je .LBB15_1
-; X64-NEXT: .LBB15_4: # %res_block
-; X64-NEXT: xorl %edx, %edx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB15_1: # %loadbb1
-; X64-NEXT: movl 8(%rdi), %eax
-; X64-NEXT: movl 8(%rsi), %ecx
-; X64-NEXT: bswapl %eax
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: jne .LBB15_2
+; X64-NEXT: # %bb.1: # %loadbb1
+; X64-NEXT: movl 8(%rdi), %ecx
+; X64-NEXT: movl 8(%rsi), %edx
; X64-NEXT: bswapl %ecx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB15_4
-; X64-NEXT: # %bb.2:
+; X64-NEXT: bswapl %edx
+; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: je .LBB15_3
+; X64-NEXT: .LBB15_2: # %res_block
; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: setae %al
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: .LBB15_3: # %endblock
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 12) nounwind
ret i32 %m
@@ -504,27 +501,26 @@ define i32 @length16(i8* %X, i8* %Y) nounwind !prof !14 {
;
; X64-LABEL: length16:
; X64: # %bb.0:
-; X64-NEXT: movq (%rdi), %rax
-; X64-NEXT: movq (%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq (%rdi), %rcx
+; X64-NEXT: movq (%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: je .LBB16_1
-; X64-NEXT: .LBB16_4: # %res_block
-; X64-NEXT: xorl %edx, %edx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB16_1: # %loadbb1
-; X64-NEXT: movq 8(%rdi), %rax
-; X64-NEXT: movq 8(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: jne .LBB16_2
+; X64-NEXT: # %bb.1: # %loadbb1
+; X64-NEXT: movq 8(%rdi), %rcx
+; X64-NEXT: movq 8(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB16_4
-; X64-NEXT: # %bb.2:
+; X64-NEXT: bswapq %rdx
; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: je .LBB16_3
+; X64-NEXT: .LBB16_2: # %res_block
+; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: setae %al
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: .LBB16_3: # %endblock
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 16) nounwind
ret i32 %m
diff --git a/llvm/test/CodeGen/X86/memcmp.ll b/llvm/test/CodeGen/X86/memcmp.ll
index 233183b7f9ad..de604ded08d1 100644
--- a/llvm/test/CodeGen/X86/memcmp.ll
+++ b/llvm/test/CodeGen/X86/memcmp.ll
@@ -285,19 +285,19 @@ define i32 @length3(i8* %X, i8* %Y) nounwind {
; X86-NEXT: rolw $8, %dx
; X86-NEXT: rolw $8, %si
; X86-NEXT: cmpw %si, %dx
-; X86-NEXT: je .LBB11_1
-; X86-NEXT: # %bb.3: # %res_block
-; X86-NEXT: setae %al
-; X86-NEXT: movzbl %al, %eax
-; X86-NEXT: leal -1(%eax,%eax), %eax
-; X86-NEXT: popl %esi
-; X86-NEXT: retl
-; X86-NEXT: .LBB11_1: # %loadbb1
+; X86-NEXT: jne .LBB11_3
+; X86-NEXT: # %bb.1: # %loadbb1
; X86-NEXT: movzbl 2(%eax), %eax
; X86-NEXT: movzbl 2(%ecx), %ecx
; X86-NEXT: subl %ecx, %eax
; X86-NEXT: popl %esi
; X86-NEXT: retl
+; X86-NEXT: .LBB11_3: # %res_block
+; X86-NEXT: setae %al
+; X86-NEXT: movzbl %al, %eax
+; X86-NEXT: leal -1(%eax,%eax), %eax
+; X86-NEXT: popl %esi
+; X86-NEXT: retl
;
; X64-LABEL: length3:
; X64: # %bb.0:
@@ -306,16 +306,16 @@ define i32 @length3(i8* %X, i8* %Y) nounwind {
; X64-NEXT: rolw $8, %ax
; X64-NEXT: rolw $8, %cx
; X64-NEXT: cmpw %cx, %ax
-; X64-NEXT: je .LBB11_1
-; X64-NEXT: # %bb.3: # %res_block
-; X64-NEXT: setae %al
-; X64-NEXT: movzbl %al, %eax
-; X64-NEXT: leal -1(%rax,%rax), %eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB11_1: # %loadbb1
+; X64-NEXT: jne .LBB11_3
+; X64-NEXT: # %bb.1: # %loadbb1
; X64-NEXT: movzbl 2(%rdi), %eax
; X64-NEXT: movzbl 2(%rsi), %ecx
; X64-NEXT: subl %ecx, %eax
+; X64-NEXT: retq
+; X64-NEXT: .LBB11_3: # %res_block
+; X64-NEXT: setae %al
+; X64-NEXT: movzbl %al, %eax
+; X64-NEXT: leal -1(%rax,%rax), %eax
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 3) nounwind
ret i32 %m
@@ -500,19 +500,19 @@ define i32 @length5(i8* %X, i8* %Y) nounwind {
; X86-NEXT: bswapl %edx
; X86-NEXT: bswapl %esi
; X86-NEXT: cmpl %esi, %edx
-; X86-NEXT: je .LBB18_1
-; X86-NEXT: # %bb.3: # %res_block
-; X86-NEXT: setae %al
-; X86-NEXT: movzbl %al, %eax
-; X86-NEXT: leal -1(%eax,%eax), %eax
-; X86-NEXT: popl %esi
-; X86-NEXT: retl
-; X86-NEXT: .LBB18_1: # %loadbb1
+; X86-NEXT: jne .LBB18_3
+; X86-NEXT: # %bb.1: # %loadbb1
; X86-NEXT: movzbl 4(%eax), %eax
; X86-NEXT: movzbl 4(%ecx), %ecx
; X86-NEXT: subl %ecx, %eax
; X86-NEXT: popl %esi
; X86-NEXT: retl
+; X86-NEXT: .LBB18_3: # %res_block
+; X86-NEXT: setae %al
+; X86-NEXT: movzbl %al, %eax
+; X86-NEXT: leal -1(%eax,%eax), %eax
+; X86-NEXT: popl %esi
+; X86-NEXT: retl
;
; X64-LABEL: length5:
; X64: # %bb.0:
@@ -521,16 +521,16 @@ define i32 @length5(i8* %X, i8* %Y) nounwind {
; X64-NEXT: bswapl %eax
; X64-NEXT: bswapl %ecx
; X64-NEXT: cmpl %ecx, %eax
-; X64-NEXT: je .LBB18_1
-; X64-NEXT: # %bb.3: # %res_block
-; X64-NEXT: setae %al
-; X64-NEXT: movzbl %al, %eax
-; X64-NEXT: leal -1(%rax,%rax), %eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB18_1: # %loadbb1
+; X64-NEXT: jne .LBB18_3
+; X64-NEXT: # %bb.1: # %loadbb1
; X64-NEXT: movzbl 4(%rdi), %eax
; X64-NEXT: movzbl 4(%rsi), %ecx
; X64-NEXT: subl %ecx, %eax
+; X64-NEXT: retq
+; X64-NEXT: .LBB18_3: # %res_block
+; X64-NEXT: setae %al
+; X64-NEXT: movzbl %al, %eax
+; X64-NEXT: leal -1(%rax,%rax), %eax
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 5) nounwind
ret i32 %m
@@ -576,16 +576,16 @@ define i1 @length5_lt(i8* %X, i8* %Y) nounwind {
; X86-NEXT: bswapl %edx
; X86-NEXT: bswapl %esi
; X86-NEXT: cmpl %esi, %edx
-; X86-NEXT: je .LBB20_1
-; X86-NEXT: # %bb.3: # %res_block
-; X86-NEXT: setae %al
-; X86-NEXT: movzbl %al, %eax
-; X86-NEXT: leal -1(%eax,%eax), %eax
-; X86-NEXT: jmp .LBB20_2
-; X86-NEXT: .LBB20_1: # %loadbb1
+; X86-NEXT: jne .LBB20_3
+; X86-NEXT: # %bb.1: # %loadbb1
; X86-NEXT: movzbl 4(%eax), %eax
; X86-NEXT: movzbl 4(%ecx), %ecx
; X86-NEXT: subl %ecx, %eax
+; X86-NEXT: jmp .LBB20_2
+; X86-NEXT: .LBB20_3: # %res_block
+; X86-NEXT: setae %al
+; X86-NEXT: movzbl %al, %eax
+; X86-NEXT: leal -1(%eax,%eax), %eax
; X86-NEXT: .LBB20_2: # %endblock
; X86-NEXT: shrl $31, %eax
; X86-NEXT: # kill: def $al killed $al killed $eax
@@ -599,20 +599,20 @@ define i1 @length5_lt(i8* %X, i8* %Y) nounwind {
; X64-NEXT: bswapl %eax
; X64-NEXT: bswapl %ecx
; X64-NEXT: cmpl %ecx, %eax
-; X64-NEXT: je .LBB20_1
-; X64-NEXT: # %bb.3: # %res_block
-; X64-NEXT: setae %al
-; X64-NEXT: movzbl %al, %eax
-; X64-NEXT: leal -1(%rax,%rax), %eax
-; X64-NEXT: shrl $31, %eax
-; X64-NEXT: # kill: def $al killed $al killed $eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB20_1: # %loadbb1
+; X64-NEXT: jne .LBB20_3
+; X64-NEXT: # %bb.1: # %loadbb1
; X64-NEXT: movzbl 4(%rdi), %eax
; X64-NEXT: movzbl 4(%rsi), %ecx
; X64-NEXT: subl %ecx, %eax
; X64-NEXT: shrl $31, %eax
; X64-NEXT: # kill: def $al killed $al killed $eax
+; X64-NEXT: retq
+; X64-NEXT: .LBB20_3: # %res_block
+; X64-NEXT: setae %al
+; X64-NEXT: movzbl %al, %eax
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: shrl $31, %eax
+; X64-NEXT: # kill: def $al killed $al killed $eax
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 5) nounwind
%c = icmp slt i32 %m, 0
@@ -623,56 +623,53 @@ define i32 @length7(i8* %X, i8* %Y) nounwind {
; X86-LABEL: length7:
; X86: # %bb.0:
; X86-NEXT: pushl %esi
-; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
+; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl {{[0-9]+}}(%esp), %esi
-; X86-NEXT: movl (%esi), %eax
-; X86-NEXT: movl (%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl (%esi), %ecx
+; X86-NEXT: movl (%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: jne .LBB21_4
+; X86-NEXT: bswapl %edx
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: jne .LBB21_2
; X86-NEXT: # %bb.1: # %loadbb1
-; X86-NEXT: movl 3(%esi), %eax
-; X86-NEXT: movl 3(%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl 3(%esi), %ecx
+; X86-NEXT: movl 3(%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: je .LBB21_2
-; X86-NEXT: .LBB21_4: # %res_block
-; X86-NEXT: xorl %edx, %edx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: setae %dl
-; X86-NEXT: leal -1(%edx,%edx), %eax
-; X86-NEXT: popl %esi
-; X86-NEXT: retl
-; X86-NEXT: .LBB21_2:
+; X86-NEXT: bswapl %edx
; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: je .LBB21_3
+; X86-NEXT: .LBB21_2: # %res_block
+; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: setae %al
+; X86-NEXT: leal -1(%eax,%eax), %eax
+; X86-NEXT: .LBB21_3: # %endblock
; X86-NEXT: popl %esi
; X86-NEXT: retl
;
; X64-LABEL: length7:
; X64: # %bb.0:
-; X64-NEXT: movl (%rdi), %eax
-; X64-NEXT: movl (%rsi), %ecx
-; X64-NEXT: bswapl %eax
+; X64-NEXT: movl (%rdi), %ecx
+; X64-NEXT: movl (%rsi), %edx
; X64-NEXT: bswapl %ecx
-; X64-NEXT: cmpl %ecx, %eax
-; X64-NEXT: jne .LBB21_4
+; X64-NEXT: bswapl %edx
+; X64-NEXT: cmpl %edx, %ecx
+; X64-NEXT: jne .LBB21_2
; X64-NEXT: # %bb.1: # %loadbb1
-; X64-NEXT: movl 3(%rdi), %eax
-; X64-NEXT: movl 3(%rsi), %ecx
-; X64-NEXT: bswapl %eax
+; X64-NEXT: movl 3(%rdi), %ecx
+; X64-NEXT: movl 3(%rsi), %edx
; X64-NEXT: bswapl %ecx
-; X64-NEXT: cmpl %ecx, %eax
-; X64-NEXT: je .LBB21_2
-; X64-NEXT: .LBB21_4: # %res_block
-; X64-NEXT: xorl %edx, %edx
-; X64-NEXT: cmpl %ecx, %eax
-; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB21_2:
+; X64-NEXT: bswapl %edx
; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpl %edx, %ecx
+; X64-NEXT: je .LBB21_3
+; X64-NEXT: .LBB21_2: # %res_block
+; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpl %edx, %ecx
+; X64-NEXT: setae %al
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: .LBB21_3: # %endblock
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 7) nounwind
ret i32 %m
@@ -682,60 +679,55 @@ define i1 @length7_lt(i8* %X, i8* %Y) nounwind {
; X86-LABEL: length7_lt:
; X86: # %bb.0:
; X86-NEXT: pushl %esi
-; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
+; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl {{[0-9]+}}(%esp), %esi
-; X86-NEXT: movl (%esi), %eax
-; X86-NEXT: movl (%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl (%esi), %ecx
+; X86-NEXT: movl (%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: jne .LBB22_4
+; X86-NEXT: bswapl %edx
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: jne .LBB22_2
; X86-NEXT: # %bb.1: # %loadbb1
-; X86-NEXT: movl 3(%esi), %eax
-; X86-NEXT: movl 3(%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl 3(%esi), %ecx
+; X86-NEXT: movl 3(%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: je .LBB22_2
-; X86-NEXT: .LBB22_4: # %res_block
-; X86-NEXT: xorl %edx, %edx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: setae %dl
-; X86-NEXT: leal -1(%edx,%edx), %eax
+; X86-NEXT: bswapl %edx
+; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: je .LBB22_3
+; X86-NEXT: .LBB22_2: # %res_block
+; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: setae %al
+; X86-NEXT: leal -1(%eax,%eax), %eax
; X86-NEXT: .LBB22_3: # %endblock
; X86-NEXT: shrl $31, %eax
; X86-NEXT: # kill: def $al killed $al killed $eax
; X86-NEXT: popl %esi
; X86-NEXT: retl
-; X86-NEXT: .LBB22_2:
-; X86-NEXT: xorl %eax, %eax
-; X86-NEXT: jmp .LBB22_3
;
; X64-LABEL: length7_lt:
; X64: # %bb.0:
-; X64-NEXT: movl (%rdi), %eax
-; X64-NEXT: movl (%rsi), %ecx
-; X64-NEXT: bswapl %eax
+; X64-NEXT: movl (%rdi), %ecx
+; X64-NEXT: movl (%rsi), %edx
; X64-NEXT: bswapl %ecx
-; X64-NEXT: cmpl %ecx, %eax
+; X64-NEXT: bswapl %edx
+; X64-NEXT: cmpl %edx, %ecx
; X64-NEXT: jne .LBB22_2
; X64-NEXT: # %bb.1: # %loadbb1
-; X64-NEXT: movl 3(%rdi), %eax
-; X64-NEXT: movl 3(%rsi), %ecx
-; X64-NEXT: bswapl %eax
+; X64-NEXT: movl 3(%rdi), %ecx
+; X64-NEXT: movl 3(%rsi), %edx
; X64-NEXT: bswapl %ecx
-; X64-NEXT: cmpl %ecx, %eax
+; X64-NEXT: bswapl %edx
+; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpl %edx, %ecx
; X64-NEXT: je .LBB22_3
; X64-NEXT: .LBB22_2: # %res_block
-; X64-NEXT: xorl %edx, %edx
-; X64-NEXT: cmpl %ecx, %eax
-; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
-; X64-NEXT: shrl $31, %eax
-; X64-NEXT: # kill: def $al killed $al killed $eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB22_3:
; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpl %edx, %ecx
+; X64-NEXT: setae %al
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: .LBB22_3: # %endblock
; X64-NEXT: shrl $31, %eax
; X64-NEXT: # kill: def $al killed $al killed $eax
; X64-NEXT: retq
@@ -775,30 +767,28 @@ define i32 @length8(i8* %X, i8* %Y) nounwind {
; X86-LABEL: length8:
; X86: # %bb.0:
; X86-NEXT: pushl %esi
-; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
+; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl {{[0-9]+}}(%esp), %esi
-; X86-NEXT: movl (%esi), %eax
-; X86-NEXT: movl (%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl (%esi), %ecx
+; X86-NEXT: movl (%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: jne .LBB24_4
+; X86-NEXT: bswapl %edx
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: jne .LBB24_2
; X86-NEXT: # %bb.1: # %loadbb1
-; X86-NEXT: movl 4(%esi), %eax
-; X86-NEXT: movl 4(%edx), %ecx
-; X86-NEXT: bswapl %eax
+; X86-NEXT: movl 4(%esi), %ecx
+; X86-NEXT: movl 4(%eax), %edx
; X86-NEXT: bswapl %ecx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: je .LBB24_2
-; X86-NEXT: .LBB24_4: # %res_block
-; X86-NEXT: xorl %edx, %edx
-; X86-NEXT: cmpl %ecx, %eax
-; X86-NEXT: setae %dl
-; X86-NEXT: leal -1(%edx,%edx), %eax
-; X86-NEXT: popl %esi
-; X86-NEXT: retl
-; X86-NEXT: .LBB24_2:
+; X86-NEXT: bswapl %edx
+; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: je .LBB24_3
+; X86-NEXT: .LBB24_2: # %res_block
; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: cmpl %edx, %ecx
+; X86-NEXT: setae %al
+; X86-NEXT: leal -1(%eax,%eax), %eax
+; X86-NEXT: .LBB24_3: # %endblock
; X86-NEXT: popl %esi
; X86-NEXT: retl
;
@@ -987,27 +977,26 @@ define i32 @length12(i8* %X, i8* %Y) nounwind {
;
; X64-LABEL: length12:
; X64: # %bb.0:
-; X64-NEXT: movq (%rdi), %rax
-; X64-NEXT: movq (%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq (%rdi), %rcx
+; X64-NEXT: movq (%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB31_4
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: jne .LBB31_2
; X64-NEXT: # %bb.1: # %loadbb1
-; X64-NEXT: movl 8(%rdi), %eax
-; X64-NEXT: movl 8(%rsi), %ecx
-; X64-NEXT: bswapl %eax
+; X64-NEXT: movl 8(%rdi), %ecx
+; X64-NEXT: movl 8(%rsi), %edx
; X64-NEXT: bswapl %ecx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: je .LBB31_2
-; X64-NEXT: .LBB31_4: # %res_block
-; X64-NEXT: xorl %edx, %edx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB31_2:
+; X64-NEXT: bswapl %edx
; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: je .LBB31_3
+; X64-NEXT: .LBB31_2: # %res_block
+; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: setae %al
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: .LBB31_3: # %endblock
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 12) nounwind
ret i32 %m
@@ -1080,27 +1069,26 @@ define i32 @length15(i8* %X, i8* %Y) nounwind {
;
; X64-LABEL: length15:
; X64: # %bb.0:
-; X64-NEXT: movq (%rdi), %rax
-; X64-NEXT: movq (%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq (%rdi), %rcx
+; X64-NEXT: movq (%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB34_4
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: jne .LBB34_2
; X64-NEXT: # %bb.1: # %loadbb1
-; X64-NEXT: movq 7(%rdi), %rax
-; X64-NEXT: movq 7(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq 7(%rdi), %rcx
+; X64-NEXT: movq 7(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: je .LBB34_2
-; X64-NEXT: .LBB34_4: # %res_block
-; X64-NEXT: xorl %edx, %edx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB34_2:
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: je .LBB34_3
+; X64-NEXT: .LBB34_2: # %res_block
; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: setae %al
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: .LBB34_3: # %endblock
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 15) nounwind
ret i32 %m
@@ -1121,29 +1109,26 @@ define i1 @length15_lt(i8* %X, i8* %Y) nounwind {
;
; X64-LABEL: length15_lt:
; X64: # %bb.0:
-; X64-NEXT: movq (%rdi), %rax
-; X64-NEXT: movq (%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq (%rdi), %rcx
+; X64-NEXT: movq (%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
; X64-NEXT: jne .LBB35_2
; X64-NEXT: # %bb.1: # %loadbb1
-; X64-NEXT: movq 7(%rdi), %rax
-; X64-NEXT: movq 7(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq 7(%rdi), %rcx
+; X64-NEXT: movq 7(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
; X64-NEXT: je .LBB35_3
; X64-NEXT: .LBB35_2: # %res_block
-; X64-NEXT: xorl %edx, %edx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
-; X64-NEXT: shrl $31, %eax
-; X64-NEXT: # kill: def $al killed $al killed $eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB35_3:
; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: setae %al
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: .LBB35_3: # %endblock
; X64-NEXT: shrl $31, %eax
; X64-NEXT: # kill: def $al killed $al killed $eax
; X64-NEXT: retq
@@ -1271,27 +1256,26 @@ define i32 @length16(i8* %X, i8* %Y) nounwind {
;
; X64-LABEL: length16:
; X64: # %bb.0:
-; X64-NEXT: movq (%rdi), %rax
-; X64-NEXT: movq (%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq (%rdi), %rcx
+; X64-NEXT: movq (%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB39_4
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: jne .LBB39_2
; X64-NEXT: # %bb.1: # %loadbb1
-; X64-NEXT: movq 8(%rdi), %rax
-; X64-NEXT: movq 8(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq 8(%rdi), %rcx
+; X64-NEXT: movq 8(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: je .LBB39_2
-; X64-NEXT: .LBB39_4: # %res_block
-; X64-NEXT: xorl %edx, %edx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB39_2:
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: je .LBB39_3
+; X64-NEXT: .LBB39_2: # %res_block
; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: setae %al
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: .LBB39_3: # %endblock
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 16) nounwind
ret i32 %m
@@ -1401,29 +1385,26 @@ define i1 @length16_lt(i8* %x, i8* %y) nounwind {
;
; X64-LABEL: length16_lt:
; X64: # %bb.0:
-; X64-NEXT: movq (%rdi), %rax
-; X64-NEXT: movq (%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq (%rdi), %rcx
+; X64-NEXT: movq (%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: cmpq %rdx, %rcx
; X64-NEXT: jne .LBB41_2
; X64-NEXT: # %bb.1: # %loadbb1
-; X64-NEXT: movq 8(%rdi), %rax
-; X64-NEXT: movq 8(%rsi), %rcx
-; X64-NEXT: bswapq %rax
+; X64-NEXT: movq 8(%rdi), %rcx
+; X64-NEXT: movq 8(%rsi), %rdx
; X64-NEXT: bswapq %rcx
-; X64-NEXT: cmpq %rcx, %rax
+; X64-NEXT: bswapq %rdx
+; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
; X64-NEXT: je .LBB41_3
; X64-NEXT: .LBB41_2: # %res_block
-; X64-NEXT: xorl %edx, %edx
-; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
-; X64-NEXT: shrl $31, %eax
-; X64-NEXT: # kill: def $al killed $al killed $eax
-; X64-NEXT: retq
-; X64-NEXT: .LBB41_3:
; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: cmpq %rdx, %rcx
+; X64-NEXT: setae %al
+; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: .LBB41_3: # %endblock
; X64-NEXT: shrl $31, %eax
; X64-NEXT: # kill: def $al killed $al killed $eax
; X64-NEXT: retq
@@ -1452,26 +1433,24 @@ define i1 @length16_gt(i8* %x, i8* %y) nounwind {
; X64-NEXT: bswapq %rax
; X64-NEXT: bswapq %rcx
; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: jne .LBB42_4
+; X64-NEXT: jne .LBB42_2
; X64-NEXT: # %bb.1: # %loadbb1
; X64-NEXT: movq 8(%rdi), %rax
; X64-NEXT: movq 8(%rsi), %rcx
; X64-NEXT: bswapq %rax
; X64-NEXT: bswapq %rcx
+; X64-NEXT: xorl %edx, %edx
; X64-NEXT: cmpq %rcx, %rax
-; X64-NEXT: je .LBB42_2
-; X64-NEXT: .LBB42_4: # %res_block
+; X64-NEXT: je .LBB42_3
+; X64-NEXT: .LBB42_2: # %res_block
; X64-NEXT: xorl %edx, %edx
; X64-NEXT: cmpq %rcx, %rax
; X64-NEXT: setae %dl
-; X64-NEXT: leal -1(%rdx,%rdx), %eax
+; X64-NEXT: leal -1(%rdx,%rdx), %edx
; X64-NEXT: .LBB42_3: # %endblock
-; X64-NEXT: testl %eax, %eax
+; X64-NEXT: testl %edx, %edx
; X64-NEXT: setg %al
; X64-NEXT: retq
-; X64-NEXT: .LBB42_2:
-; X64-NEXT: xorl %eax, %eax
-; X64-NEXT: jmp .LBB42_3
%call = tail call i32 @memcmp(i8* %x, i8* %y, i64 16) nounwind
%cmp = icmp sgt i32 %call, 0
ret i1 %cmp
diff --git a/llvm/test/CodeGen/X86/neg_cmp.ll b/llvm/test/CodeGen/X86/neg_cmp.ll
index 8cc1ed7ff123..47fa7fbb88f0 100644
--- a/llvm/test/CodeGen/X86/neg_cmp.ll
+++ b/llvm/test/CodeGen/X86/neg_cmp.ll
@@ -10,11 +10,11 @@ define void @neg_cmp(i32 %x, i32 %y) nounwind {
; CHECK-LABEL: neg_cmp:
; CHECK: # %bb.0:
; CHECK-NEXT: addl %esi, %edi
-; CHECK-NEXT: je .LBB0_2
-; CHECK-NEXT: # %bb.1: # %if.end
-; CHECK-NEXT: retq
-; CHECK-NEXT: .LBB0_2: # %if.then
+; CHECK-NEXT: jne .LBB0_1
+; CHECK-NEXT: # %bb.2: # %if.then
; CHECK-NEXT: jmp g # TAILCALL
+; CHECK-NEXT: .LBB0_1: # %if.end
+; CHECK-NEXT: retq
%sub = sub i32 0, %y
%cmp = icmp eq i32 %x, %sub
br i1 %cmp, label %if.then, label %if.end
@@ -31,11 +31,11 @@ define void @neg_cmp_commuted(i32 %x, i32 %y) nounwind {
; CHECK-LABEL: neg_cmp_commuted:
; CHECK: # %bb.0:
; CHECK-NEXT: addl %esi, %edi
-; CHECK-NEXT: je .LBB1_2
-; CHECK-NEXT: # %bb.1: # %if.end
-; CHECK-NEXT: retq
-; CHECK-NEXT: .LBB1_2: # %if.then
+; CHECK-NEXT: jne .LBB1_1
+; CHECK-NEXT: # %bb.2: # %if.then
; CHECK-NEXT: jmp g # TAILCALL
+; CHECK-NEXT: .LBB1_1: # %if.end
+; CHECK-NEXT: retq
%sub = sub i32 0, %y
%cmp = icmp eq i32 %sub, %x
br i1 %cmp, label %if.then, label %if.end
diff --git a/llvm/test/CodeGen/X86/nobt.ll b/llvm/test/CodeGen/X86/nobt.ll
index 5e3926dd07a8..b994b7f950fd 100644
--- a/llvm/test/CodeGen/X86/nobt.ll
+++ b/llvm/test/CodeGen/X86/nobt.ll
@@ -9,11 +9,10 @@ define void @test2(i32 %x, i32 %n) nounwind {
; CHECK: # %bb.0: # %entry
; CHECK-NEXT: xorl %eax, %eax
; CHECK-NEXT: testb %al, %al
-; CHECK-NEXT: je .LBB0_1
-; CHECK-NEXT: # %bb.2: # %UnifiedReturnBlock
-; CHECK-NEXT: retl
-; CHECK-NEXT: .LBB0_1: # %bb
+; CHECK-NEXT: jne .LBB0_2
+; CHECK-NEXT: # %bb.1: # %bb
; CHECK-NEXT: calll foo
+; CHECK-NEXT: .LBB0_2: # %UnifiedReturnBlock
; CHECK-NEXT: retl
entry:
%tmp1 = and i32 %x, 1
@@ -35,11 +34,10 @@ define void @test3(i32 %x, i32 %n) nounwind {
; CHECK: # %bb.0: # %entry
; CHECK-NEXT: xorl %eax, %eax
; CHECK-NEXT: testb %al, %al
-; CHECK-NEXT: je .LBB1_1
-; CHECK-NEXT: # %bb.2: # %UnifiedReturnBlock
-; CHECK-NEXT: retl
-; CHECK-NEXT: .LBB1_1: # %bb
+; CHECK-NEXT: jne .LBB1_2
+; CHECK-NEXT: # %bb.1: # %bb
; CHECK-NEXT: calll foo
+; CHECK-NEXT: .LBB1_2: # %UnifiedReturnBlock
; CHECK-NEXT: retl
entry:
%tmp1 = and i32 %x, 1
diff --git a/llvm/test/CodeGen/X86/pr29170.ll b/llvm/test/CodeGen/X86/pr29170.ll
index d5cfc51bf416..dfbad021d287 100644
--- a/llvm/test/CodeGen/X86/pr29170.ll
+++ b/llvm/test/CodeGen/X86/pr29170.ll
@@ -11,11 +11,8 @@ define i32 @main() {
; CHECK: # %bb.0: # %entry
; CHECK-NEXT: xorl %eax, %eax
; CHECK-NEXT: testb %al, %al
-; CHECK-NEXT: je .LBB0_1
-; CHECK-NEXT: .LBB0_3: # %if.else
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: retl
-; CHECK-NEXT: .LBB0_1: # %go
+; CHECK-NEXT: jne .LBB0_3
+; CHECK-NEXT: # %bb.1: # %go
; CHECK-NEXT: movl $-1, %ecx
; CHECK-NEXT: movsbl b, %edx
; CHECK-NEXT: notl %ecx
@@ -26,6 +23,9 @@ define i32 @main() {
; CHECK-NEXT: # %bb.2: # %if.then
; CHECK-NEXT: movl $42, %eax
; CHECK-NEXT: retl
+; CHECK-NEXT: .LBB0_3: # %if.else
+; CHECK-NEXT: xorl %eax, %eax
+; CHECK-NEXT: retl
entry:
%true = icmp eq i32 0, 0
%const = bitcast i64 -4294967296 to i64
diff --git a/llvm/test/CodeGen/X86/wide-integer-cmp.ll b/llvm/test/CodeGen/X86/wide-integer-cmp.ll
index cc4857a16c58..a15d633d8538 100644
--- a/llvm/test/CodeGen/X86/wide-integer-cmp.ll
+++ b/llvm/test/CodeGen/X86/wide-integer-cmp.ll
@@ -9,13 +9,13 @@ define i32 @branch_eq(i64 %a, i64 %b) {
; CHECK-NEXT: xorl {{[0-9]+}}(%esp), %ecx
; CHECK-NEXT: xorl {{[0-9]+}}(%esp), %eax
; CHECK-NEXT: orl %ecx, %eax
-; CHECK-NEXT: je .LBB0_1
-; CHECK-NEXT: # %bb.2: # %bb2
-; CHECK-NEXT: movl $2, %eax
-; CHECK-NEXT: retl
-; CHECK-NEXT: .LBB0_1: # %bb1
+; CHECK-NEXT: jne .LBB0_2
+; CHECK-NEXT: # %bb.1: # %bb1
; CHECK-NEXT: movl $1, %eax
; CHECK-NEXT: retl
+; CHECK-NEXT: .LBB0_2: # %bb2
+; CHECK-NEXT: movl $2, %eax
+; CHECK-NEXT: retl
entry:
%cmp = icmp eq i64 %a, %b
br i1 %cmp, label %bb1, label %bb2
@@ -55,13 +55,13 @@ define i32 @branch_ule(i64 %a, i64 %b) {
; CHECK-NEXT: movl {{[0-9]+}}(%esp), %ecx
; CHECK-NEXT: cmpl {{[0-9]+}}(%esp), %eax
; CHECK-NEXT: sbbl {{[0-9]+}}(%esp), %ecx
-; CHECK-NEXT: jae .LBB2_1
-; CHECK-NEXT: # %bb.2: # %bb2
-; CHECK-NEXT: movl $2, %eax
-; CHECK-NEXT: retl
-; CHECK-NEXT: .LBB2_1: # %bb1
+; CHECK-NEXT: jb .LBB2_2
+; CHECK-NEXT: # %bb.1: # %bb1
; CHECK-NEXT: movl $1, %eax
; CHECK-NEXT: retl
+; CHECK-NEXT: .LBB2_2: # %bb2
+; CHECK-NEXT: movl $2, %eax
+; CHECK-NEXT: retl
entry:
%cmp = icmp ule i64 %a, %b
br i1 %cmp, label %bb1, label %bb2
diff --git a/llvm/test/DebugInfo/COFF/fpo-shrink-wrap.ll b/llvm/test/DebugInfo/COFF/fpo-shrink-wrap.ll
index ea89c76c4d93..719d69d16a62 100644
--- a/llvm/test/DebugInfo/COFF/fpo-shrink-wrap.ll
+++ b/llvm/test/DebugInfo/COFF/fpo-shrink-wrap.ll
@@ -16,11 +16,8 @@
; ASM: .cv_loc 0 1 3 9 # t.c:3:9
; ASM: movl %ecx, %eax
; ASM: cmpl %edx, %ecx
-; ASM: jge LBB0_1
+; ASM: jl [[EPILOGUE:LBB0_[0-9]+]]
-; ASM: retl $8
-
-; ASM: LBB0_1:
; ASM: pushl %ebx
; ASM: .cv_fpo_pushreg %ebx
; ASM: pushl %edi
@@ -34,7 +31,9 @@
; ASM: popl %esi
; ASM: popl %edi
; ASM: popl %ebx
-; ASM: Ltmp11:
+; ASM: [[EPILOGUE]]: # %return
+; ASM: retl $8
+; ASM: Ltmp10:
; ASM: .cv_fpo_endproc
; Note how RvaStart advances 7 bytes to skip the shrink-wrapped portion.
@@ -42,7 +41,7 @@
; OBJ: FrameData {
; OBJ: RvaStart: 0x0
; OBJ: CodeSize: 0x36
-; OBJ: PrologSize: 0xC
+; OBJ: PrologSize: 0x9
; OBJ: FrameFunc [
; OBJ-NEXT: $T0 .raSearch =
; OBJ-NEXT: $eip $T0 ^ =
@@ -50,8 +49,8 @@
; OBJ-NEXT: ]
; OBJ: }
; OBJ: FrameData {
-; OBJ: RvaStart: 0xA
-; OBJ: CodeSize: 0x2C
+; OBJ: RvaStart: 0x7
+; OBJ: CodeSize: 0x2F
; OBJ: PrologSize: 0x2
; OBJ: FrameFunc [
; OBJ-NEXT: $T0 .raSearch =
@@ -61,8 +60,8 @@
; OBJ-NEXT: ]
; OBJ: }
; OBJ: FrameData {
-; OBJ: RvaStart: 0xB
-; OBJ: CodeSize: 0x2B
+; OBJ: RvaStart: 0x8
+; OBJ: CodeSize: 0x2E
; OBJ: PrologSize: 0x1
; OBJ: FrameFunc [
; OBJ-NEXT: $T0 .raSearch =
@@ -73,8 +72,8 @@
; OBJ-NEXT: ]
; OBJ: }
; OBJ: FrameData {
-; OBJ: RvaStart: 0xC
-; OBJ: CodeSize: 0x2A
+; OBJ: RvaStart: 0x9
+; OBJ: CodeSize: 0x2D
; OBJ: PrologSize: 0x0
; OBJ: FrameFunc [
; OBJ-NEXT: $T0 .raSearch =
diff --git a/llvm/test/Transforms/PGOProfile/counter_promo_mexits.ll b/llvm/test/Transforms/PGOProfile/counter_promo_mexits.ll
index f462eba5a3ab..bb799757a47c 100644
--- a/llvm/test/Transforms/PGOProfile/counter_promo_mexits.ll
+++ b/llvm/test/Transforms/PGOProfile/counter_promo_mexits.ll
@@ -1,95 +1,27 @@
-; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
-; RUN: opt < %s -pgo-instr-gen -instrprof -do-counter-promotion=true -speculative-counter-promotion-max-exiting=3 -S | FileCheck %s --check-prefix=PROMO
-; RUN: opt < %s --passes=pgo-instr-gen,instrprof -do-counter-promotion=true -speculative-counter-promotion-max-exiting=3 -S | FileCheck %s --check-prefix=PROMO
+; RUN: opt < %s -pgo-instr-gen -instrprof -do-counter-promotion=true -speculative-counter-promotion-max-exiting=3 -S | FileCheck --check-prefix=PROMO %s
+; RUN: opt < %s --passes=pgo-instr-gen,instrprof -do-counter-promotion=true -speculative-counter-promotion-max-exiting=3 -S | FileCheck --check-prefix=PROMO %s
@g = common local_unnamed_addr global i32 0, align 4
define void @foo(i32 %arg) local_unnamed_addr {
-; PROMO-LABEL: @foo(
-; PROMO-NEXT: bb:
-; PROMO-NEXT: [[T:%.*]] = add nsw i32 [[ARG:%.*]], -1
-; PROMO-NEXT: br label [[BB1:%.*]]
-; PROMO: bb1:
-; PROMO-NEXT: [[PGOCOUNT213:%.*]] = phi i64 [ 0, [[BB:%.*]] ], [ [[PGOCOUNT212:%.*]], [[BB11:%.*]] ]
-; PROMO-NEXT: [[PGOCOUNT19:%.*]] = phi i64 [ 0, [[BB]] ], [ [[PGOCOUNT18:%.*]], [[BB11]] ]
-; PROMO-NEXT: [[PGOCOUNT6:%.*]] = phi i64 [ 0, [[BB]] ], [ [[PGOCOUNT5:%.*]], [[BB11]] ]
-; PROMO-NEXT: [[T2:%.*]] = phi i32 [ 0, [[BB]] ], [ [[T12:%.*]], [[BB11]] ]
-; PROMO-NEXT: [[T3:%.*]] = icmp sgt i32 [[T2]], [[ARG]]
-; PROMO-NEXT: br i1 [[T3]], label [[BB7:%.*]], label [[BB4:%.*]]
-; PROMO: bb4:
-; PROMO-NEXT: [[TMP0:%.*]] = add i64 [[PGOCOUNT6]], 1
-; PROMO-NEXT: tail call void @bar(i32 1)
-; PROMO-NEXT: [[T5:%.*]] = load i32, i32* @g, align 4
-; PROMO-NEXT: [[T6:%.*]] = icmp sgt i32 [[T5]], 100
-; PROMO-NEXT: br i1 [[T6]], label [[BB15_0:%.*]], label [[BB11]]
-; PROMO: bb7:
-; PROMO-NEXT: [[T8:%.*]] = icmp slt i32 [[T2]], [[T]]
-; PROMO-NEXT: br i1 [[T8]], label [[BB9:%.*]], label [[BB10:%.*]]
-; PROMO: bb9:
-; PROMO-NEXT: [[TMP1:%.*]] = add i64 [[PGOCOUNT19]], 1
-; PROMO-NEXT: tail call void @bar(i32 2)
-; PROMO-NEXT: br label [[BB11]]
-; PROMO: bb10:
-; PROMO-NEXT: [[TMP2:%.*]] = add i64 [[PGOCOUNT213]], 1
-; PROMO-NEXT: tail call void @bar(i32 3)
-; PROMO-NEXT: br label [[BB11]]
-; PROMO: bb11:
-; PROMO-NEXT: [[PGOCOUNT212]] = phi i64 [ [[TMP2]], [[BB10]] ], [ [[PGOCOUNT213]], [[BB9]] ], [ [[PGOCOUNT213]], [[BB4]] ]
-; PROMO-NEXT: [[PGOCOUNT18]] = phi i64 [ [[PGOCOUNT19]], [[BB10]] ], [ [[TMP1]], [[BB9]] ], [ [[PGOCOUNT19]], [[BB4]] ]
-; PROMO-NEXT: [[PGOCOUNT5]] = phi i64 [ [[PGOCOUNT6]], [[BB10]] ], [ [[PGOCOUNT6]], [[BB9]] ], [ [[TMP0]], [[BB4]] ]
-; PROMO-NEXT: [[T12]] = add nuw nsw i32 [[T2]], 1
-; PROMO-NEXT: [[T13:%.*]] = icmp slt i32 [[T2]], 99
-; PROMO-NEXT: br i1 [[T13]], label [[BB1]], label [[BB14:%.*]]
-; PROMO: bb14:
-; PROMO-NEXT: [[PGOCOUNT_PROMOTED7:%.*]] = load i64, i64* getelementptr inbounds ([5 x i64], [5 x i64]* @__profc_foo, i64 0, i64 1), align 4
-; PROMO-NEXT: [[TMP3:%.*]] = add i64 [[PGOCOUNT_PROMOTED7]], [[PGOCOUNT5]]
-; PROMO-NEXT: store i64 [[TMP3]], i64* getelementptr inbounds ([5 x i64], [5 x i64]* @__profc_foo, i64 0, i64 1), align 4
-; PROMO-NEXT: [[PGOCOUNT_PROMOTED11:%.*]] = load i64, i64* getelementptr inbounds ([5 x i64], [5 x i64]* @__profc_foo, i64 0, i64 0), align 4
-; PROMO-NEXT: [[TMP4:%.*]] = add i64 [[PGOCOUNT_PROMOTED11]], [[PGOCOUNT18]]
-; PROMO-NEXT: store i64 [[TMP4]], i64* getelementptr inbounds ([5 x i64], [5 x i64]* @__profc_foo, i64 0, i64 0), align 4
-; PROMO-NEXT: [[PGOCOUNT_PROMOTED15:%.*]] = load i64, i64* getelementptr inbounds ([5 x i64], [5 x i64]* @__profc_foo, i64 0, i64 2), align 4
-; PROMO-NEXT: [[TMP5:%.*]] = add i64 [[PGOCOUNT_PROMOTED15]], [[PGOCOUNT212]]
-; PROMO-NEXT: store i64 [[TMP5]], i64* getelementptr inbounds ([5 x i64], [5 x i64]* @__profc_foo, i64 0, i64 2), align 4
-; PROMO-NEXT: [[PGOCOUNT3:%.*]] = load i64, i64* getelementptr inbounds ([5 x i64], [5 x i64]* @__profc_foo, i64 0, i64 3), align 4
-; PROMO-NEXT: [[TMP6:%.*]] = add i64 [[PGOCOUNT3]], 1
-; PROMO-NEXT: store i64 [[TMP6]], i64* getelementptr inbounds ([5 x i64], [5 x i64]* @__profc_foo, i64 0, i64 3), align 4
-; PROMO-NEXT: tail call void @bar(i32 0)
-; PROMO-NEXT: br label [[BB15:%.*]]
-; PROMO: bb15_0:
-; PROMO-NEXT: [[PGOCOUNT_PROMOTED:%.*]] = load i64, i64* getelementptr inbounds ([5 x i64], [5 x i64]* @__profc_foo, i64 0, i64 1), align 4
-; PROMO-NEXT: [[TMP7:%.*]] = add i64 [[PGOCOUNT_PROMOTED]], [[TMP0]]
-; PROMO-NEXT: store i64 [[TMP7]], i64* getelementptr inbounds ([5 x i64], [5 x i64]* @__profc_foo, i64 0, i64 1), align 4
-; PROMO-NEXT: [[PGOCOUNT_PROMOTED10:%.*]] = load i64, i64* getelementptr inbounds ([5 x i64], [5 x i64]* @__profc_foo, i64 0, i64 0), align 4
-; PROMO-NEXT: [[TMP8:%.*]] = add i64 [[PGOCOUNT_PROMOTED10]], [[PGOCOUNT19]]
-; PROMO-NEXT: store i64 [[TMP8]], i64* getelementptr inbounds ([5 x i64], [5 x i64]* @__profc_foo, i64 0, i64 0), align 4
-; PROMO-NEXT: [[PGOCOUNT_PROMOTED14:%.*]] = load i64, i64* getelementptr inbounds ([5 x i64], [5 x i64]* @__profc_foo, i64 0, i64 2), align 4
-; PROMO-NEXT: [[TMP9:%.*]] = add i64 [[PGOCOUNT_PROMOTED14]], [[PGOCOUNT213]]
-; PROMO-NEXT: store i64 [[TMP9]], i64* getelementptr inbounds ([5 x i64], [5 x i64]* @__profc_foo, i64 0, i64 2), align 4
-; PROMO-NEXT: [[PGOCOUNT4:%.*]] = load i64, i64* getelementptr inbounds ([5 x i64], [5 x i64]* @__profc_foo, i64 0, i64 4), align 4
-; PROMO-NEXT: [[TMP10:%.*]] = add i64 [[PGOCOUNT4]], 1
-; PROMO-NEXT: store i64 [[TMP10]], i64* getelementptr inbounds ([5 x i64], [5 x i64]* @__profc_foo, i64 0, i64 4), align 4
-; PROMO-NEXT: br label [[BB15]]
-; PROMO: bb15:
-; PROMO-NEXT: tail call void @bar(i32 1)
-; PROMO-NEXT: ret void
-;
+; PROMO-LABEL: @foo
bb:
- %t = add nsw i32 %arg, -1
+ %tmp = add nsw i32 %arg, -1
br label %bb1
bb1: ; preds = %bb11, %bb
- %t2 = phi i32 [ 0, %bb ], [ %t12, %bb11 ]
- %t3 = icmp sgt i32 %t2, %arg
- br i1 %t3, label %bb7, label %bb4
+ %tmp2 = phi i32 [ 0, %bb ], [ %tmp12, %bb11 ]
+ %tmp3 = icmp sgt i32 %tmp2, %arg
+ br i1 %tmp3, label %bb7, label %bb4
bb4: ; preds = %bb1
tail call void @bar(i32 1)
- %t5 = load i32, i32* @g, align 4
- %t6 = icmp sgt i32 %t5, 100
- br i1 %t6, label %bb15_0, label %bb11
+ %tmp5 = load i32, i32* @g, align 4
+ %tmp6 = icmp sgt i32 %tmp5, 100
+ br i1 %tmp6, label %bb15_0, label %bb11
bb7: ; preds = %bb1
- %t8 = icmp slt i32 %t2, %t
- br i1 %t8, label %bb9, label %bb10
+ %tmp8 = icmp slt i32 %tmp2, %tmp
+ br i1 %tmp8, label %bb9, label %bb10
bb9: ; preds = %bb7
tail call void @bar(i32 2)
@@ -100,18 +32,45 @@ bb10: ; preds = %bb7
br label %bb11
bb11: ; preds = %bb10, %bb9, %bb4
- %t12 = add nuw nsw i32 %t2, 1
- %t13 = icmp slt i32 %t2, 99
- br i1 %t13, label %bb1, label %bb14
+ %tmp12 = add nuw nsw i32 %tmp2, 1
+ %tmp13 = icmp slt i32 %tmp2, 99
+ br i1 %tmp13, label %bb1, label %bb14
bb14: ; preds = %bb11
+; PROMO-LABEL: bb14:
tail call void @bar(i32 0)
br label %bb15
-
+; PROMO: %pgocount.promoted{{.*}} = load {{.*}} @__profc_foo{{.*}} 0)
+; PROMO-NEXT: add
+; PROMO-NEXT: store {{.*}}@__profc_foo{{.*}}0)
+; PROMO-NEXT: %pgocount.promoted{{.*}} = load {{.*}} @__profc_foo{{.*}} 1)
+; PROMO-NEXT: add
+; PROMO-NEXT: store {{.*}}@__profc_foo{{.*}}1)
+; PROMO-NEXT: %pgocount.promoted{{.*}} = load {{.*}} @__profc_foo{{.*}} 2)
+; PROMO-NEXT: add
+; PROMO-NEXT: store {{.*}}@__profc_foo{{.*}}2)
+; PROMO-NEXT: %pgocount{{.*}} = load {{.*}} @__profc_foo{{.*}} 3)
+; PROMO-NEXT: add
+; PROMO-NEXT: store {{.*}}@__profc_foo{{.*}}3)
bb15_0: ; preds = %bb11
+; PROMO-LABEL: bb15_0:
br label %bb15
+; PROMO: %pgocount.promoted{{.*}} = load {{.*}} @__profc_foo{{.*}} 0)
+; PROMO-NEXT: add
+; PROMO-NEXT: store {{.*}}@__profc_foo{{.*}}0)
+; PROMO-NEXT: %pgocount.promoted{{.*}} = load {{.*}} @__profc_foo{{.*}} 1)
+; PROMO-NEXT: add
+; PROMO-NEXT: store {{.*}}@__profc_foo{{.*}}1)
+; PROMO-NEXT: %pgocount.promoted{{.*}} = load {{.*}} @__profc_foo{{.*}} 2)
+; PROMO-NEXT: add
+; PROMO-NEXT: store {{.*}}@__profc_foo{{.*}}2)
+; PROMO-NEXT: %pgocount{{.*}} = load {{.*}} @__profc_foo{{.*}} 4)
+; PROMO-NEXT: add
+; PROMO-NEXT: store {{.*}}@__profc_foo{{.*}}4)
+; PROMO-NOT: @__profc_foo
+
bb15: ; preds = %bb14, %bb4
tail call void @bar(i32 1)
diff --git a/llvm/test/Transforms/PGOProfile/landingpad.ll b/llvm/test/Transforms/PGOProfile/landingpad.ll
index 5191e118210f..a0ca799fa8a4 100644
--- a/llvm/test/Transforms/PGOProfile/landingpad.ll
+++ b/llvm/test/Transforms/PGOProfile/landingpad.ll
@@ -85,8 +85,7 @@ catch.dispatch:
; GEN: catch.dispatch:
; GEN-NOT: call void @llvm.instrprof.increment
%tmp3 = call i32 @llvm.eh.typeid.for(i8* bitcast (i8** @_ZTIi to i8*))
- %c = icmp ne i32 %tmp2, %tmp3
- %matches = xor i1 %c, -1
+ %matches = icmp eq i32 %tmp2, %tmp3
br i1 %matches, label %catch, label %eh.resume
; USE: br i1 %matches, label %catch, label %eh.resume
; USE-SAME: !prof ![[BW_CATCH_DISPATCH:[0-9]+]]
More information about the llvm-commits
mailing list