[llvm] r344955 - [x86] add test for PR25498 and complete checks; NFC
Sanjay Patel via llvm-commits
llvm-commits at lists.llvm.org
Mon Oct 22 14:11:16 PDT 2018
Author: spatel
Date: Mon Oct 22 14:11:15 2018
New Revision: 344955
URL: http://llvm.org/viewvc/llvm-project?rev=344955&view=rev
Log:
[x86] add test for PR25498 and complete checks; NFC
Might as well test the actual codegen instead of just the absence of crashing.
Modified:
llvm/trunk/test/CodeGen/X86/scheduler-backtracking.ll
Modified: llvm/trunk/test/CodeGen/X86/scheduler-backtracking.ll
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/X86/scheduler-backtracking.ll?rev=344955&r1=344954&r2=344955&view=diff
==============================================================================
--- llvm/trunk/test/CodeGen/X86/scheduler-backtracking.ll (original)
+++ llvm/trunk/test/CodeGen/X86/scheduler-backtracking.ll Mon Oct 22 14:11:15 2018
@@ -1,15 +1,462 @@
-; RUN: llc -mtriple=x86_64-- < %s -pre-RA-sched=list-ilp | FileCheck %s
-; RUN: llc -mtriple=x86_64-- < %s -pre-RA-sched=list-hybrid | FileCheck %s
-; RUN: llc -mtriple=x86_64-- < %s -pre-RA-sched=source | FileCheck %s
-; RUN: llc -mtriple=x86_64-- < %s -pre-RA-sched=list-burr | FileCheck %s
-; RUN: llc -mtriple=x86_64-- < %s -pre-RA-sched=linearize | FileCheck %s
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
+; RUN: llc -mtriple=x86_64-- < %s -pre-RA-sched=list-ilp | FileCheck %s --check-prefix=ILP
+; RUN: llc -mtriple=x86_64-- < %s -pre-RA-sched=list-hybrid | FileCheck %s --check-prefix=HYBRID
+; RUN: llc -mtriple=x86_64-- < %s -pre-RA-sched=list-burr | FileCheck %s --check-prefix=BURR
+; RUN: llc -mtriple=x86_64-- < %s -pre-RA-sched=source | FileCheck %s --check-prefix=SRC
+; RUN: llc -mtriple=x86_64-- < %s -pre-RA-sched=linearize | FileCheck %s --check-prefix=LIN
; PR22304 https://llvm.org/bugs/show_bug.cgi?id=22304
; Tests checking backtracking in source scheduler. llc used to crash on them.
-; CHECK-LABEL: test1
-define i256 @test1(i256 %a) {
- %b = add i256 %a, 1
+define i256 @test1(i256 %a) nounwind {
+; ILP-LABEL: test1:
+; ILP: # %bb.0:
+; ILP-NEXT: pushq %rbp
+; ILP-NEXT: pushq %r15
+; ILP-NEXT: pushq %r14
+; ILP-NEXT: pushq %r13
+; ILP-NEXT: pushq %r12
+; ILP-NEXT: pushq %rbx
+; ILP-NEXT: movq %rcx, %r9
+; ILP-NEXT: movq %rdi, {{[-0-9]+}}(%r{{[sb]}}p) # 8-byte Spill
+; ILP-NEXT: xorl %eax, %eax
+; ILP-NEXT: addq $1, %rsi
+; ILP-NEXT: adcq $0, %rdx
+; ILP-NEXT: adcq $0, %r9
+; ILP-NEXT: adcq $0, %r8
+; ILP-NEXT: leal 1(%rsi,%rsi), %edi
+; ILP-NEXT: movl $1, %ebp
+; ILP-NEXT: xorl %r14d, %r14d
+; ILP-NEXT: movl %edi, %ecx
+; ILP-NEXT: shldq %cl, %rbp, %r14
+; ILP-NEXT: movl $1, %r11d
+; ILP-NEXT: shlq %cl, %r11
+; ILP-NEXT: movb $-128, %r10b
+; ILP-NEXT: subb %dil, %r10b
+; ILP-NEXT: movq %r9, %r13
+; ILP-NEXT: movl %r10d, %ecx
+; ILP-NEXT: shlq %cl, %r13
+; ILP-NEXT: movl $1, %r12d
+; ILP-NEXT: shrdq %cl, %rax, %r12
+; ILP-NEXT: xorl %r15d, %r15d
+; ILP-NEXT: movl %edi, %ecx
+; ILP-NEXT: shldq %cl, %r15, %r15
+; ILP-NEXT: movq %rsi, %rbx
+; ILP-NEXT: shrdq %cl, %rdx, %rbx
+; ILP-NEXT: shrq %cl, %rdx
+; ILP-NEXT: addb $-128, %cl
+; ILP-NEXT: shrdq %cl, %r8, %r9
+; ILP-NEXT: testb $64, %dil
+; ILP-NEXT: cmovneq %r11, %r14
+; ILP-NEXT: cmoveq %rbx, %rdx
+; ILP-NEXT: cmovneq %rax, %r15
+; ILP-NEXT: cmovneq %rax, %r11
+; ILP-NEXT: testb $64, %r10b
+; ILP-NEXT: cmovneq %rax, %r12
+; ILP-NEXT: cmovneq %rax, %r13
+; ILP-NEXT: movl $1, %ebx
+; ILP-NEXT: shlq %cl, %rbx
+; ILP-NEXT: orl %edx, %r13d
+; ILP-NEXT: xorl %edx, %edx
+; ILP-NEXT: movl $1, %ebp
+; ILP-NEXT: shldq %cl, %rbp, %rdx
+; ILP-NEXT: shrq %cl, %r8
+; ILP-NEXT: testb $64, %cl
+; ILP-NEXT: cmoveq %r9, %r8
+; ILP-NEXT: cmovneq %rbx, %rdx
+; ILP-NEXT: cmovneq %rax, %rbx
+; ILP-NEXT: testb %dil, %dil
+; ILP-NEXT: cmovsq %rax, %r14
+; ILP-NEXT: cmovsq %rax, %r11
+; ILP-NEXT: jns .LBB0_2
+; ILP-NEXT: # %bb.1:
+; ILP-NEXT: movl %r8d, %r13d
+; ILP-NEXT: .LBB0_2:
+; ILP-NEXT: je .LBB0_4
+; ILP-NEXT: # %bb.3:
+; ILP-NEXT: movl %r13d, %esi
+; ILP-NEXT: .LBB0_4:
+; ILP-NEXT: cmovnsq %r12, %rbx
+; ILP-NEXT: cmoveq %rax, %rbx
+; ILP-NEXT: cmovnsq %r15, %rdx
+; ILP-NEXT: cmoveq %rax, %rdx
+; ILP-NEXT: testb $1, %sil
+; ILP-NEXT: cmovneq %rax, %rdx
+; ILP-NEXT: movq {{[-0-9]+}}(%r{{[sb]}}p), %rax # 8-byte Reload
+; ILP-NEXT: movq %rdx, 24(%rax)
+; ILP-NEXT: cmovneq %rax, %rbx
+; ILP-NEXT: movq %rbx, 16(%rax)
+; ILP-NEXT: cmovneq %rax, %r14
+; ILP-NEXT: movq %r14, 8(%rax)
+; ILP-NEXT: cmovneq %rax, %r11
+; ILP-NEXT: movq %r11, (%rax)
+; ILP-NEXT: popq %rbx
+; ILP-NEXT: popq %r12
+; ILP-NEXT: popq %r13
+; ILP-NEXT: popq %r14
+; ILP-NEXT: popq %r15
+; ILP-NEXT: popq %rbp
+; ILP-NEXT: retq
+;
+; HYBRID-LABEL: test1:
+; HYBRID: # %bb.0:
+; HYBRID-NEXT: pushq %rbp
+; HYBRID-NEXT: pushq %r15
+; HYBRID-NEXT: pushq %r14
+; HYBRID-NEXT: pushq %r13
+; HYBRID-NEXT: pushq %r12
+; HYBRID-NEXT: pushq %rbx
+; HYBRID-NEXT: movq %rcx, %r9
+; HYBRID-NEXT: movq %rdi, %rax
+; HYBRID-NEXT: addq $1, %rsi
+; HYBRID-NEXT: adcq $0, %rdx
+; HYBRID-NEXT: adcq $0, %r9
+; HYBRID-NEXT: adcq $0, %r8
+; HYBRID-NEXT: xorl %r10d, %r10d
+; HYBRID-NEXT: leal 1(%rsi,%rsi), %edi
+; HYBRID-NEXT: xorl %r14d, %r14d
+; HYBRID-NEXT: movl %edi, %ecx
+; HYBRID-NEXT: shldq %cl, %r14, %r14
+; HYBRID-NEXT: testb $64, %dil
+; HYBRID-NEXT: cmovneq %r10, %r14
+; HYBRID-NEXT: movl $1, %ebp
+; HYBRID-NEXT: movl $1, %r12d
+; HYBRID-NEXT: shlq %cl, %r12
+; HYBRID-NEXT: testb $64, %dil
+; HYBRID-NEXT: movq %r12, %r11
+; HYBRID-NEXT: cmovneq %r10, %r11
+; HYBRID-NEXT: movq %rsi, %rbx
+; HYBRID-NEXT: shrdq %cl, %rdx, %rbx
+; HYBRID-NEXT: shrq %cl, %rdx
+; HYBRID-NEXT: testb $64, %dil
+; HYBRID-NEXT: cmoveq %rbx, %rdx
+; HYBRID-NEXT: xorl %r15d, %r15d
+; HYBRID-NEXT: shldq %cl, %rbp, %r15
+; HYBRID-NEXT: testb $64, %dil
+; HYBRID-NEXT: cmovneq %r12, %r15
+; HYBRID-NEXT: movb $-128, %cl
+; HYBRID-NEXT: subb %dil, %cl
+; HYBRID-NEXT: movq %r9, %r13
+; HYBRID-NEXT: shlq %cl, %r13
+; HYBRID-NEXT: movl $1, %r12d
+; HYBRID-NEXT: shrdq %cl, %r10, %r12
+; HYBRID-NEXT: testb $64, %cl
+; HYBRID-NEXT: cmovneq %r10, %r12
+; HYBRID-NEXT: cmovneq %r10, %r13
+; HYBRID-NEXT: orl %edx, %r13d
+; HYBRID-NEXT: movl %edi, %ecx
+; HYBRID-NEXT: addb $-128, %cl
+; HYBRID-NEXT: shrdq %cl, %r8, %r9
+; HYBRID-NEXT: shrq %cl, %r8
+; HYBRID-NEXT: xorl %edx, %edx
+; HYBRID-NEXT: shldq %cl, %rbp, %rdx
+; HYBRID-NEXT: shlq %cl, %rbp
+; HYBRID-NEXT: testb $64, %cl
+; HYBRID-NEXT: cmovneq %rbp, %rdx
+; HYBRID-NEXT: cmoveq %r9, %r8
+; HYBRID-NEXT: cmovneq %r10, %rbp
+; HYBRID-NEXT: testb %dil, %dil
+; HYBRID-NEXT: jns .LBB0_2
+; HYBRID-NEXT: # %bb.1:
+; HYBRID-NEXT: movl %r8d, %r13d
+; HYBRID-NEXT: .LBB0_2:
+; HYBRID-NEXT: je .LBB0_4
+; HYBRID-NEXT: # %bb.3:
+; HYBRID-NEXT: movl %r13d, %esi
+; HYBRID-NEXT: .LBB0_4:
+; HYBRID-NEXT: cmovsq %r10, %r15
+; HYBRID-NEXT: cmovnsq %r12, %rbp
+; HYBRID-NEXT: cmoveq %r10, %rbp
+; HYBRID-NEXT: cmovnsq %r14, %rdx
+; HYBRID-NEXT: cmoveq %r10, %rdx
+; HYBRID-NEXT: cmovsq %r10, %r11
+; HYBRID-NEXT: testb $1, %sil
+; HYBRID-NEXT: cmovneq %rax, %rdx
+; HYBRID-NEXT: movq %rdx, 24(%rax)
+; HYBRID-NEXT: cmovneq %rax, %rbp
+; HYBRID-NEXT: movq %rbp, 16(%rax)
+; HYBRID-NEXT: cmovneq %rax, %r15
+; HYBRID-NEXT: movq %r15, 8(%rax)
+; HYBRID-NEXT: cmovneq %rax, %r11
+; HYBRID-NEXT: movq %r11, (%rax)
+; HYBRID-NEXT: popq %rbx
+; HYBRID-NEXT: popq %r12
+; HYBRID-NEXT: popq %r13
+; HYBRID-NEXT: popq %r14
+; HYBRID-NEXT: popq %r15
+; HYBRID-NEXT: popq %rbp
+; HYBRID-NEXT: retq
+;
+; BURR-LABEL: test1:
+; BURR: # %bb.0:
+; BURR-NEXT: pushq %rbp
+; BURR-NEXT: pushq %r15
+; BURR-NEXT: pushq %r14
+; BURR-NEXT: pushq %r13
+; BURR-NEXT: pushq %r12
+; BURR-NEXT: pushq %rbx
+; BURR-NEXT: movq %rcx, %r9
+; BURR-NEXT: movq %rdi, %rax
+; BURR-NEXT: addq $1, %rsi
+; BURR-NEXT: adcq $0, %rdx
+; BURR-NEXT: adcq $0, %r9
+; BURR-NEXT: adcq $0, %r8
+; BURR-NEXT: xorl %r10d, %r10d
+; BURR-NEXT: leal 1(%rsi,%rsi), %edi
+; BURR-NEXT: xorl %r14d, %r14d
+; BURR-NEXT: movl %edi, %ecx
+; BURR-NEXT: shldq %cl, %r14, %r14
+; BURR-NEXT: testb $64, %dil
+; BURR-NEXT: cmovneq %r10, %r14
+; BURR-NEXT: movl $1, %ebp
+; BURR-NEXT: movl $1, %r12d
+; BURR-NEXT: shlq %cl, %r12
+; BURR-NEXT: testb $64, %dil
+; BURR-NEXT: movq %r12, %r11
+; BURR-NEXT: cmovneq %r10, %r11
+; BURR-NEXT: movq %rsi, %rbx
+; BURR-NEXT: shrdq %cl, %rdx, %rbx
+; BURR-NEXT: shrq %cl, %rdx
+; BURR-NEXT: testb $64, %dil
+; BURR-NEXT: cmoveq %rbx, %rdx
+; BURR-NEXT: xorl %r15d, %r15d
+; BURR-NEXT: shldq %cl, %rbp, %r15
+; BURR-NEXT: testb $64, %dil
+; BURR-NEXT: cmovneq %r12, %r15
+; BURR-NEXT: movb $-128, %cl
+; BURR-NEXT: subb %dil, %cl
+; BURR-NEXT: movq %r9, %r13
+; BURR-NEXT: shlq %cl, %r13
+; BURR-NEXT: movl $1, %r12d
+; BURR-NEXT: shrdq %cl, %r10, %r12
+; BURR-NEXT: testb $64, %cl
+; BURR-NEXT: cmovneq %r10, %r12
+; BURR-NEXT: cmovneq %r10, %r13
+; BURR-NEXT: orl %edx, %r13d
+; BURR-NEXT: movl %edi, %ecx
+; BURR-NEXT: addb $-128, %cl
+; BURR-NEXT: shrdq %cl, %r8, %r9
+; BURR-NEXT: xorl %edx, %edx
+; BURR-NEXT: shldq %cl, %rbp, %rdx
+; BURR-NEXT: shrq %cl, %r8
+; BURR-NEXT: shlq %cl, %rbp
+; BURR-NEXT: testb $64, %cl
+; BURR-NEXT: cmovneq %rbp, %rdx
+; BURR-NEXT: cmoveq %r9, %r8
+; BURR-NEXT: cmovneq %r10, %rbp
+; BURR-NEXT: testb %dil, %dil
+; BURR-NEXT: jns .LBB0_2
+; BURR-NEXT: # %bb.1:
+; BURR-NEXT: movl %r8d, %r13d
+; BURR-NEXT: .LBB0_2:
+; BURR-NEXT: je .LBB0_4
+; BURR-NEXT: # %bb.3:
+; BURR-NEXT: movl %r13d, %esi
+; BURR-NEXT: .LBB0_4:
+; BURR-NEXT: cmovsq %r10, %r15
+; BURR-NEXT: cmovnsq %r12, %rbp
+; BURR-NEXT: cmoveq %r10, %rbp
+; BURR-NEXT: cmovnsq %r14, %rdx
+; BURR-NEXT: cmoveq %r10, %rdx
+; BURR-NEXT: cmovsq %r10, %r11
+; BURR-NEXT: testb $1, %sil
+; BURR-NEXT: cmovneq %rax, %rdx
+; BURR-NEXT: movq %rdx, 24(%rax)
+; BURR-NEXT: cmovneq %rax, %rbp
+; BURR-NEXT: movq %rbp, 16(%rax)
+; BURR-NEXT: cmovneq %rax, %r15
+; BURR-NEXT: movq %r15, 8(%rax)
+; BURR-NEXT: cmovneq %rax, %r11
+; BURR-NEXT: movq %r11, (%rax)
+; BURR-NEXT: popq %rbx
+; BURR-NEXT: popq %r12
+; BURR-NEXT: popq %r13
+; BURR-NEXT: popq %r14
+; BURR-NEXT: popq %r15
+; BURR-NEXT: popq %rbp
+; BURR-NEXT: retq
+;
+; SRC-LABEL: test1:
+; SRC: # %bb.0:
+; SRC-NEXT: pushq %rbp
+; SRC-NEXT: pushq %r15
+; SRC-NEXT: pushq %r14
+; SRC-NEXT: pushq %r13
+; SRC-NEXT: pushq %r12
+; SRC-NEXT: pushq %rbx
+; SRC-NEXT: movq %rcx, %r9
+; SRC-NEXT: movq %rdi, %rax
+; SRC-NEXT: addq $1, %rsi
+; SRC-NEXT: adcq $0, %rdx
+; SRC-NEXT: adcq $0, %r9
+; SRC-NEXT: adcq $0, %r8
+; SRC-NEXT: leal 1(%rsi,%rsi), %r11d
+; SRC-NEXT: movb $-128, %r10b
+; SRC-NEXT: subb %r11b, %r10b
+; SRC-NEXT: movq %r9, %r12
+; SRC-NEXT: movl %r10d, %ecx
+; SRC-NEXT: shlq %cl, %r12
+; SRC-NEXT: movq %rsi, %rbp
+; SRC-NEXT: movl %r11d, %ecx
+; SRC-NEXT: shrdq %cl, %rdx, %rbp
+; SRC-NEXT: shrq %cl, %rdx
+; SRC-NEXT: xorl %r15d, %r15d
+; SRC-NEXT: movl $1, %edi
+; SRC-NEXT: xorl %r14d, %r14d
+; SRC-NEXT: shldq %cl, %rdi, %r14
+; SRC-NEXT: xorl %r13d, %r13d
+; SRC-NEXT: shldq %cl, %r13, %r13
+; SRC-NEXT: movl $1, %ebx
+; SRC-NEXT: shlq %cl, %rbx
+; SRC-NEXT: testb $64, %r11b
+; SRC-NEXT: cmoveq %rbp, %rdx
+; SRC-NEXT: cmovneq %rbx, %r14
+; SRC-NEXT: cmovneq %r15, %rbx
+; SRC-NEXT: cmovneq %r15, %r13
+; SRC-NEXT: movl $1, %ebp
+; SRC-NEXT: movl %r10d, %ecx
+; SRC-NEXT: shrdq %cl, %r15, %rbp
+; SRC-NEXT: testb $64, %r10b
+; SRC-NEXT: cmovneq %r15, %r12
+; SRC-NEXT: cmovneq %r15, %rbp
+; SRC-NEXT: orl %edx, %r12d
+; SRC-NEXT: movl %r11d, %ecx
+; SRC-NEXT: addb $-128, %cl
+; SRC-NEXT: shrdq %cl, %r8, %r9
+; SRC-NEXT: shrq %cl, %r8
+; SRC-NEXT: xorl %edx, %edx
+; SRC-NEXT: shldq %cl, %rdi, %rdx
+; SRC-NEXT: shlq %cl, %rdi
+; SRC-NEXT: testb $64, %cl
+; SRC-NEXT: cmoveq %r9, %r8
+; SRC-NEXT: cmovneq %rdi, %rdx
+; SRC-NEXT: cmovneq %r15, %rdi
+; SRC-NEXT: testb %r11b, %r11b
+; SRC-NEXT: jns .LBB0_2
+; SRC-NEXT: # %bb.1:
+; SRC-NEXT: movl %r8d, %r12d
+; SRC-NEXT: .LBB0_2:
+; SRC-NEXT: je .LBB0_4
+; SRC-NEXT: # %bb.3:
+; SRC-NEXT: movl %r12d, %esi
+; SRC-NEXT: .LBB0_4:
+; SRC-NEXT: cmovnsq %r13, %rdx
+; SRC-NEXT: cmoveq %r15, %rdx
+; SRC-NEXT: cmovnsq %rbp, %rdi
+; SRC-NEXT: cmoveq %r15, %rdi
+; SRC-NEXT: cmovsq %r15, %r14
+; SRC-NEXT: cmovsq %r15, %rbx
+; SRC-NEXT: testb $1, %sil
+; SRC-NEXT: cmovneq %rax, %rbx
+; SRC-NEXT: cmovneq %rax, %r14
+; SRC-NEXT: cmovneq %rax, %rdi
+; SRC-NEXT: cmovneq %rax, %rdx
+; SRC-NEXT: movq %rdx, 24(%rax)
+; SRC-NEXT: movq %rdi, 16(%rax)
+; SRC-NEXT: movq %r14, 8(%rax)
+; SRC-NEXT: movq %rbx, (%rax)
+; SRC-NEXT: popq %rbx
+; SRC-NEXT: popq %r12
+; SRC-NEXT: popq %r13
+; SRC-NEXT: popq %r14
+; SRC-NEXT: popq %r15
+; SRC-NEXT: popq %rbp
+; SRC-NEXT: retq
+;
+; LIN-LABEL: test1:
+; LIN: # %bb.0:
+; LIN-NEXT: pushq %rbp
+; LIN-NEXT: pushq %r15
+; LIN-NEXT: pushq %r14
+; LIN-NEXT: pushq %r12
+; LIN-NEXT: pushq %rbx
+; LIN-NEXT: movq %rcx, %r9
+; LIN-NEXT: movq %rdi, %rax
+; LIN-NEXT: xorl %r15d, %r15d
+; LIN-NEXT: movl $1, %r14d
+; LIN-NEXT: addq $1, %rsi
+; LIN-NEXT: leal 1(%rsi,%rsi), %ebp
+; LIN-NEXT: movl $1, %r12d
+; LIN-NEXT: movl %ebp, %ecx
+; LIN-NEXT: shlq %cl, %r12
+; LIN-NEXT: testb $64, %bpl
+; LIN-NEXT: movq %r12, %rbx
+; LIN-NEXT: cmovneq %r15, %rbx
+; LIN-NEXT: testb %bpl, %bpl
+; LIN-NEXT: cmovsq %r15, %rbx
+; LIN-NEXT: adcq $0, %rdx
+; LIN-NEXT: adcq $0, %r9
+; LIN-NEXT: adcq $0, %r8
+; LIN-NEXT: movl %ebp, %r10d
+; LIN-NEXT: addb $-128, %r10b
+; LIN-NEXT: movq %r9, %rdi
+; LIN-NEXT: movl %r10d, %ecx
+; LIN-NEXT: shrdq %cl, %r8, %rdi
+; LIN-NEXT: shrq %cl, %r8
+; LIN-NEXT: testb $64, %r10b
+; LIN-NEXT: cmoveq %rdi, %r8
+; LIN-NEXT: movq %rsi, %rdi
+; LIN-NEXT: movl %ebp, %ecx
+; LIN-NEXT: shrdq %cl, %rdx, %rdi
+; LIN-NEXT: shrq %cl, %rdx
+; LIN-NEXT: cmoveq %rdi, %rdx
+; LIN-NEXT: movb $-128, %r11b
+; LIN-NEXT: subb %bpl, %r11b
+; LIN-NEXT: movl %r11d, %ecx
+; LIN-NEXT: shlq %cl, %r9
+; LIN-NEXT: testb $64, %r11b
+; LIN-NEXT: cmovneq %r15, %r9
+; LIN-NEXT: orl %edx, %r9d
+; LIN-NEXT: jns .LBB0_2
+; LIN-NEXT: # %bb.1:
+; LIN-NEXT: movl %r8d, %r9d
+; LIN-NEXT: .LBB0_2:
+; LIN-NEXT: je .LBB0_4
+; LIN-NEXT: # %bb.3:
+; LIN-NEXT: movl %r9d, %esi
+; LIN-NEXT: .LBB0_4:
+; LIN-NEXT: testb $1, %sil
+; LIN-NEXT: cmovneq %rax, %rbx
+; LIN-NEXT: movq %rbx, (%rax)
+; LIN-NEXT: xorl %edx, %edx
+; LIN-NEXT: movl %ebp, %ecx
+; LIN-NEXT: shldq %cl, %r14, %rdx
+; LIN-NEXT: cmovneq %r12, %rdx
+; LIN-NEXT: cmovsq %r15, %rdx
+; LIN-NEXT: cmovneq %rax, %rdx
+; LIN-NEXT: movq %rdx, 8(%rax)
+; LIN-NEXT: movl $1, %edx
+; LIN-NEXT: movl %r10d, %ecx
+; LIN-NEXT: shlq %cl, %rdx
+; LIN-NEXT: movq %rdx, %rsi
+; LIN-NEXT: cmovneq %r15, %rsi
+; LIN-NEXT: movl $1, %edi
+; LIN-NEXT: movl %r11d, %ecx
+; LIN-NEXT: shrdq %cl, %r15, %rdi
+; LIN-NEXT: cmovneq %r15, %rdi
+; LIN-NEXT: cmovsq %rsi, %rdi
+; LIN-NEXT: cmoveq %r15, %rdi
+; LIN-NEXT: cmovneq %rax, %rdi
+; LIN-NEXT: movq %rdi, 16(%rax)
+; LIN-NEXT: xorl %esi, %esi
+; LIN-NEXT: movl %r10d, %ecx
+; LIN-NEXT: shldq %cl, %r14, %rsi
+; LIN-NEXT: cmovneq %rdx, %rsi
+; LIN-NEXT: xorl %edx, %edx
+; LIN-NEXT: movl %ebp, %ecx
+; LIN-NEXT: shldq %cl, %rdx, %rdx
+; LIN-NEXT: cmovneq %r15, %rdx
+; LIN-NEXT: cmovsq %rsi, %rdx
+; LIN-NEXT: cmoveq %r15, %rdx
+; LIN-NEXT: cmovneq %rax, %rdx
+; LIN-NEXT: movq %rdx, 24(%rax)
+; LIN-NEXT: popq %rbx
+; LIN-NEXT: popq %r12
+; LIN-NEXT: popq %r14
+; LIN-NEXT: popq %r15
+; LIN-NEXT: popq %rbp
+; LIN-NEXT: retq
+ %b = add i256 %a, 1
%m = shl i256 %b, 1
%p = add i256 %m, 1
%v = lshr i256 %b, %p
@@ -19,16 +466,436 @@ define i256 @test1(i256 %a) {
ret i256 %f
}
-; CHECK-LABEL: test2
-define i256 @test2(i256 %a) {
+define i256 @test2(i256 %a) nounwind {
+; ILP-LABEL: test2:
+; ILP: # %bb.0:
+; ILP-NEXT: movq %rdi, %rax
+; ILP-NEXT: xorl %edi, %edi
+; ILP-NEXT: movq %rsi, %r11
+; ILP-NEXT: negq %r11
+; ILP-NEXT: movl $0, %r10d
+; ILP-NEXT: sbbq %rdx, %r10
+; ILP-NEXT: movl $0, %r9d
+; ILP-NEXT: sbbq %rcx, %r9
+; ILP-NEXT: sbbq %r8, %rdi
+; ILP-NEXT: andq %rcx, %r9
+; ILP-NEXT: bsrq %r9, %rcx
+; ILP-NEXT: xorq $63, %rcx
+; ILP-NEXT: andq %r8, %rdi
+; ILP-NEXT: bsrq %rdi, %r8
+; ILP-NEXT: andq %rdx, %r10
+; ILP-NEXT: bsrq %r10, %rdx
+; ILP-NEXT: xorq $63, %r8
+; ILP-NEXT: addq $64, %rcx
+; ILP-NEXT: testq %rdi, %rdi
+; ILP-NEXT: movq $0, 24(%rax)
+; ILP-NEXT: movq $0, 16(%rax)
+; ILP-NEXT: movq $0, 8(%rax)
+; ILP-NEXT: cmovneq %r8, %rcx
+; ILP-NEXT: xorq $63, %rdx
+; ILP-NEXT: andq %rsi, %r11
+; ILP-NEXT: movl $127, %r8d
+; ILP-NEXT: bsrq %r11, %rsi
+; ILP-NEXT: cmoveq %r8, %rsi
+; ILP-NEXT: xorq $63, %rsi
+; ILP-NEXT: addq $64, %rsi
+; ILP-NEXT: testq %r10, %r10
+; ILP-NEXT: cmovneq %rdx, %rsi
+; ILP-NEXT: subq $-128, %rsi
+; ILP-NEXT: orq %r9, %rdi
+; ILP-NEXT: cmovneq %rcx, %rsi
+; ILP-NEXT: movq %rsi, (%rax)
+; ILP-NEXT: retq
+;
+; HYBRID-LABEL: test2:
+; HYBRID: # %bb.0:
+; HYBRID-NEXT: movq %rdi, %rax
+; HYBRID-NEXT: xorl %r9d, %r9d
+; HYBRID-NEXT: movq %rsi, %r11
+; HYBRID-NEXT: negq %r11
+; HYBRID-NEXT: movl $0, %r10d
+; HYBRID-NEXT: sbbq %rdx, %r10
+; HYBRID-NEXT: movl $0, %edi
+; HYBRID-NEXT: sbbq %rcx, %rdi
+; HYBRID-NEXT: sbbq %r8, %r9
+; HYBRID-NEXT: andq %r8, %r9
+; HYBRID-NEXT: bsrq %r9, %r8
+; HYBRID-NEXT: xorq $63, %r8
+; HYBRID-NEXT: andq %rcx, %rdi
+; HYBRID-NEXT: bsrq %rdi, %rcx
+; HYBRID-NEXT: xorq $63, %rcx
+; HYBRID-NEXT: addq $64, %rcx
+; HYBRID-NEXT: testq %r9, %r9
+; HYBRID-NEXT: cmovneq %r8, %rcx
+; HYBRID-NEXT: andq %rdx, %r10
+; HYBRID-NEXT: bsrq %r10, %rdx
+; HYBRID-NEXT: xorq $63, %rdx
+; HYBRID-NEXT: andq %rsi, %r11
+; HYBRID-NEXT: movl $127, %r8d
+; HYBRID-NEXT: bsrq %r11, %rsi
+; HYBRID-NEXT: cmoveq %r8, %rsi
+; HYBRID-NEXT: xorq $63, %rsi
+; HYBRID-NEXT: addq $64, %rsi
+; HYBRID-NEXT: testq %r10, %r10
+; HYBRID-NEXT: cmovneq %rdx, %rsi
+; HYBRID-NEXT: subq $-128, %rsi
+; HYBRID-NEXT: orq %r9, %rdi
+; HYBRID-NEXT: cmovneq %rcx, %rsi
+; HYBRID-NEXT: movq %rsi, (%rax)
+; HYBRID-NEXT: movq $0, 24(%rax)
+; HYBRID-NEXT: movq $0, 16(%rax)
+; HYBRID-NEXT: movq $0, 8(%rax)
+; HYBRID-NEXT: retq
+;
+; BURR-LABEL: test2:
+; BURR: # %bb.0:
+; BURR-NEXT: movq %rdi, %rax
+; BURR-NEXT: xorl %r9d, %r9d
+; BURR-NEXT: movq %rsi, %r11
+; BURR-NEXT: negq %r11
+; BURR-NEXT: movl $0, %r10d
+; BURR-NEXT: sbbq %rdx, %r10
+; BURR-NEXT: movl $0, %edi
+; BURR-NEXT: sbbq %rcx, %rdi
+; BURR-NEXT: sbbq %r8, %r9
+; BURR-NEXT: andq %r8, %r9
+; BURR-NEXT: bsrq %r9, %r8
+; BURR-NEXT: xorq $63, %r8
+; BURR-NEXT: andq %rcx, %rdi
+; BURR-NEXT: bsrq %rdi, %rcx
+; BURR-NEXT: xorq $63, %rcx
+; BURR-NEXT: addq $64, %rcx
+; BURR-NEXT: testq %r9, %r9
+; BURR-NEXT: cmovneq %r8, %rcx
+; BURR-NEXT: andq %rdx, %r10
+; BURR-NEXT: bsrq %r10, %rdx
+; BURR-NEXT: xorq $63, %rdx
+; BURR-NEXT: andq %rsi, %r11
+; BURR-NEXT: movl $127, %r8d
+; BURR-NEXT: bsrq %r11, %rsi
+; BURR-NEXT: cmoveq %r8, %rsi
+; BURR-NEXT: xorq $63, %rsi
+; BURR-NEXT: addq $64, %rsi
+; BURR-NEXT: testq %r10, %r10
+; BURR-NEXT: cmovneq %rdx, %rsi
+; BURR-NEXT: subq $-128, %rsi
+; BURR-NEXT: orq %r9, %rdi
+; BURR-NEXT: cmovneq %rcx, %rsi
+; BURR-NEXT: movq %rsi, (%rax)
+; BURR-NEXT: movq $0, 24(%rax)
+; BURR-NEXT: movq $0, 16(%rax)
+; BURR-NEXT: movq $0, 8(%rax)
+; BURR-NEXT: retq
+;
+; SRC-LABEL: test2:
+; SRC: # %bb.0:
+; SRC-NEXT: movq %rdi, %rax
+; SRC-NEXT: xorl %edi, %edi
+; SRC-NEXT: movq %rsi, %r11
+; SRC-NEXT: negq %r11
+; SRC-NEXT: movl $0, %r10d
+; SRC-NEXT: sbbq %rdx, %r10
+; SRC-NEXT: movl $0, %r9d
+; SRC-NEXT: sbbq %rcx, %r9
+; SRC-NEXT: sbbq %r8, %rdi
+; SRC-NEXT: andq %rdx, %r10
+; SRC-NEXT: andq %rcx, %r9
+; SRC-NEXT: andq %r8, %rdi
+; SRC-NEXT: andq %rsi, %r11
+; SRC-NEXT: bsrq %rdi, %rcx
+; SRC-NEXT: xorq $63, %rcx
+; SRC-NEXT: bsrq %r9, %rdx
+; SRC-NEXT: xorq $63, %rdx
+; SRC-NEXT: addq $64, %rdx
+; SRC-NEXT: testq %rdi, %rdi
+; SRC-NEXT: cmovneq %rcx, %rdx
+; SRC-NEXT: bsrq %r10, %rcx
+; SRC-NEXT: xorq $63, %rcx
+; SRC-NEXT: bsrq %r11, %r8
+; SRC-NEXT: movl $127, %esi
+; SRC-NEXT: cmovneq %r8, %rsi
+; SRC-NEXT: xorq $63, %rsi
+; SRC-NEXT: addq $64, %rsi
+; SRC-NEXT: testq %r10, %r10
+; SRC-NEXT: cmovneq %rcx, %rsi
+; SRC-NEXT: subq $-128, %rsi
+; SRC-NEXT: orq %r9, %rdi
+; SRC-NEXT: cmovneq %rdx, %rsi
+; SRC-NEXT: movq %rsi, (%rax)
+; SRC-NEXT: movq $0, 24(%rax)
+; SRC-NEXT: movq $0, 16(%rax)
+; SRC-NEXT: movq $0, 8(%rax)
+; SRC-NEXT: retq
+;
+; LIN-LABEL: test2:
+; LIN: # %bb.0:
+; LIN-NEXT: movq %rdi, %rax
+; LIN-NEXT: movq %rsi, %rdi
+; LIN-NEXT: negq %rdi
+; LIN-NEXT: andq %rsi, %rdi
+; LIN-NEXT: bsrq %rdi, %rsi
+; LIN-NEXT: movl $127, %edi
+; LIN-NEXT: cmovneq %rsi, %rdi
+; LIN-NEXT: xorq $63, %rdi
+; LIN-NEXT: addq $64, %rdi
+; LIN-NEXT: xorl %r9d, %r9d
+; LIN-NEXT: movl $0, %esi
+; LIN-NEXT: sbbq %rdx, %rsi
+; LIN-NEXT: andq %rdx, %rsi
+; LIN-NEXT: bsrq %rsi, %rdx
+; LIN-NEXT: xorq $63, %rdx
+; LIN-NEXT: testq %rsi, %rsi
+; LIN-NEXT: cmoveq %rdi, %rdx
+; LIN-NEXT: subq $-128, %rdx
+; LIN-NEXT: movl $0, %esi
+; LIN-NEXT: sbbq %rcx, %rsi
+; LIN-NEXT: andq %rcx, %rsi
+; LIN-NEXT: bsrq %rsi, %rcx
+; LIN-NEXT: xorq $63, %rcx
+; LIN-NEXT: addq $64, %rcx
+; LIN-NEXT: sbbq %r8, %r9
+; LIN-NEXT: andq %r8, %r9
+; LIN-NEXT: bsrq %r9, %rdi
+; LIN-NEXT: xorq $63, %rdi
+; LIN-NEXT: testq %r9, %r9
+; LIN-NEXT: cmoveq %rcx, %rdi
+; LIN-NEXT: orq %rsi, %r9
+; LIN-NEXT: cmoveq %rdx, %rdi
+; LIN-NEXT: movq %rdi, (%rax)
+; LIN-NEXT: movq $0, 8(%rax)
+; LIN-NEXT: movq $0, 16(%rax)
+; LIN-NEXT: movq $0, 24(%rax)
+; LIN-NEXT: retq
%b = sub i256 0, %a
%c = and i256 %b, %a
%d = call i256 @llvm.ctlz.i256(i256 %c, i1 false)
ret i256 %d
}
-; CHECK-LABEL: test3
-define i256 @test3(i256 %n) {
+define i256 @test3(i256 %n) nounwind {
+; ILP-LABEL: test3:
+; ILP: # %bb.0:
+; ILP-NEXT: movq %rdi, %rax
+; ILP-NEXT: xorl %r10d, %r10d
+; ILP-NEXT: movq %rsi, %r9
+; ILP-NEXT: negq %r9
+; ILP-NEXT: movl $0, %r11d
+; ILP-NEXT: sbbq %rdx, %r11
+; ILP-NEXT: movl $0, %edi
+; ILP-NEXT: sbbq %rcx, %rdi
+; ILP-NEXT: sbbq %r8, %r10
+; ILP-NEXT: notq %rcx
+; ILP-NEXT: andq %rdi, %rcx
+; ILP-NEXT: bsrq %rcx, %rdi
+; ILP-NEXT: notq %rdx
+; ILP-NEXT: andq %r11, %rdx
+; ILP-NEXT: xorq $63, %rdi
+; ILP-NEXT: notq %r8
+; ILP-NEXT: andq %r10, %r8
+; ILP-NEXT: bsrq %r8, %r10
+; ILP-NEXT: xorq $63, %r10
+; ILP-NEXT: addq $64, %rdi
+; ILP-NEXT: bsrq %rdx, %r11
+; ILP-NEXT: notq %rsi
+; ILP-NEXT: testq %r8, %r8
+; ILP-NEXT: movq $0, 24(%rax)
+; ILP-NEXT: movq $0, 16(%rax)
+; ILP-NEXT: movq $0, 8(%rax)
+; ILP-NEXT: cmovneq %r10, %rdi
+; ILP-NEXT: xorq $63, %r11
+; ILP-NEXT: andq %r9, %rsi
+; ILP-NEXT: movl $127, %r9d
+; ILP-NEXT: bsrq %rsi, %rsi
+; ILP-NEXT: cmoveq %r9, %rsi
+; ILP-NEXT: xorq $63, %rsi
+; ILP-NEXT: addq $64, %rsi
+; ILP-NEXT: testq %rdx, %rdx
+; ILP-NEXT: cmovneq %r11, %rsi
+; ILP-NEXT: subq $-128, %rsi
+; ILP-NEXT: orq %rcx, %r8
+; ILP-NEXT: cmovneq %rdi, %rsi
+; ILP-NEXT: movq %rsi, (%rax)
+; ILP-NEXT: retq
+;
+; HYBRID-LABEL: test3:
+; HYBRID: # %bb.0:
+; HYBRID-NEXT: pushq %rbx
+; HYBRID-NEXT: movq %rdi, %rax
+; HYBRID-NEXT: xorl %edi, %edi
+; HYBRID-NEXT: movq %rsi, %r9
+; HYBRID-NEXT: negq %r9
+; HYBRID-NEXT: movl $0, %r10d
+; HYBRID-NEXT: sbbq %rdx, %r10
+; HYBRID-NEXT: movl $0, %r11d
+; HYBRID-NEXT: sbbq %rcx, %r11
+; HYBRID-NEXT: sbbq %r8, %rdi
+; HYBRID-NEXT: notq %r8
+; HYBRID-NEXT: andq %rdi, %r8
+; HYBRID-NEXT: bsrq %r8, %rbx
+; HYBRID-NEXT: xorq $63, %rbx
+; HYBRID-NEXT: notq %rcx
+; HYBRID-NEXT: andq %r11, %rcx
+; HYBRID-NEXT: bsrq %rcx, %rdi
+; HYBRID-NEXT: xorq $63, %rdi
+; HYBRID-NEXT: addq $64, %rdi
+; HYBRID-NEXT: testq %r8, %r8
+; HYBRID-NEXT: cmovneq %rbx, %rdi
+; HYBRID-NEXT: notq %rdx
+; HYBRID-NEXT: andq %r10, %rdx
+; HYBRID-NEXT: bsrq %rdx, %rbx
+; HYBRID-NEXT: xorq $63, %rbx
+; HYBRID-NEXT: notq %rsi
+; HYBRID-NEXT: andq %r9, %rsi
+; HYBRID-NEXT: movl $127, %r9d
+; HYBRID-NEXT: bsrq %rsi, %rsi
+; HYBRID-NEXT: cmoveq %r9, %rsi
+; HYBRID-NEXT: xorq $63, %rsi
+; HYBRID-NEXT: addq $64, %rsi
+; HYBRID-NEXT: testq %rdx, %rdx
+; HYBRID-NEXT: cmovneq %rbx, %rsi
+; HYBRID-NEXT: subq $-128, %rsi
+; HYBRID-NEXT: orq %r8, %rcx
+; HYBRID-NEXT: cmovneq %rdi, %rsi
+; HYBRID-NEXT: movq %rsi, (%rax)
+; HYBRID-NEXT: movq $0, 24(%rax)
+; HYBRID-NEXT: movq $0, 16(%rax)
+; HYBRID-NEXT: movq $0, 8(%rax)
+; HYBRID-NEXT: popq %rbx
+; HYBRID-NEXT: retq
+;
+; BURR-LABEL: test3:
+; BURR: # %bb.0:
+; BURR-NEXT: pushq %rbx
+; BURR-NEXT: movq %rdi, %rax
+; BURR-NEXT: xorl %edi, %edi
+; BURR-NEXT: movq %rsi, %r9
+; BURR-NEXT: negq %r9
+; BURR-NEXT: movl $0, %r10d
+; BURR-NEXT: sbbq %rdx, %r10
+; BURR-NEXT: movl $0, %r11d
+; BURR-NEXT: sbbq %rcx, %r11
+; BURR-NEXT: sbbq %r8, %rdi
+; BURR-NEXT: notq %r8
+; BURR-NEXT: andq %rdi, %r8
+; BURR-NEXT: bsrq %r8, %rbx
+; BURR-NEXT: xorq $63, %rbx
+; BURR-NEXT: notq %rcx
+; BURR-NEXT: andq %r11, %rcx
+; BURR-NEXT: bsrq %rcx, %rdi
+; BURR-NEXT: xorq $63, %rdi
+; BURR-NEXT: addq $64, %rdi
+; BURR-NEXT: testq %r8, %r8
+; BURR-NEXT: cmovneq %rbx, %rdi
+; BURR-NEXT: notq %rdx
+; BURR-NEXT: andq %r10, %rdx
+; BURR-NEXT: bsrq %rdx, %rbx
+; BURR-NEXT: xorq $63, %rbx
+; BURR-NEXT: notq %rsi
+; BURR-NEXT: andq %r9, %rsi
+; BURR-NEXT: movl $127, %r9d
+; BURR-NEXT: bsrq %rsi, %rsi
+; BURR-NEXT: cmoveq %r9, %rsi
+; BURR-NEXT: xorq $63, %rsi
+; BURR-NEXT: addq $64, %rsi
+; BURR-NEXT: testq %rdx, %rdx
+; BURR-NEXT: cmovneq %rbx, %rsi
+; BURR-NEXT: subq $-128, %rsi
+; BURR-NEXT: orq %r8, %rcx
+; BURR-NEXT: cmovneq %rdi, %rsi
+; BURR-NEXT: movq %rsi, (%rax)
+; BURR-NEXT: movq $0, 24(%rax)
+; BURR-NEXT: movq $0, 16(%rax)
+; BURR-NEXT: movq $0, 8(%rax)
+; BURR-NEXT: popq %rbx
+; BURR-NEXT: retq
+;
+; SRC-LABEL: test3:
+; SRC: # %bb.0:
+; SRC-NEXT: movq %rdi, %rax
+; SRC-NEXT: movq %rsi, %r9
+; SRC-NEXT: notq %r9
+; SRC-NEXT: xorl %r10d, %r10d
+; SRC-NEXT: negq %rsi
+; SRC-NEXT: movl $0, %r11d
+; SRC-NEXT: sbbq %rdx, %r11
+; SRC-NEXT: notq %rdx
+; SRC-NEXT: movl $0, %edi
+; SRC-NEXT: sbbq %rcx, %rdi
+; SRC-NEXT: notq %rcx
+; SRC-NEXT: sbbq %r8, %r10
+; SRC-NEXT: notq %r8
+; SRC-NEXT: andq %r11, %rdx
+; SRC-NEXT: andq %rdi, %rcx
+; SRC-NEXT: andq %r10, %r8
+; SRC-NEXT: andq %r9, %rsi
+; SRC-NEXT: bsrq %r8, %r9
+; SRC-NEXT: xorq $63, %r9
+; SRC-NEXT: bsrq %rcx, %rdi
+; SRC-NEXT: xorq $63, %rdi
+; SRC-NEXT: addq $64, %rdi
+; SRC-NEXT: testq %r8, %r8
+; SRC-NEXT: cmovneq %r9, %rdi
+; SRC-NEXT: bsrq %rdx, %r9
+; SRC-NEXT: xorq $63, %r9
+; SRC-NEXT: bsrq %rsi, %r10
+; SRC-NEXT: movl $127, %esi
+; SRC-NEXT: cmovneq %r10, %rsi
+; SRC-NEXT: xorq $63, %rsi
+; SRC-NEXT: addq $64, %rsi
+; SRC-NEXT: testq %rdx, %rdx
+; SRC-NEXT: cmovneq %r9, %rsi
+; SRC-NEXT: subq $-128, %rsi
+; SRC-NEXT: orq %rcx, %r8
+; SRC-NEXT: cmovneq %rdi, %rsi
+; SRC-NEXT: movq %rsi, (%rax)
+; SRC-NEXT: movq $0, 24(%rax)
+; SRC-NEXT: movq $0, 16(%rax)
+; SRC-NEXT: movq $0, 8(%rax)
+; SRC-NEXT: retq
+;
+; LIN-LABEL: test3:
+; LIN: # %bb.0:
+; LIN-NEXT: movq %rdi, %rax
+; LIN-NEXT: movq %rsi, %rdi
+; LIN-NEXT: negq %rdi
+; LIN-NEXT: notq %rsi
+; LIN-NEXT: andq %rdi, %rsi
+; LIN-NEXT: bsrq %rsi, %rsi
+; LIN-NEXT: movl $127, %edi
+; LIN-NEXT: cmovneq %rsi, %rdi
+; LIN-NEXT: xorq $63, %rdi
+; LIN-NEXT: addq $64, %rdi
+; LIN-NEXT: xorl %r9d, %r9d
+; LIN-NEXT: movl $0, %esi
+; LIN-NEXT: sbbq %rdx, %rsi
+; LIN-NEXT: notq %rdx
+; LIN-NEXT: andq %rsi, %rdx
+; LIN-NEXT: bsrq %rdx, %rsi
+; LIN-NEXT: xorq $63, %rsi
+; LIN-NEXT: testq %rdx, %rdx
+; LIN-NEXT: cmoveq %rdi, %rsi
+; LIN-NEXT: subq $-128, %rsi
+; LIN-NEXT: movl $0, %edx
+; LIN-NEXT: sbbq %rcx, %rdx
+; LIN-NEXT: notq %rcx
+; LIN-NEXT: andq %rdx, %rcx
+; LIN-NEXT: bsrq %rcx, %rdx
+; LIN-NEXT: xorq $63, %rdx
+; LIN-NEXT: addq $64, %rdx
+; LIN-NEXT: sbbq %r8, %r9
+; LIN-NEXT: notq %r8
+; LIN-NEXT: andq %r9, %r8
+; LIN-NEXT: bsrq %r8, %rdi
+; LIN-NEXT: xorq $63, %rdi
+; LIN-NEXT: testq %r8, %r8
+; LIN-NEXT: cmoveq %rdx, %rdi
+; LIN-NEXT: orq %rcx, %r8
+; LIN-NEXT: cmoveq %rsi, %rdi
+; LIN-NEXT: movq %rdi, (%rax)
+; LIN-NEXT: movq $0, 8(%rax)
+; LIN-NEXT: movq $0, 16(%rax)
+; LIN-NEXT: movq $0, 24(%rax)
+; LIN-NEXT: retq
%m = sub i256 -1, %n
%x = sub i256 0, %n
%y = and i256 %x, %m
@@ -38,8 +905,91 @@ define i256 @test3(i256 %n) {
declare i256 @llvm.ctlz.i256(i256, i1) nounwind readnone
-; CHECK-LABEL: test4
-define i64 @test4(i64 %a, i64 %b) {
+define i64 @test4(i64 %a, i64 %b) nounwind {
+; ILP-LABEL: test4:
+; ILP: # %bb.0:
+; ILP-NEXT: xorl %ecx, %ecx
+; ILP-NEXT: xorl %edx, %edx
+; ILP-NEXT: addq $1, %rsi
+; ILP-NEXT: setb %dl
+; ILP-NEXT: movl $2, %eax
+; ILP-NEXT: cmpq %rdi, %rsi
+; ILP-NEXT: sbbq $0, %rdx
+; ILP-NEXT: movl $0, %edx
+; ILP-NEXT: sbbq $0, %rdx
+; ILP-NEXT: sbbq $0, %rcx
+; ILP-NEXT: setae %cl
+; ILP-NEXT: movzbl %cl, %ecx
+; ILP-NEXT: subq %rcx, %rax
+; ILP-NEXT: retq
+;
+; HYBRID-LABEL: test4:
+; HYBRID: # %bb.0:
+; HYBRID-NEXT: xorl %eax, %eax
+; HYBRID-NEXT: xorl %ecx, %ecx
+; HYBRID-NEXT: addq $1, %rsi
+; HYBRID-NEXT: setb %cl
+; HYBRID-NEXT: cmpq %rdi, %rsi
+; HYBRID-NEXT: sbbq $0, %rcx
+; HYBRID-NEXT: movl $0, %ecx
+; HYBRID-NEXT: sbbq $0, %rcx
+; HYBRID-NEXT: sbbq $0, %rax
+; HYBRID-NEXT: setae %al
+; HYBRID-NEXT: movzbl %al, %ecx
+; HYBRID-NEXT: movl $2, %eax
+; HYBRID-NEXT: subq %rcx, %rax
+; HYBRID-NEXT: retq
+;
+; BURR-LABEL: test4:
+; BURR: # %bb.0:
+; BURR-NEXT: xorl %eax, %eax
+; BURR-NEXT: xorl %ecx, %ecx
+; BURR-NEXT: addq $1, %rsi
+; BURR-NEXT: setb %cl
+; BURR-NEXT: cmpq %rdi, %rsi
+; BURR-NEXT: sbbq $0, %rcx
+; BURR-NEXT: movl $0, %ecx
+; BURR-NEXT: sbbq $0, %rcx
+; BURR-NEXT: sbbq $0, %rax
+; BURR-NEXT: setae %al
+; BURR-NEXT: movzbl %al, %ecx
+; BURR-NEXT: movl $2, %eax
+; BURR-NEXT: subq %rcx, %rax
+; BURR-NEXT: retq
+;
+; SRC-LABEL: test4:
+; SRC: # %bb.0:
+; SRC-NEXT: xorl %eax, %eax
+; SRC-NEXT: addq $1, %rsi
+; SRC-NEXT: setb %al
+; SRC-NEXT: xorl %ecx, %ecx
+; SRC-NEXT: cmpq %rdi, %rsi
+; SRC-NEXT: sbbq $0, %rax
+; SRC-NEXT: movl $0, %eax
+; SRC-NEXT: sbbq $0, %rax
+; SRC-NEXT: sbbq $0, %rcx
+; SRC-NEXT: setae %al
+; SRC-NEXT: movzbl %al, %ecx
+; SRC-NEXT: movl $2, %eax
+; SRC-NEXT: subq %rcx, %rax
+; SRC-NEXT: retq
+;
+; LIN-LABEL: test4:
+; LIN: # %bb.0:
+; LIN-NEXT: movl $2, %eax
+; LIN-NEXT: xorl %ecx, %ecx
+; LIN-NEXT: xorl %edx, %edx
+; LIN-NEXT: addq $1, %rsi
+; LIN-NEXT: setb %dl
+; LIN-NEXT: cmpq %rdi, %rsi
+; LIN-NEXT: sbbq $0, %rdx
+; LIN-NEXT: movl $0, %edx
+; LIN-NEXT: sbbq $0, %rdx
+; LIN-NEXT: sbbq $0, %rcx
+; LIN-NEXT: setae %cl
+; LIN-NEXT: movzbl %cl, %ecx
+; LIN-NEXT: subq %rcx, %rax
+; LIN-NEXT: retq
%r = zext i64 %b to i256
%u = add i256 %r, 1
%w = and i256 %u, 1461501637330902918203684832716283019655932542975
@@ -49,3 +999,252 @@ define i64 @test4(i64 %a, i64 %b) {
%z = add i64 %y, 1
ret i64 %z
}
+
+define i256 @PR25498(i256 %a) nounwind {
+; ILP-LABEL: PR25498:
+; ILP: # %bb.0:
+; ILP-NEXT: pushq %rbx
+; ILP-NEXT: movq %rdi, %rax
+; ILP-NEXT: xorl %r9d, %r9d
+; ILP-NEXT: movq %rsi, %rbx
+; ILP-NEXT: negq %rbx
+; ILP-NEXT: movl $0, %r11d
+; ILP-NEXT: sbbq %rdx, %r11
+; ILP-NEXT: movl $0, %r10d
+; ILP-NEXT: sbbq %rcx, %r10
+; ILP-NEXT: movl $0, %edi
+; ILP-NEXT: sbbq %r8, %rdi
+; ILP-NEXT: orq %r8, %rdx
+; ILP-NEXT: orq %rcx, %rsi
+; ILP-NEXT: orq %rdx, %rsi
+; ILP-NEXT: je .LBB4_1
+; ILP-NEXT: # %bb.2: # %cond.false
+; ILP-NEXT: bsrq %r11, %rdx
+; ILP-NEXT: bsrq %rdi, %rcx
+; ILP-NEXT: xorq $63, %rcx
+; ILP-NEXT: bsrq %r10, %rsi
+; ILP-NEXT: xorq $63, %rsi
+; ILP-NEXT: addq $64, %rsi
+; ILP-NEXT: testq %rdi, %rdi
+; ILP-NEXT: cmovneq %rcx, %rsi
+; ILP-NEXT: xorq $63, %rdx
+; ILP-NEXT: bsrq %rbx, %rcx
+; ILP-NEXT: xorq $63, %rcx
+; ILP-NEXT: addq $64, %rcx
+; ILP-NEXT: testq %r11, %r11
+; ILP-NEXT: cmovneq %rdx, %rcx
+; ILP-NEXT: subq $-128, %rcx
+; ILP-NEXT: xorl %r9d, %r9d
+; ILP-NEXT: orq %rdi, %r10
+; ILP-NEXT: cmovneq %rsi, %rcx
+; ILP-NEXT: jmp .LBB4_3
+; ILP-NEXT: .LBB4_1:
+; ILP-NEXT: movl $256, %ecx # imm = 0x100
+; ILP-NEXT: .LBB4_3: # %cond.end
+; ILP-NEXT: movq %rcx, (%rax)
+; ILP-NEXT: movq %r9, 8(%rax)
+; ILP-NEXT: movq %r9, 16(%rax)
+; ILP-NEXT: movq %r9, 24(%rax)
+; ILP-NEXT: popq %rbx
+; ILP-NEXT: retq
+;
+; HYBRID-LABEL: PR25498:
+; HYBRID: # %bb.0:
+; HYBRID-NEXT: pushq %rbx
+; HYBRID-NEXT: movq %rdi, %rax
+; HYBRID-NEXT: xorl %r9d, %r9d
+; HYBRID-NEXT: movq %rsi, %rbx
+; HYBRID-NEXT: negq %rbx
+; HYBRID-NEXT: movl $0, %r11d
+; HYBRID-NEXT: sbbq %rdx, %r11
+; HYBRID-NEXT: movl $0, %r10d
+; HYBRID-NEXT: sbbq %rcx, %r10
+; HYBRID-NEXT: movl $0, %edi
+; HYBRID-NEXT: sbbq %r8, %rdi
+; HYBRID-NEXT: orq %r8, %rdx
+; HYBRID-NEXT: orq %rcx, %rsi
+; HYBRID-NEXT: orq %rdx, %rsi
+; HYBRID-NEXT: je .LBB4_1
+; HYBRID-NEXT: # %bb.2: # %cond.false
+; HYBRID-NEXT: bsrq %rdi, %rcx
+; HYBRID-NEXT: xorq $63, %rcx
+; HYBRID-NEXT: bsrq %r10, %rdx
+; HYBRID-NEXT: xorq $63, %rdx
+; HYBRID-NEXT: addq $64, %rdx
+; HYBRID-NEXT: testq %rdi, %rdi
+; HYBRID-NEXT: cmovneq %rcx, %rdx
+; HYBRID-NEXT: bsrq %r11, %rsi
+; HYBRID-NEXT: xorq $63, %rsi
+; HYBRID-NEXT: bsrq %rbx, %rcx
+; HYBRID-NEXT: xorq $63, %rcx
+; HYBRID-NEXT: addq $64, %rcx
+; HYBRID-NEXT: testq %r11, %r11
+; HYBRID-NEXT: cmovneq %rsi, %rcx
+; HYBRID-NEXT: subq $-128, %rcx
+; HYBRID-NEXT: orq %rdi, %r10
+; HYBRID-NEXT: cmovneq %rdx, %rcx
+; HYBRID-NEXT: xorl %r9d, %r9d
+; HYBRID-NEXT: jmp .LBB4_3
+; HYBRID-NEXT: .LBB4_1:
+; HYBRID-NEXT: movl $256, %ecx # imm = 0x100
+; HYBRID-NEXT: .LBB4_3: # %cond.end
+; HYBRID-NEXT: movq %rcx, (%rax)
+; HYBRID-NEXT: movq %r9, 8(%rax)
+; HYBRID-NEXT: movq %r9, 16(%rax)
+; HYBRID-NEXT: movq %r9, 24(%rax)
+; HYBRID-NEXT: popq %rbx
+; HYBRID-NEXT: retq
+;
+; BURR-LABEL: PR25498:
+; BURR: # %bb.0:
+; BURR-NEXT: pushq %rbx
+; BURR-NEXT: movq %rdi, %rax
+; BURR-NEXT: xorl %r9d, %r9d
+; BURR-NEXT: movq %rsi, %rbx
+; BURR-NEXT: negq %rbx
+; BURR-NEXT: movl $0, %r11d
+; BURR-NEXT: sbbq %rdx, %r11
+; BURR-NEXT: movl $0, %r10d
+; BURR-NEXT: sbbq %rcx, %r10
+; BURR-NEXT: movl $0, %edi
+; BURR-NEXT: sbbq %r8, %rdi
+; BURR-NEXT: orq %r8, %rdx
+; BURR-NEXT: orq %rcx, %rsi
+; BURR-NEXT: orq %rdx, %rsi
+; BURR-NEXT: je .LBB4_1
+; BURR-NEXT: # %bb.2: # %cond.false
+; BURR-NEXT: bsrq %rdi, %rcx
+; BURR-NEXT: xorq $63, %rcx
+; BURR-NEXT: bsrq %r10, %rdx
+; BURR-NEXT: xorq $63, %rdx
+; BURR-NEXT: addq $64, %rdx
+; BURR-NEXT: testq %rdi, %rdi
+; BURR-NEXT: cmovneq %rcx, %rdx
+; BURR-NEXT: bsrq %r11, %rsi
+; BURR-NEXT: xorq $63, %rsi
+; BURR-NEXT: bsrq %rbx, %rcx
+; BURR-NEXT: xorq $63, %rcx
+; BURR-NEXT: addq $64, %rcx
+; BURR-NEXT: testq %r11, %r11
+; BURR-NEXT: cmovneq %rsi, %rcx
+; BURR-NEXT: subq $-128, %rcx
+; BURR-NEXT: orq %rdi, %r10
+; BURR-NEXT: cmovneq %rdx, %rcx
+; BURR-NEXT: xorl %r9d, %r9d
+; BURR-NEXT: jmp .LBB4_3
+; BURR-NEXT: .LBB4_1:
+; BURR-NEXT: movl $256, %ecx # imm = 0x100
+; BURR-NEXT: .LBB4_3: # %cond.end
+; BURR-NEXT: movq %rcx, (%rax)
+; BURR-NEXT: movq %r9, 8(%rax)
+; BURR-NEXT: movq %r9, 16(%rax)
+; BURR-NEXT: movq %r9, 24(%rax)
+; BURR-NEXT: popq %rbx
+; BURR-NEXT: retq
+;
+; SRC-LABEL: PR25498:
+; SRC: # %bb.0:
+; SRC-NEXT: pushq %rbx
+; SRC-NEXT: movq %rdi, %rax
+; SRC-NEXT: xorl %r9d, %r9d
+; SRC-NEXT: movq %rsi, %rbx
+; SRC-NEXT: negq %rbx
+; SRC-NEXT: movl $0, %r11d
+; SRC-NEXT: sbbq %rdx, %r11
+; SRC-NEXT: movl $0, %r10d
+; SRC-NEXT: sbbq %rcx, %r10
+; SRC-NEXT: movl $0, %edi
+; SRC-NEXT: sbbq %r8, %rdi
+; SRC-NEXT: orq %r8, %rdx
+; SRC-NEXT: orq %rcx, %rsi
+; SRC-NEXT: orq %rdx, %rsi
+; SRC-NEXT: je .LBB4_1
+; SRC-NEXT: # %bb.2: # %cond.false
+; SRC-NEXT: bsrq %rdi, %rcx
+; SRC-NEXT: xorq $63, %rcx
+; SRC-NEXT: bsrq %r10, %rdx
+; SRC-NEXT: xorq $63, %rdx
+; SRC-NEXT: addq $64, %rdx
+; SRC-NEXT: testq %rdi, %rdi
+; SRC-NEXT: cmovneq %rcx, %rdx
+; SRC-NEXT: bsrq %r11, %rsi
+; SRC-NEXT: xorq $63, %rsi
+; SRC-NEXT: bsrq %rbx, %rcx
+; SRC-NEXT: xorq $63, %rcx
+; SRC-NEXT: addq $64, %rcx
+; SRC-NEXT: testq %r11, %r11
+; SRC-NEXT: cmovneq %rsi, %rcx
+; SRC-NEXT: subq $-128, %rcx
+; SRC-NEXT: orq %rdi, %r10
+; SRC-NEXT: cmovneq %rdx, %rcx
+; SRC-NEXT: xorl %r9d, %r9d
+; SRC-NEXT: jmp .LBB4_3
+; SRC-NEXT: .LBB4_1:
+; SRC-NEXT: movl $256, %ecx # imm = 0x100
+; SRC-NEXT: .LBB4_3: # %cond.end
+; SRC-NEXT: movq %rcx, (%rax)
+; SRC-NEXT: movq %r9, 8(%rax)
+; SRC-NEXT: movq %r9, 16(%rax)
+; SRC-NEXT: movq %r9, 24(%rax)
+; SRC-NEXT: popq %rbx
+; SRC-NEXT: retq
+;
+; LIN-LABEL: PR25498:
+; LIN: # %bb.0:
+; LIN-NEXT: pushq %rbx
+; LIN-NEXT: movq %rdi, %rax
+; LIN-NEXT: movq %rsi, %rbx
+; LIN-NEXT: negq %rbx
+; LIN-NEXT: xorl %r9d, %r9d
+; LIN-NEXT: movl $0, %edi
+; LIN-NEXT: sbbq %rdx, %rdi
+; LIN-NEXT: movl $0, %r10d
+; LIN-NEXT: sbbq %rcx, %r10
+; LIN-NEXT: movl $0, %r11d
+; LIN-NEXT: sbbq %r8, %r11
+; LIN-NEXT: orq %rcx, %rsi
+; LIN-NEXT: orq %r8, %rdx
+; LIN-NEXT: orq %rsi, %rdx
+; LIN-NEXT: je .LBB4_1
+; LIN-NEXT: # %bb.2: # %cond.false
+; LIN-NEXT: bsrq %rbx, %rcx
+; LIN-NEXT: xorq $63, %rcx
+; LIN-NEXT: addq $64, %rcx
+; LIN-NEXT: bsrq %rdi, %rdx
+; LIN-NEXT: xorq $63, %rdx
+; LIN-NEXT: testq %rdi, %rdi
+; LIN-NEXT: cmoveq %rcx, %rdx
+; LIN-NEXT: subq $-128, %rdx
+; LIN-NEXT: bsrq %r10, %rsi
+; LIN-NEXT: xorq $63, %rsi
+; LIN-NEXT: addq $64, %rsi
+; LIN-NEXT: bsrq %r11, %rcx
+; LIN-NEXT: xorq $63, %rcx
+; LIN-NEXT: testq %r11, %r11
+; LIN-NEXT: cmoveq %rsi, %rcx
+; LIN-NEXT: orq %r11, %r10
+; LIN-NEXT: cmoveq %rdx, %rcx
+; LIN-NEXT: xorl %r9d, %r9d
+; LIN-NEXT: jmp .LBB4_3
+; LIN-NEXT: .LBB4_1:
+; LIN-NEXT: movl $256, %ecx # imm = 0x100
+; LIN-NEXT: .LBB4_3: # %cond.end
+; LIN-NEXT: movq %rcx, (%rax)
+; LIN-NEXT: movq %r9, 8(%rax)
+; LIN-NEXT: movq %r9, 16(%rax)
+; LIN-NEXT: movq %r9, 24(%rax)
+; LIN-NEXT: popq %rbx
+; LIN-NEXT: retq
+ %b = sub i256 0, %a
+ %cmpz = icmp eq i256 %b, 0
+ br i1 %cmpz, label %cond.end, label %cond.false
+
+cond.false:
+ %d = call i256 @llvm.ctlz.i256(i256 %b, i1 true)
+ br label %cond.end
+
+cond.end:
+ %ctz = phi i256 [ 256, %0 ], [ %d, %cond.false ]
+ ret i256 %ctz
+}
+
More information about the llvm-commits
mailing list