[llvm] r272508 - [X86][BMI] Added fast-isel tests for BMI1 intrinsics

Simon Pilgrim via llvm-commits llvm-commits at lists.llvm.org
Sun Jun 12 02:56:05 PDT 2016


Author: rksimon
Date: Sun Jun 12 04:56:05 2016
New Revision: 272508

URL: http://llvm.org/viewvc/llvm-project?rev=272508&view=rev
Log:
[X86][BMI] Added fast-isel tests for BMI1 intrinsics

A lot of the codegen is pretty awful for these as they are mostly implemented as generic bit twiddling ops 

Added:
    llvm/trunk/test/CodeGen/X86/bmi-intrinsics-fast-isel-x86_64.ll
    llvm/trunk/test/CodeGen/X86/bmi-intrinsics-fast-isel.ll

Added: llvm/trunk/test/CodeGen/X86/bmi-intrinsics-fast-isel-x86_64.ll
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/X86/bmi-intrinsics-fast-isel-x86_64.ll?rev=272508&view=auto
==============================================================================
--- llvm/trunk/test/CodeGen/X86/bmi-intrinsics-fast-isel-x86_64.ll (added)
+++ llvm/trunk/test/CodeGen/X86/bmi-intrinsics-fast-isel-x86_64.ll Sun Jun 12 04:56:05 2016
@@ -0,0 +1,165 @@
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
+; RUN: llc < %s -fast-isel -mtriple=x86_64-unknown-unknown -mattr=+bmi | FileCheck %s --check-prefix=X64
+
+; NOTE: This should use IR equivalent to what is generated by clang/test/CodeGen/bmi-builtins.c
+
+;
+; AMD Intrinsics
+;
+
+define i64 @test__andn_u64(i64 %a0, i64 %a1) {
+; X64-LABEL: test__andn_u64:
+; X64:       # BB#0:
+; X64-NEXT:    xorq $-1, %rdi
+; X64-NEXT:    andq %rsi, %rdi
+; X64-NEXT:    movq %rdi, %rax
+; X64-NEXT:    retq
+  %xor = xor i64 %a0, -1
+  %res = and i64 %xor, %a1
+  ret i64 %res
+}
+
+define i64 @test__bextr_u64(i64 %a0, i64 %a1) {
+; X64-LABEL: test__bextr_u64:
+; X64:       # BB#0:
+; X64-NEXT:    bextrq %rsi, %rdi, %rax
+; X64-NEXT:    retq
+  %res = call i64 @llvm.x86.bmi.bextr.64(i64 %a0, i64 %a1)
+  ret i64 %res
+}
+
+define i64 @test__blsi_u64(i64 %a0) {
+; X64-LABEL: test__blsi_u64:
+; X64:       # BB#0:
+; X64-NEXT:    xorl %eax, %eax
+; X64-NEXT:    subq %rdi, %rax
+; X64-NEXT:    andq %rdi, %rax
+; X64-NEXT:    retq
+  %neg = sub i64 0, %a0
+  %res = and i64 %a0, %neg
+  ret i64 %res
+}
+
+define i64 @test__blsmsk_u64(i64 %a0) {
+; X64-LABEL: test__blsmsk_u64:
+; X64:       # BB#0:
+; X64-NEXT:    movq %rdi, %rax
+; X64-NEXT:    subq $1, %rax
+; X64-NEXT:    xorq %rdi, %rax
+; X64-NEXT:    retq
+  %dec = sub i64 %a0, 1
+  %res = xor i64 %a0, %dec
+  ret i64 %res
+}
+
+define i64 @test__blsr_u64(i64 %a0) {
+; X64-LABEL: test__blsr_u64:
+; X64:       # BB#0:
+; X64-NEXT:    movq %rdi, %rax
+; X64-NEXT:    subq $1, %rax
+; X64-NEXT:    andq %rdi, %rax
+; X64-NEXT:    retq
+  %dec = sub i64 %a0, 1
+  %res = and i64 %a0, %dec
+  ret i64 %res
+}
+
+define i64 @test__tzcnt_u64(i64 %a0) {
+; X64-LABEL: test__tzcnt_u64:
+; X64:       # BB#0:
+; X64-NEXT:    movl $64, %ecx
+; X64-NEXT:    tzcntq %rdi, %rax
+; X64-NEXT:    cmovbq %rcx, %rax
+; X64-NEXT:    retq
+  %cmp = icmp ne i64 %a0, 0
+  %cttz = call i64 @llvm.cttz.i64(i64 %a0, i1 true)
+  %res = select i1 %cmp, i64 %cttz, i64 64
+  ret i64 %res
+}
+
+;
+; Intel intrinsics
+;
+
+define i64 @test_andn_u64(i64 %a0, i64 %a1) {
+; X64-LABEL: test_andn_u64:
+; X64:       # BB#0:
+; X64-NEXT:    xorq $-1, %rdi
+; X64-NEXT:    andq %rsi, %rdi
+; X64-NEXT:    movq %rdi, %rax
+; X64-NEXT:    retq
+  %xor = xor i64 %a0, -1
+  %res = and i64 %xor, %a1
+  ret i64 %res
+}
+
+define i64 @test_bextr_u64(i64 %a0, i32 %a1, i32 %a2) {
+; X64-LABEL: test_bextr_u64:
+; X64:       # BB#0:
+; X64-NEXT:    andl $255, %esi
+; X64-NEXT:    andl $255, %edx
+; X64-NEXT:    shll $8, %edx
+; X64-NEXT:    orl %esi, %edx
+; X64-NEXT:    movl %edx, %eax
+; X64-NEXT:    bextrq %rax, %rdi, %rax
+; X64-NEXT:    retq
+  %and1 = and i32 %a1, 255
+  %and2 = and i32 %a2, 255
+  %shl = shl i32 %and2, 8
+  %or = or i32 %and1, %shl
+  %zext = zext i32 %or to i64
+  %res = call i64 @llvm.x86.bmi.bextr.64(i64 %a0, i64 %zext)
+  ret i64 %res
+}
+
+define i64 @test_blsi_u64(i64 %a0) {
+; X64-LABEL: test_blsi_u64:
+; X64:       # BB#0:
+; X64-NEXT:    xorl %eax, %eax
+; X64-NEXT:    subq %rdi, %rax
+; X64-NEXT:    andq %rdi, %rax
+; X64-NEXT:    retq
+  %neg = sub i64 0, %a0
+  %res = and i64 %a0, %neg
+  ret i64 %res
+}
+
+define i64 @test_blsmsk_u64(i64 %a0) {
+; X64-LABEL: test_blsmsk_u64:
+; X64:       # BB#0:
+; X64-NEXT:    movq %rdi, %rax
+; X64-NEXT:    subq $1, %rax
+; X64-NEXT:    xorq %rdi, %rax
+; X64-NEXT:    retq
+  %dec = sub i64 %a0, 1
+  %res = xor i64 %a0, %dec
+  ret i64 %res
+}
+
+define i64 @test_blsr_u64(i64 %a0) {
+; X64-LABEL: test_blsr_u64:
+; X64:       # BB#0:
+; X64-NEXT:    movq %rdi, %rax
+; X64-NEXT:    subq $1, %rax
+; X64-NEXT:    andq %rdi, %rax
+; X64-NEXT:    retq
+  %dec = sub i64 %a0, 1
+  %res = and i64 %a0, %dec
+  ret i64 %res
+}
+
+define i64 @test_tzcnt_u64(i64 %a0) {
+; X64-LABEL: test_tzcnt_u64:
+; X64:       # BB#0:
+; X64-NEXT:    movl $64, %ecx
+; X64-NEXT:    tzcntq %rdi, %rax
+; X64-NEXT:    cmovbq %rcx, %rax
+; X64-NEXT:    retq
+  %cmp = icmp ne i64 %a0, 0
+  %cttz = call i64 @llvm.cttz.i64(i64 %a0, i1 true)
+  %res = select i1 %cmp, i64 %cttz, i64 64
+  ret i64 %res
+}
+
+declare i64 @llvm.cttz.i64(i64, i1)
+declare i64 @llvm.x86.bmi.bextr.64(i64, i64)

Added: llvm/trunk/test/CodeGen/X86/bmi-intrinsics-fast-isel.ll
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/X86/bmi-intrinsics-fast-isel.ll?rev=272508&view=auto
==============================================================================
--- llvm/trunk/test/CodeGen/X86/bmi-intrinsics-fast-isel.ll (added)
+++ llvm/trunk/test/CodeGen/X86/bmi-intrinsics-fast-isel.ll Sun Jun 12 04:56:05 2016
@@ -0,0 +1,326 @@
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
+; RUN: llc < %s -fast-isel -mtriple=i686-unknown-unknown -mattr=+bmi | FileCheck %s --check-prefix=X32
+; RUN: llc < %s -fast-isel -mtriple=x86_64-unknown-unknown -mattr=+bmi | FileCheck %s --check-prefix=X64
+
+; NOTE: This should use IR equivalent to what is generated by clang/test/CodeGen/bmi-builtins.c
+
+;
+; AMD Intrinsics
+;
+
+define i16 @test__tzcnt_u16(i16 %a0) {
+; X32-LABEL: test__tzcnt_u16:
+; X32:       # BB#0:
+; X32-NEXT:    movzwl {{[0-9]+}}(%esp), %eax
+; X32-NEXT:    movzwl %ax, %ecx
+; X32-NEXT:    cmpl $0, %ecx
+; X32-NEXT:    jne .LBB0_1
+; X32-NEXT:  # BB#2:
+; X32-NEXT:    movw $16, %ax
+; X32-NEXT:    retl
+; X32-NEXT:  .LBB0_1:
+; X32-NEXT:    tzcntw %ax, %ax
+; X32-NEXT:    retl
+;
+; X64-LABEL: test__tzcnt_u16:
+; X64:       # BB#0:
+; X64-NEXT:    movw $16, %cx
+; X64-NEXT:    movzwl %di, %edx
+; X64-NEXT:    tzcntw %dx, %ax
+; X64-NEXT:    cmpl $0, %edx
+; X64-NEXT:    cmovew %cx, %ax
+; X64-NEXT:    retq
+  %zext = zext i16 %a0 to i32
+  %cmp = icmp ne i32 %zext, 0
+  %cttz = call i16 @llvm.cttz.i16(i16 %a0, i1 true)
+  %res = select i1 %cmp, i16 %cttz, i16 16
+  ret i16 %res
+}
+
+define i32 @test__andn_u32(i32 %a0, i32 %a1) {
+; X32-LABEL: test__andn_u32:
+; X32:       # BB#0:
+; X32-NEXT:    movl {{[0-9]+}}(%esp), %eax
+; X32-NEXT:    xorl $-1, %eax
+; X32-NEXT:    andl {{[0-9]+}}(%esp), %eax
+; X32-NEXT:    retl
+;
+; X64-LABEL: test__andn_u32:
+; X64:       # BB#0:
+; X64-NEXT:    xorl $-1, %edi
+; X64-NEXT:    andl %esi, %edi
+; X64-NEXT:    movl %edi, %eax
+; X64-NEXT:    retq
+  %xor = xor i32 %a0, -1
+  %res = and i32 %xor, %a1
+  ret i32 %res
+}
+
+define i32 @test__bextr_u32(i32 %a0, i32 %a1) {
+; X32-LABEL: test__bextr_u32:
+; X32:       # BB#0:
+; X32-NEXT:    movl {{[0-9]+}}(%esp), %eax
+; X32-NEXT:    bextrl %eax, {{[0-9]+}}(%esp), %eax
+; X32-NEXT:    retl
+;
+; X64-LABEL: test__bextr_u32:
+; X64:       # BB#0:
+; X64-NEXT:    bextrl %esi, %edi, %eax
+; X64-NEXT:    retq
+  %res = call i32 @llvm.x86.bmi.bextr.32(i32 %a0, i32 %a1)
+  ret i32 %res
+}
+
+define i32 @test__blsi_u32(i32 %a0) {
+; X32-LABEL: test__blsi_u32:
+; X32:       # BB#0:
+; X32-NEXT:    movl {{[0-9]+}}(%esp), %ecx
+; X32-NEXT:    xorl %eax, %eax
+; X32-NEXT:    subl %ecx, %eax
+; X32-NEXT:    andl %ecx, %eax
+; X32-NEXT:    retl
+;
+; X64-LABEL: test__blsi_u32:
+; X64:       # BB#0:
+; X64-NEXT:    xorl %eax, %eax
+; X64-NEXT:    subl %edi, %eax
+; X64-NEXT:    andl %edi, %eax
+; X64-NEXT:    retq
+  %neg = sub i32 0, %a0
+  %res = and i32 %a0, %neg
+  ret i32 %res
+}
+
+define i32 @test__blsmsk_u32(i32 %a0) {
+; X32-LABEL: test__blsmsk_u32:
+; X32:       # BB#0:
+; X32-NEXT:    movl {{[0-9]+}}(%esp), %ecx
+; X32-NEXT:    movl %ecx, %eax
+; X32-NEXT:    subl $1, %eax
+; X32-NEXT:    xorl %ecx, %eax
+; X32-NEXT:    retl
+;
+; X64-LABEL: test__blsmsk_u32:
+; X64:       # BB#0:
+; X64-NEXT:    movl %edi, %eax
+; X64-NEXT:    subl $1, %eax
+; X64-NEXT:    xorl %edi, %eax
+; X64-NEXT:    retq
+  %dec = sub i32 %a0, 1
+  %res = xor i32 %a0, %dec
+  ret i32 %res
+}
+
+define i32 @test__blsr_u32(i32 %a0) {
+; X32-LABEL: test__blsr_u32:
+; X32:       # BB#0:
+; X32-NEXT:    movl {{[0-9]+}}(%esp), %ecx
+; X32-NEXT:    movl %ecx, %eax
+; X32-NEXT:    subl $1, %eax
+; X32-NEXT:    andl %ecx, %eax
+; X32-NEXT:    retl
+;
+; X64-LABEL: test__blsr_u32:
+; X64:       # BB#0:
+; X64-NEXT:    movl %edi, %eax
+; X64-NEXT:    subl $1, %eax
+; X64-NEXT:    andl %edi, %eax
+; X64-NEXT:    retq
+  %dec = sub i32 %a0, 1
+  %res = and i32 %a0, %dec
+  ret i32 %res
+}
+
+define i32 @test__tzcnt_u32(i32 %a0) {
+; X32-LABEL: test__tzcnt_u32:
+; X32:       # BB#0:
+; X32-NEXT:    movl {{[0-9]+}}(%esp), %eax
+; X32-NEXT:    cmpl $0, %eax
+; X32-NEXT:    jne .LBB6_1
+; X32-NEXT:  # BB#2:
+; X32-NEXT:    movl $32, %eax
+; X32-NEXT:    retl
+; X32-NEXT:  .LBB6_1:
+; X32-NEXT:    tzcntl %eax, %eax
+; X32-NEXT:    retl
+;
+; X64-LABEL: test__tzcnt_u32:
+; X64:       # BB#0:
+; X64-NEXT:    movl $32, %ecx
+; X64-NEXT:    tzcntl %edi, %eax
+; X64-NEXT:    cmovbl %ecx, %eax
+; X64-NEXT:    retq
+  %cmp = icmp ne i32 %a0, 0
+  %cttz = call i32 @llvm.cttz.i32(i32 %a0, i1 true)
+  %res = select i1 %cmp, i32 %cttz, i32 32
+  ret i32 %res
+}
+
+;
+; Intel intrinsics
+;
+
+define i16 @test_tzcnt_u16(i16 %a0) {
+; X32-LABEL: test_tzcnt_u16:
+; X32:       # BB#0:
+; X32-NEXT:    movzwl {{[0-9]+}}(%esp), %eax
+; X32-NEXT:    movzwl %ax, %ecx
+; X32-NEXT:    cmpl $0, %ecx
+; X32-NEXT:    jne .LBB7_1
+; X32-NEXT:  # BB#2:
+; X32-NEXT:    movw $16, %ax
+; X32-NEXT:    retl
+; X32-NEXT:  .LBB7_1:
+; X32-NEXT:    tzcntw %ax, %ax
+; X32-NEXT:    retl
+;
+; X64-LABEL: test_tzcnt_u16:
+; X64:       # BB#0:
+; X64-NEXT:    movw $16, %cx
+; X64-NEXT:    movzwl %di, %edx
+; X64-NEXT:    tzcntw %dx, %ax
+; X64-NEXT:    cmpl $0, %edx
+; X64-NEXT:    cmovew %cx, %ax
+; X64-NEXT:    retq
+  %zext = zext i16 %a0 to i32
+  %cmp = icmp ne i32 %zext, 0
+  %cttz = call i16 @llvm.cttz.i16(i16 %a0, i1 true)
+  %res = select i1 %cmp, i16 %cttz, i16 16
+  ret i16 %res
+}
+
+define i32 @test_andn_u32(i32 %a0, i32 %a1) {
+; X32-LABEL: test_andn_u32:
+; X32:       # BB#0:
+; X32-NEXT:    movl {{[0-9]+}}(%esp), %eax
+; X32-NEXT:    xorl $-1, %eax
+; X32-NEXT:    andl {{[0-9]+}}(%esp), %eax
+; X32-NEXT:    retl
+;
+; X64-LABEL: test_andn_u32:
+; X64:       # BB#0:
+; X64-NEXT:    xorl $-1, %edi
+; X64-NEXT:    andl %esi, %edi
+; X64-NEXT:    movl %edi, %eax
+; X64-NEXT:    retq
+  %xor = xor i32 %a0, -1
+  %res = and i32 %xor, %a1
+  ret i32 %res
+}
+
+define i32 @test_bextr_u32(i32 %a0, i32 %a1, i32 %a2) {
+; X32-LABEL: test_bextr_u32:
+; X32:       # BB#0:
+; X32-NEXT:    movl {{[0-9]+}}(%esp), %eax
+; X32-NEXT:    movl {{[0-9]+}}(%esp), %ecx
+; X32-NEXT:    andl $255, %ecx
+; X32-NEXT:    andl $255, %eax
+; X32-NEXT:    shll $8, %eax
+; X32-NEXT:    orl %ecx, %eax
+; X32-NEXT:    bextrl %eax, {{[0-9]+}}(%esp), %eax
+; X32-NEXT:    retl
+;
+; X64-LABEL: test_bextr_u32:
+; X64:       # BB#0:
+; X64-NEXT:    andl $255, %esi
+; X64-NEXT:    andl $255, %edx
+; X64-NEXT:    shll $8, %edx
+; X64-NEXT:    orl %esi, %edx
+; X64-NEXT:    bextrl %edx, %edi, %eax
+; X64-NEXT:    retq
+  %and1 = and i32 %a1, 255
+  %and2 = and i32 %a2, 255
+  %shl = shl i32 %and2, 8
+  %or = or i32 %and1, %shl
+  %res = call i32 @llvm.x86.bmi.bextr.32(i32 %a0, i32 %or)
+  ret i32 %res
+}
+
+define i32 @test_blsi_u32(i32 %a0) {
+; X32-LABEL: test_blsi_u32:
+; X32:       # BB#0:
+; X32-NEXT:    movl {{[0-9]+}}(%esp), %ecx
+; X32-NEXT:    xorl %eax, %eax
+; X32-NEXT:    subl %ecx, %eax
+; X32-NEXT:    andl %ecx, %eax
+; X32-NEXT:    retl
+;
+; X64-LABEL: test_blsi_u32:
+; X64:       # BB#0:
+; X64-NEXT:    xorl %eax, %eax
+; X64-NEXT:    subl %edi, %eax
+; X64-NEXT:    andl %edi, %eax
+; X64-NEXT:    retq
+  %neg = sub i32 0, %a0
+  %res = and i32 %a0, %neg
+  ret i32 %res
+}
+
+define i32 @test_blsmsk_u32(i32 %a0) {
+; X32-LABEL: test_blsmsk_u32:
+; X32:       # BB#0:
+; X32-NEXT:    movl {{[0-9]+}}(%esp), %ecx
+; X32-NEXT:    movl %ecx, %eax
+; X32-NEXT:    subl $1, %eax
+; X32-NEXT:    xorl %ecx, %eax
+; X32-NEXT:    retl
+;
+; X64-LABEL: test_blsmsk_u32:
+; X64:       # BB#0:
+; X64-NEXT:    movl %edi, %eax
+; X64-NEXT:    subl $1, %eax
+; X64-NEXT:    xorl %edi, %eax
+; X64-NEXT:    retq
+  %dec = sub i32 %a0, 1
+  %res = xor i32 %a0, %dec
+  ret i32 %res
+}
+
+define i32 @test_blsr_u32(i32 %a0) {
+; X32-LABEL: test_blsr_u32:
+; X32:       # BB#0:
+; X32-NEXT:    movl {{[0-9]+}}(%esp), %ecx
+; X32-NEXT:    movl %ecx, %eax
+; X32-NEXT:    subl $1, %eax
+; X32-NEXT:    andl %ecx, %eax
+; X32-NEXT:    retl
+;
+; X64-LABEL: test_blsr_u32:
+; X64:       # BB#0:
+; X64-NEXT:    movl %edi, %eax
+; X64-NEXT:    subl $1, %eax
+; X64-NEXT:    andl %edi, %eax
+; X64-NEXT:    retq
+  %dec = sub i32 %a0, 1
+  %res = and i32 %a0, %dec
+  ret i32 %res
+}
+
+define i32 @test_tzcnt_u32(i32 %a0) {
+; X32-LABEL: test_tzcnt_u32:
+; X32:       # BB#0:
+; X32-NEXT:    movl {{[0-9]+}}(%esp), %eax
+; X32-NEXT:    cmpl $0, %eax
+; X32-NEXT:    jne .LBB13_1
+; X32-NEXT:  # BB#2:
+; X32-NEXT:    movl $32, %eax
+; X32-NEXT:    retl
+; X32-NEXT:  .LBB13_1:
+; X32-NEXT:    tzcntl %eax, %eax
+; X32-NEXT:    retl
+;
+; X64-LABEL: test_tzcnt_u32:
+; X64:       # BB#0:
+; X64-NEXT:    movl $32, %ecx
+; X64-NEXT:    tzcntl %edi, %eax
+; X64-NEXT:    cmovbl %ecx, %eax
+; X64-NEXT:    retq
+  %cmp = icmp ne i32 %a0, 0
+  %cttz = call i32 @llvm.cttz.i32(i32 %a0, i1 true)
+  %res = select i1 %cmp, i32 %cttz, i32 32
+  ret i32 %res
+}
+
+declare i16 @llvm.cttz.i16(i16, i1)
+declare i32 @llvm.cttz.i32(i32, i1)
+declare i32 @llvm.x86.bmi.bextr.32(i32, i32)




More information about the llvm-commits mailing list