[llvm] r278165 - add test cases for missed vselect optimizations (PR28895)

Sanjay Patel via llvm-commits llvm-commits at lists.llvm.org
Tue Aug 9 14:07:17 PDT 2016


Author: spatel
Date: Tue Aug  9 16:07:17 2016
New Revision: 278165

URL: http://llvm.org/viewvc/llvm-project?rev=278165&view=rev
Log:
add test cases for missed vselect optimizations (PR28895)

Modified:
    llvm/trunk/test/CodeGen/X86/select-with-and-or.ll

Modified: llvm/trunk/test/CodeGen/X86/select-with-and-or.ll
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/X86/select-with-and-or.ll?rev=278165&r1=278164&r2=278165&view=diff
==============================================================================
--- llvm/trunk/test/CodeGen/X86/select-with-and-or.ll (original)
+++ llvm/trunk/test/CodeGen/X86/select-with-and-or.ll Tue Aug  9 16:07:17 2016
@@ -84,3 +84,100 @@ define <4 x i32> @test7(<4 x float> %a,
   ret <4 x i32> %r
 }
 
+; FIXME: None of these should use vblendvpd.
+; Repeat all with FP types.
+
+define <2 x double> @test1f(<2 x double> %a, <2 x double> %b, <2 x double> %c) {
+; CHECK-LABEL: test1f:
+; CHECK:       # BB#0:
+; CHECK-NEXT:    vcmpnlepd %xmm0, %xmm1, %xmm0
+; CHECK-NEXT:    vxorpd %xmm1, %xmm1, %xmm1
+; CHECK-NEXT:    vblendvpd %xmm0, %xmm2, %xmm1, %xmm0
+; CHECK-NEXT:    retq
+;
+  %f = fcmp ult <2 x double> %a, %b
+  %r = select <2 x i1> %f, <2 x double> %c, <2 x double> zeroinitializer
+  ret <2 x double> %r
+}
+
+define <2 x double> @test2f(<2 x double> %a, <2 x double> %b, <2 x double> %c) {
+; CHECK-LABEL: test2f:
+; CHECK:       # BB#0:
+; CHECK-NEXT:    vcmpnlepd %xmm0, %xmm1, %xmm0
+; CHECK-NEXT:    vpcmpeqd %xmm1, %xmm1, %xmm1
+; CHECK-NEXT:    vblendvpd %xmm0, %xmm1, %xmm2, %xmm0
+; CHECK-NEXT:    retq
+;
+  %f = fcmp ult <2 x double> %a, %b
+  %r = select <2 x i1> %f, <2 x double> <double 0xffffffffffffffff, double 0xffffffffffffffff>, <2 x double> %c
+  ret <2 x double> %r
+}
+
+define <2 x double> @test3f(<2 x double> %a, <2 x double> %b, <2 x double> %c) {
+; CHECK-LABEL: test3f:
+; CHECK:       # BB#0:
+; CHECK-NEXT:    vcmpnlepd %xmm0, %xmm1, %xmm0
+; CHECK-NEXT:    vxorpd %xmm1, %xmm1, %xmm1
+; CHECK-NEXT:    vblendvpd %xmm0, %xmm1, %xmm2, %xmm0
+; CHECK-NEXT:    retq
+;
+  %f = fcmp ult <2 x double> %a, %b
+  %r = select <2 x i1> %f, <2 x double> zeroinitializer, <2 x double> %c
+  ret <2 x double> %r
+}
+
+define <2 x double> @test4f(<2 x double> %a, <2 x double> %b, <2 x double> %c) {
+; CHECK-LABEL: test4f:
+; CHECK:       # BB#0:
+; CHECK-NEXT:    vcmpnlepd %xmm0, %xmm1, %xmm0
+; CHECK-NEXT:    vpcmpeqd %xmm1, %xmm1, %xmm1
+; CHECK-NEXT:    vblendvpd %xmm0, %xmm2, %xmm1, %xmm0
+; CHECK-NEXT:    retq
+;
+  %f = fcmp ult <2 x double> %a, %b
+  %r = select <2 x i1> %f, <2 x double> %c, <2 x double> <double 0xffffffffffffffff, double 0xffffffffffffffff>
+  ret <2 x double> %r
+}
+
+define <2 x double> @test5f(<2 x double> %a, <2 x double> %b, <2 x double> %c) {
+; CHECK-LABEL: test5f:
+; CHECK:       # BB#0:
+; CHECK-NEXT:    vcmpnlepd %xmm0, %xmm1, %xmm0
+; CHECK-NEXT:    vpcmpeqd %xmm1, %xmm1, %xmm1
+; CHECK-NEXT:    vxorpd %xmm2, %xmm2, %xmm2
+; CHECK-NEXT:    vblendvpd %xmm0, %xmm1, %xmm2, %xmm0
+; CHECK-NEXT:    retq
+;
+  %f = fcmp ult <2 x double> %a, %b
+  %r = select <2 x i1> %f, <2 x double> <double 0xffffffffffffffff, double 0xffffffffffffffff>, <2 x double> zeroinitializer
+  ret <2 x double> %r
+}
+
+define <2 x double> @test6f(<2 x double> %a, <2 x double> %b, <2 x double> %c) {
+; CHECK-LABEL: test6f:
+; CHECK:       # BB#0:
+; CHECK-NEXT:    vcmpnlepd %xmm0, %xmm1, %xmm0
+; CHECK-NEXT:    vxorpd %xmm1, %xmm1, %xmm1
+; CHECK-NEXT:    vpcmpeqd %xmm2, %xmm2, %xmm2
+; CHECK-NEXT:    vblendvpd %xmm0, %xmm1, %xmm2, %xmm0
+; CHECK-NEXT:    retq
+;
+  %f = fcmp ult <2 x double> %a, %b
+  %r = select <2 x i1> %f, <2 x double> zeroinitializer, <2 x double> <double 0xffffffffffffffff, double 0xffffffffffffffff>
+  ret <2 x double> %r
+}
+
+define <2 x double> @test7f(<2 x double> %a, <2 x double> %b, <2 x double>* %p) {
+; CHECK-LABEL: test7f:
+; CHECK:       # BB#0:
+; CHECK-NEXT:    vcmpnlepd %xmm0, %xmm1, %xmm0
+; CHECK-NEXT:    vxorpd %xmm1, %xmm1, %xmm1
+; CHECK-NEXT:    vblendvpd %xmm0, (%rdi), %xmm1, %xmm0
+; CHECK-NEXT:    retq
+;
+  %f = fcmp ult <2 x double> %a, %b
+  %l = load <2 x double>, <2 x double>* %p, align 16
+  %r = select <2 x i1> %f, <2 x double> %l, <2 x double> zeroinitializer
+  ret <2 x double> %r
+}
+




More information about the llvm-commits mailing list