[llvm] 9d23224 - [X86] Add test cases for PR47874. NFC

Craig Topper via llvm-commits llvm-commits at lists.llvm.org
Sun Oct 18 12:52:12 PDT 2020


Author: Craig Topper
Date: 2020-10-18T12:45:01-07:00
New Revision: 9d23224bf656358e409b620803f6b49341258dea

URL: https://github.com/llvm/llvm-project/commit/9d23224bf656358e409b620803f6b49341258dea
DIFF: https://github.com/llvm/llvm-project/commit/9d23224bf656358e409b620803f6b49341258dea.diff

LOG: [X86] Add test cases for PR47874. NFC

Added: 
    llvm/test/CodeGen/X86/pr47874.ll

Modified: 
    

Removed: 
    


################################################################################
diff  --git a/llvm/test/CodeGen/X86/pr47874.ll b/llvm/test/CodeGen/X86/pr47874.ll
new file mode 100644
index 000000000000..76bfd4613e1c
--- /dev/null
+++ b/llvm/test/CodeGen/X86/pr47874.ll
@@ -0,0 +1,276 @@
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
+; RUN: llc < %s -mtriple=x86_64-apple-darwin19.6.0 | FileCheck %s --check-prefixes=CHECK,SSE2
+; RUN: llc < %s -mtriple=x86_64-apple-darwin19.6.0 -mattr=avx | FileCheck %s --check-prefixes=CHECK,AVX
+; RUN: llc < %s -mtriple=x86_64-apple-darwin19.6.0 -mattr=avx512f | FileCheck %s --check-prefixes=CHECK,AVX
+
+define void @a(float* %arg, i32 %arg1) {
+; SSE2-LABEL: a:
+; SSE2:       ## %bb.0: ## %bb
+; SSE2-NEXT:    testl %esi, %esi
+; SSE2-NEXT:    jle LBB0_3
+; SSE2-NEXT:  ## %bb.1: ## %bb2
+; SSE2-NEXT:    movd %esi, %xmm0
+; SSE2-NEXT:    movd %xmm0, {{[-0-9]+}}(%r{{[sb]}}p) ## 4-byte Folded Spill
+; SSE2-NEXT:    movl %esi, %eax
+; SSE2-NEXT:    .p2align 4, 0x90
+; SSE2-NEXT:  LBB0_2: ## %bb6
+; SSE2-NEXT:    ## =>This Inner Loop Header: Depth=1
+; SSE2-NEXT:    ## InlineAsm Start
+; SSE2-NEXT:    ## InlineAsm End
+; SSE2-NEXT:    movss {{.*#+}} xmm0 = mem[0],zero,zero,zero
+; SSE2-NEXT:    addss {{[-0-9]+}}(%r{{[sb]}}p), %xmm0 ## 4-byte Folded Reload
+; SSE2-NEXT:    movss %xmm0, (%rdi)
+; SSE2-NEXT:    addq $4, %rdi
+; SSE2-NEXT:    decq %rax
+; SSE2-NEXT:    jne LBB0_2
+; SSE2-NEXT:  LBB0_3: ## %bb5
+; SSE2-NEXT:    retq
+;
+; AVX-LABEL: a:
+; AVX:       ## %bb.0: ## %bb
+; AVX-NEXT:    testl %esi, %esi
+; AVX-NEXT:    jle LBB0_3
+; AVX-NEXT:  ## %bb.1: ## %bb2
+; AVX-NEXT:    vmovd %esi, %xmm0
+; AVX-NEXT:    vmovd %xmm0, {{[-0-9]+}}(%r{{[sb]}}p) ## 4-byte Folded Spill
+; AVX-NEXT:    movl %esi, %eax
+; AVX-NEXT:    .p2align 4, 0x90
+; AVX-NEXT:  LBB0_2: ## %bb6
+; AVX-NEXT:    ## =>This Inner Loop Header: Depth=1
+; AVX-NEXT:    ## InlineAsm Start
+; AVX-NEXT:    ## InlineAsm End
+; AVX-NEXT:    vmovss {{[-0-9]+}}(%r{{[sb]}}p), %xmm0 ## 4-byte Reload
+; AVX-NEXT:    ## xmm0 = mem[0],zero,zero,zero
+; AVX-NEXT:    vaddss (%rdi), %xmm0, %xmm0
+; AVX-NEXT:    vmovss %xmm0, (%rdi)
+; AVX-NEXT:    addq $4, %rdi
+; AVX-NEXT:    decq %rax
+; AVX-NEXT:    jne LBB0_2
+; AVX-NEXT:  LBB0_3: ## %bb5
+; AVX-NEXT:    retq
+bb:
+  %i = icmp sgt i32 %arg1, 0
+  br i1 %i, label %bb2, label %bb5
+
+bb2:                                              ; preds = %bb
+  %i3 = bitcast i32 %arg1 to float
+  %i4 = zext i32 %arg1 to i64
+  br label %bb6
+
+bb5:                                              ; preds = %bb6, %bb
+  ret void
+
+bb6:                                              ; preds = %bb6, %bb2
+  %i7 = phi i64 [ 0, %bb2 ], [ %i11, %bb6 ]
+  tail call void asm sideeffect "", "~{xmm0},~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{xmm16},~{xmm17},~{xmm18},~{xmm19},~{xmm20},~{xmm21},~{xmm22},~{xmm23},~{xmm24},~{xmm25},~{xmm26},~{xmm27},~{xmm28},~{xmm29},~{xmm30},~{xmm31},~{dirflag},~{fpsr},~{flags}"()
+  %i8 = getelementptr inbounds float, float* %arg, i64 %i7
+  %i9 = load float, float* %i8, align 4
+  %i10 = fadd float %i9, %i3
+  store float %i10, float* %i8, align 4
+  %i11 = add nuw nsw i64 %i7, 1
+  %i12 = icmp eq i64 %i11, %i4
+  br i1 %i12, label %bb5, label %bb6
+}
+
+define void @b(double* %arg, i64 %arg1) {
+; SSE2-LABEL: b:
+; SSE2:       ## %bb.0: ## %bb
+; SSE2-NEXT:    testq %rsi, %rsi
+; SSE2-NEXT:    jle LBB1_3
+; SSE2-NEXT:  ## %bb.1: ## %bb2
+; SSE2-NEXT:    movq %rsi, %xmm0
+; SSE2-NEXT:    movq %xmm0, {{[-0-9]+}}(%r{{[sb]}}p) ## 8-byte Folded Spill
+; SSE2-NEXT:    .p2align 4, 0x90
+; SSE2-NEXT:  LBB1_2: ## %bb6
+; SSE2-NEXT:    ## =>This Inner Loop Header: Depth=1
+; SSE2-NEXT:    ## InlineAsm Start
+; SSE2-NEXT:    ## InlineAsm End
+; SSE2-NEXT:    movsd {{.*#+}} xmm0 = mem[0],zero
+; SSE2-NEXT:    addsd {{[-0-9]+}}(%r{{[sb]}}p), %xmm0 ## 8-byte Folded Reload
+; SSE2-NEXT:    movsd %xmm0, (%rdi)
+; SSE2-NEXT:    addq $8, %rdi
+; SSE2-NEXT:    decq %rsi
+; SSE2-NEXT:    jne LBB1_2
+; SSE2-NEXT:  LBB1_3: ## %bb5
+; SSE2-NEXT:    retq
+;
+; AVX-LABEL: b:
+; AVX:       ## %bb.0: ## %bb
+; AVX-NEXT:    testq %rsi, %rsi
+; AVX-NEXT:    jle LBB1_3
+; AVX-NEXT:  ## %bb.1: ## %bb2
+; AVX-NEXT:    vmovq %rsi, %xmm0
+; AVX-NEXT:    vmovq %xmm0, {{[-0-9]+}}(%r{{[sb]}}p) ## 8-byte Folded Spill
+; AVX-NEXT:    .p2align 4, 0x90
+; AVX-NEXT:  LBB1_2: ## %bb6
+; AVX-NEXT:    ## =>This Inner Loop Header: Depth=1
+; AVX-NEXT:    ## InlineAsm Start
+; AVX-NEXT:    ## InlineAsm End
+; AVX-NEXT:    vmovsd {{[-0-9]+}}(%r{{[sb]}}p), %xmm0 ## 8-byte Reload
+; AVX-NEXT:    ## xmm0 = mem[0],zero
+; AVX-NEXT:    vaddsd (%rdi), %xmm0, %xmm0
+; AVX-NEXT:    vmovsd %xmm0, (%rdi)
+; AVX-NEXT:    addq $8, %rdi
+; AVX-NEXT:    decq %rsi
+; AVX-NEXT:    jne LBB1_2
+; AVX-NEXT:  LBB1_3: ## %bb5
+; AVX-NEXT:    retq
+bb:
+  %i = icmp sgt i64 %arg1, 0
+  br i1 %i, label %bb2, label %bb5
+
+bb2:                                              ; preds = %bb
+  %i3 = bitcast i64 %arg1 to double
+  br label %bb6
+
+bb5:                                              ; preds = %bb6, %bb
+  ret void
+
+bb6:                                              ; preds = %bb6, %bb2
+  %i7 = phi i64 [ 0, %bb2 ], [ %i11, %bb6 ]
+  tail call void asm sideeffect "", "~{xmm0},~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{xmm16},~{xmm17},~{xmm18},~{xmm19},~{xmm20},~{xmm21},~{xmm22},~{xmm23},~{xmm24},~{xmm25},~{xmm26},~{xmm27},~{xmm28},~{xmm29},~{xmm30},~{xmm31},~{dirflag},~{fpsr},~{flags}"()
+  %i8 = getelementptr inbounds double, double* %arg, i64 %i7
+  %i9 = load double, double* %i8, align 4
+  %i10 = fadd double %i9, %i3
+  store double %i10, double* %i8, align 4
+  %i11 = add nuw nsw i64 %i7, 1
+  %i12 = icmp eq i64 %i11, %arg1
+  br i1 %i12, label %bb5, label %bb6
+}
+
+define void @c(<4 x float>* %arg, <4 x float>* %arg1, i32 %arg2) {
+; SSE2-LABEL: c:
+; SSE2:       ## %bb.0: ## %bb
+; SSE2-NEXT:    testl %edx, %edx
+; SSE2-NEXT:    jle LBB2_3
+; SSE2-NEXT:  ## %bb.1: ## %bb4
+; SSE2-NEXT:    movd %edx, %xmm0
+; SSE2-NEXT:    movdqa %xmm0, {{[-0-9]+}}(%r{{[sb]}}p) ## 16-byte Spill
+; SSE2-NEXT:    movl %edx, %eax
+; SSE2-NEXT:    .p2align 4, 0x90
+; SSE2-NEXT:  LBB2_2: ## %bb8
+; SSE2-NEXT:    ## =>This Inner Loop Header: Depth=1
+; SSE2-NEXT:    ## InlineAsm Start
+; SSE2-NEXT:    ## InlineAsm End
+; SSE2-NEXT:    movaps (%rdi), %xmm0
+; SSE2-NEXT:    addss {{[-0-9]+}}(%r{{[sb]}}p), %xmm0 ## 16-byte Folded Reload
+; SSE2-NEXT:    movaps %xmm0, (%rdi)
+; SSE2-NEXT:    addq $16, %rdi
+; SSE2-NEXT:    decq %rax
+; SSE2-NEXT:    jne LBB2_2
+; SSE2-NEXT:  LBB2_3: ## %bb7
+; SSE2-NEXT:    retq
+;
+; AVX-LABEL: c:
+; AVX:       ## %bb.0: ## %bb
+; AVX-NEXT:    testl %edx, %edx
+; AVX-NEXT:    jle LBB2_3
+; AVX-NEXT:  ## %bb.1: ## %bb4
+; AVX-NEXT:    vmovd %edx, %xmm0
+; AVX-NEXT:    vmovdqa %xmm0, {{[-0-9]+}}(%r{{[sb]}}p) ## 16-byte Spill
+; AVX-NEXT:    movl %edx, %eax
+; AVX-NEXT:    .p2align 4, 0x90
+; AVX-NEXT:  LBB2_2: ## %bb8
+; AVX-NEXT:    ## =>This Inner Loop Header: Depth=1
+; AVX-NEXT:    ## InlineAsm Start
+; AVX-NEXT:    ## InlineAsm End
+; AVX-NEXT:    vmovaps (%rdi), %xmm0
+; AVX-NEXT:    vaddss {{[-0-9]+}}(%r{{[sb]}}p), %xmm0, %xmm0 ## 16-byte Folded Reload
+; AVX-NEXT:    vmovaps %xmm0, (%rdi)
+; AVX-NEXT:    addq $16, %rdi
+; AVX-NEXT:    decq %rax
+; AVX-NEXT:    jne LBB2_2
+; AVX-NEXT:  LBB2_3: ## %bb7
+; AVX-NEXT:    retq
+bb:
+  %i = icmp sgt i32 %arg2, 0
+  br i1 %i, label %bb4, label %bb7
+
+bb4:                                              ; preds = %bb
+  %i5 = bitcast i32 %arg2 to float
+  %i6 = zext i32 %arg2 to i64
+  br label %bb8
+
+bb7:                                              ; preds = %bb8, %bb
+  ret void
+
+bb8:                                              ; preds = %bb8, %bb4
+  %i9 = phi i64 [ 0, %bb4 ], [ %i15, %bb8 ]
+  tail call void asm sideeffect "", "~{xmm0},~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{xmm16},~{xmm17},~{xmm18},~{xmm19},~{xmm20},~{xmm21},~{xmm22},~{xmm23},~{xmm24},~{xmm25},~{xmm26},~{xmm27},~{xmm28},~{xmm29},~{xmm30},~{xmm31},~{dirflag},~{fpsr},~{flags}"()
+  %i10 = getelementptr inbounds <4 x float>, <4 x float>* %arg, i64 %i9
+  %i11 = load <4 x float>, <4 x float>* %i10, align 16
+  %i12 = extractelement <4 x float> %i11, i32 0
+  %i13 = fadd float %i12, %i5
+  %i14 = insertelement <4 x float> %i11, float %i13, i32 0
+  store <4 x float> %i14, <4 x float>* %i10, align 16
+  %i15 = add nuw nsw i64 %i9, 1
+  %i16 = icmp eq i64 %i15, %i6
+  br i1 %i16, label %bb7, label %bb8
+}
+
+define void @d(<2 x double>* %arg, <2 x double>* %arg1, i64 %arg2) {
+; SSE2-LABEL: d:
+; SSE2:       ## %bb.0: ## %bb
+; SSE2-NEXT:    testq %rdx, %rdx
+; SSE2-NEXT:    jle LBB3_3
+; SSE2-NEXT:  ## %bb.1: ## %bb3
+; SSE2-NEXT:    movq %rdx, %xmm0
+; SSE2-NEXT:    movdqa %xmm0, {{[-0-9]+}}(%r{{[sb]}}p) ## 16-byte Spill
+; SSE2-NEXT:    .p2align 4, 0x90
+; SSE2-NEXT:  LBB3_2: ## %bb6
+; SSE2-NEXT:    ## =>This Inner Loop Header: Depth=1
+; SSE2-NEXT:    ## InlineAsm Start
+; SSE2-NEXT:    ## InlineAsm End
+; SSE2-NEXT:    movapd (%rdi), %xmm0
+; SSE2-NEXT:    addsd {{[-0-9]+}}(%r{{[sb]}}p), %xmm0 ## 16-byte Folded Reload
+; SSE2-NEXT:    movapd %xmm0, (%rdi)
+; SSE2-NEXT:    addq $16, %rdi
+; SSE2-NEXT:    decq %rdx
+; SSE2-NEXT:    jne LBB3_2
+; SSE2-NEXT:  LBB3_3: ## %bb5
+; SSE2-NEXT:    retq
+;
+; AVX-LABEL: d:
+; AVX:       ## %bb.0: ## %bb
+; AVX-NEXT:    testq %rdx, %rdx
+; AVX-NEXT:    jle LBB3_3
+; AVX-NEXT:  ## %bb.1: ## %bb3
+; AVX-NEXT:    vmovq %rdx, %xmm0
+; AVX-NEXT:    vmovdqa %xmm0, {{[-0-9]+}}(%r{{[sb]}}p) ## 16-byte Spill
+; AVX-NEXT:    .p2align 4, 0x90
+; AVX-NEXT:  LBB3_2: ## %bb6
+; AVX-NEXT:    ## =>This Inner Loop Header: Depth=1
+; AVX-NEXT:    ## InlineAsm Start
+; AVX-NEXT:    ## InlineAsm End
+; AVX-NEXT:    vmovapd (%rdi), %xmm0
+; AVX-NEXT:    vaddsd {{[-0-9]+}}(%r{{[sb]}}p), %xmm0, %xmm0 ## 16-byte Folded Reload
+; AVX-NEXT:    vmovapd %xmm0, (%rdi)
+; AVX-NEXT:    addq $16, %rdi
+; AVX-NEXT:    decq %rdx
+; AVX-NEXT:    jne LBB3_2
+; AVX-NEXT:  LBB3_3: ## %bb5
+; AVX-NEXT:    retq
+bb:
+  %i = icmp sgt i64 %arg2, 0
+  br i1 %i, label %bb3, label %bb5
+
+bb3:                                              ; preds = %bb
+  %i4 = bitcast i64 %arg2 to double
+  br label %bb6
+
+bb5:                                              ; preds = %bb6, %bb
+  ret void
+
+bb6:                                              ; preds = %bb6, %bb3
+  %i7 = phi i64 [ 0, %bb3 ], [ %i13, %bb6 ]
+  tail call void asm sideeffect "", "~{xmm0},~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{xmm16},~{xmm17},~{xmm18},~{xmm19},~{xmm20},~{xmm21},~{xmm22},~{xmm23},~{xmm24},~{xmm25},~{xmm26},~{xmm27},~{xmm28},~{xmm29},~{xmm30},~{xmm31},~{dirflag},~{fpsr},~{flags}"()
+  %i8 = getelementptr inbounds <2 x double>, <2 x double>* %arg, i64 %i7
+  %i9 = load <2 x double>, <2 x double>* %i8, align 16
+  %i10 = extractelement <2 x double> %i9, i32 0
+  %i11 = fadd double %i10, %i4
+  %i12 = insertelement <2 x double> %i9, double %i11, i32 0
+  store <2 x double> %i12, <2 x double>* %i8, align 16
+  %i13 = add nuw nsw i64 %i7, 1
+  %i14 = icmp eq i64 %i13, %arg2
+  br i1 %i14, label %bb5, label %bb6
+}


        


More information about the llvm-commits mailing list