[llvm] fe494fa - [X86][SSE] Add codegen tests dot2/3 dot product of 128-bit dereferenceable float data

Simon Pilgrim via llvm-commits llvm-commits at lists.llvm.org
Mon Jul 19 02:44:41 PDT 2021


Author: Simon Pilgrim
Date: 2021-07-19T10:44:25+01:00
New Revision: fe494fafa9025ac6e9671be300035a27d70ae469

URL: https://github.com/llvm/llvm-project/commit/fe494fafa9025ac6e9671be300035a27d70ae469
DIFF: https://github.com/llvm/llvm-project/commit/fe494fafa9025ac6e9671be300035a27d70ae469.diff

LOG: [X86][SSE] Add codegen tests dot2/3 dot product of 128-bit dereferenceable float data

Based off the codegen reports on PR51075 - hopefully we can handle some of this in SLP or VectorCombine, but we usually have to leave load combining until the backend so at least some of these patterns will still appear even then.

Added: 
    llvm/test/CodeGen/X86/load-partial-dot-product.ll

Modified: 
    

Removed: 
    


################################################################################
diff  --git a/llvm/test/CodeGen/X86/load-partial-dot-product.ll b/llvm/test/CodeGen/X86/load-partial-dot-product.ll
new file mode 100644
index 000000000000..1b296ea4fe95
--- /dev/null
+++ b/llvm/test/CodeGen/X86/load-partial-dot-product.ll
@@ -0,0 +1,429 @@
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
+; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+sse2 | FileCheck %s --check-prefixes=SSE2
+; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+ssse3 | FileCheck %s --check-prefixes=SSSE3
+; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+sse4.1 | FileCheck %s --check-prefixes=SSE41
+; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+avx | FileCheck %s --check-prefixes=AVX
+; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+avx2 | FileCheck %s --check-prefixes=AVX
+
+; Partial laod dot product patterns based off PR51075
+
+;
+; dot3(float *x, float *y) - ((x[0]*y[0])+(x[1]*y[1])+(x[2]*y[2]))
+;
+
+define float @dot3_float4(float* dereferenceable(16) %a0, float* dereferenceable(16) %a1) {
+; SSE2-LABEL: dot3_float4:
+; SSE2:       # %bb.0:
+; SSE2-NEXT:    movups (%rdi), %xmm0
+; SSE2-NEXT:    movups (%rsi), %xmm1
+; SSE2-NEXT:    mulps %xmm0, %xmm1
+; SSE2-NEXT:    movaps %xmm1, %xmm0
+; SSE2-NEXT:    shufps {{.*#+}} xmm0 = xmm0[1,1],xmm1[1,1]
+; SSE2-NEXT:    addss %xmm1, %xmm0
+; SSE2-NEXT:    movhlps {{.*#+}} xmm1 = xmm1[1,1]
+; SSE2-NEXT:    addss %xmm1, %xmm0
+; SSE2-NEXT:    retq
+;
+; SSSE3-LABEL: dot3_float4:
+; SSSE3:       # %bb.0:
+; SSSE3-NEXT:    movups (%rdi), %xmm0
+; SSSE3-NEXT:    movups (%rsi), %xmm1
+; SSSE3-NEXT:    mulps %xmm0, %xmm1
+; SSSE3-NEXT:    movshdup {{.*#+}} xmm0 = xmm1[1,1,3,3]
+; SSSE3-NEXT:    addss %xmm1, %xmm0
+; SSSE3-NEXT:    movhlps {{.*#+}} xmm1 = xmm1[1,1]
+; SSSE3-NEXT:    addss %xmm1, %xmm0
+; SSSE3-NEXT:    retq
+;
+; SSE41-LABEL: dot3_float4:
+; SSE41:       # %bb.0:
+; SSE41-NEXT:    movups (%rdi), %xmm0
+; SSE41-NEXT:    movups (%rsi), %xmm1
+; SSE41-NEXT:    mulps %xmm0, %xmm1
+; SSE41-NEXT:    movshdup {{.*#+}} xmm0 = xmm1[1,1,3,3]
+; SSE41-NEXT:    addss %xmm1, %xmm0
+; SSE41-NEXT:    movhlps {{.*#+}} xmm1 = xmm1[1,1]
+; SSE41-NEXT:    addss %xmm1, %xmm0
+; SSE41-NEXT:    retq
+;
+; AVX-LABEL: dot3_float4:
+; AVX:       # %bb.0:
+; AVX-NEXT:    vmovups (%rdi), %xmm0
+; AVX-NEXT:    vmulps (%rsi), %xmm0, %xmm0
+; AVX-NEXT:    vmovshdup {{.*#+}} xmm1 = xmm0[1,1,3,3]
+; AVX-NEXT:    vpermilpd {{.*#+}} xmm2 = xmm0[1,0]
+; AVX-NEXT:    vaddss %xmm1, %xmm0, %xmm0
+; AVX-NEXT:    vaddss %xmm2, %xmm0, %xmm0
+; AVX-NEXT:    retq
+  %bcx0123 = bitcast float* %a0 to <4 x float>*
+  %bcy0123 = bitcast float* %a1 to <4 x float>*
+  %x0123 = load <4 x float>, <4 x float>* %bcx0123, align 4
+  %y0123 = load <4 x float>, <4 x float>* %bcy0123, align 4
+  %mul0123 = fmul <4 x float> %x0123, %y0123
+  %mul0 = extractelement <4 x float> %mul0123, i32 0
+  %mul1 = extractelement <4 x float> %mul0123, i32 1
+  %mul2 = extractelement <4 x float> %mul0123, i32 2
+  %dot01 = fadd float %mul0, %mul1
+  %dot012 = fadd float %dot01, %mul2
+  ret float %dot012
+}
+
+define float @dot3_float4_as_float3(float* dereferenceable(16) %a0, float* dereferenceable(16) %a1) {
+; SSE2-LABEL: dot3_float4_as_float3:
+; SSE2:       # %bb.0:
+; SSE2-NEXT:    movups (%rdi), %xmm1
+; SSE2-NEXT:    mulps %xmm1, %xmm1
+; SSE2-NEXT:    movaps %xmm1, %xmm0
+; SSE2-NEXT:    shufps {{.*#+}} xmm0 = xmm0[1,1],xmm1[1,1]
+; SSE2-NEXT:    addss %xmm1, %xmm0
+; SSE2-NEXT:    movhlps {{.*#+}} xmm1 = xmm1[1,1]
+; SSE2-NEXT:    addss %xmm1, %xmm0
+; SSE2-NEXT:    retq
+;
+; SSSE3-LABEL: dot3_float4_as_float3:
+; SSSE3:       # %bb.0:
+; SSSE3-NEXT:    movups (%rdi), %xmm1
+; SSSE3-NEXT:    mulps %xmm1, %xmm1
+; SSSE3-NEXT:    movshdup {{.*#+}} xmm0 = xmm1[1,1,3,3]
+; SSSE3-NEXT:    addss %xmm1, %xmm0
+; SSSE3-NEXT:    movhlps {{.*#+}} xmm1 = xmm1[1,1]
+; SSSE3-NEXT:    addss %xmm1, %xmm0
+; SSSE3-NEXT:    retq
+;
+; SSE41-LABEL: dot3_float4_as_float3:
+; SSE41:       # %bb.0:
+; SSE41-NEXT:    movups (%rdi), %xmm1
+; SSE41-NEXT:    mulps %xmm1, %xmm1
+; SSE41-NEXT:    movshdup {{.*#+}} xmm0 = xmm1[1,1,3,3]
+; SSE41-NEXT:    addss %xmm1, %xmm0
+; SSE41-NEXT:    movhlps {{.*#+}} xmm1 = xmm1[1,1]
+; SSE41-NEXT:    addss %xmm1, %xmm0
+; SSE41-NEXT:    retq
+;
+; AVX-LABEL: dot3_float4_as_float3:
+; AVX:       # %bb.0:
+; AVX-NEXT:    vmovups (%rdi), %xmm0
+; AVX-NEXT:    vmulps %xmm0, %xmm0, %xmm0
+; AVX-NEXT:    vmovshdup {{.*#+}} xmm1 = xmm0[1,1,3,3]
+; AVX-NEXT:    vpermilpd {{.*#+}} xmm2 = xmm0[1,0]
+; AVX-NEXT:    vaddss %xmm1, %xmm0, %xmm0
+; AVX-NEXT:    vaddss %xmm2, %xmm0, %xmm0
+; AVX-NEXT:    retq
+  %bcx0123 = bitcast float* %a0 to <4 x float>*
+  %bcy0123 = bitcast float* %a1 to <4 x float>*
+  %x0123 = load <4 x float>, <4 x float>* %bcx0123, align 4
+  %y0123 = load <4 x float>, <4 x float>* %bcy0123, align 4
+  %x012 = shufflevector <4 x float> %x0123, <4 x float> undef, <3 x i32> <i32 0, i32 1, i32 2>
+  %y012 = shufflevector <4 x float> %x0123, <4 x float> undef, <3 x i32> <i32 0, i32 1, i32 2>
+  %mul012 = fmul <3 x float> %x012, %y012
+  %mul0 = extractelement <3 x float> %mul012, i32 0
+  %mul1 = extractelement <3 x float> %mul012, i32 1
+  %mul2 = extractelement <3 x float> %mul012, i32 2
+  %dot01 = fadd float %mul0, %mul1
+  %dot012 = fadd float %dot01, %mul2
+  ret float %dot012
+}
+
+define float @dot3_float3(float* dereferenceable(16) %a0, float* dereferenceable(16) %a1) {
+; SSE2-LABEL: dot3_float3:
+; SSE2:       # %bb.0:
+; SSE2-NEXT:    movsd {{.*#+}} xmm0 = mem[0],zero
+; SSE2-NEXT:    movss {{.*#+}} xmm1 = mem[0],zero,zero,zero
+; SSE2-NEXT:    shufps {{.*#+}} xmm1 = xmm1[0,0],xmm0[3,0]
+; SSE2-NEXT:    shufps {{.*#+}} xmm0 = xmm0[0,1],xmm1[0,2]
+; SSE2-NEXT:    movsd {{.*#+}} xmm1 = mem[0],zero
+; SSE2-NEXT:    movss {{.*#+}} xmm2 = mem[0],zero,zero,zero
+; SSE2-NEXT:    shufps {{.*#+}} xmm2 = xmm2[0,0],xmm1[3,0]
+; SSE2-NEXT:    shufps {{.*#+}} xmm1 = xmm1[0,1],xmm2[0,2]
+; SSE2-NEXT:    mulps %xmm0, %xmm1
+; SSE2-NEXT:    movaps %xmm1, %xmm0
+; SSE2-NEXT:    shufps {{.*#+}} xmm0 = xmm0[1,1],xmm1[1,1]
+; SSE2-NEXT:    addss %xmm1, %xmm0
+; SSE2-NEXT:    movhlps {{.*#+}} xmm1 = xmm1[1,1]
+; SSE2-NEXT:    addss %xmm1, %xmm0
+; SSE2-NEXT:    retq
+;
+; SSSE3-LABEL: dot3_float3:
+; SSSE3:       # %bb.0:
+; SSSE3-NEXT:    movsd {{.*#+}} xmm0 = mem[0],zero
+; SSSE3-NEXT:    movss {{.*#+}} xmm1 = mem[0],zero,zero,zero
+; SSSE3-NEXT:    shufps {{.*#+}} xmm1 = xmm1[0,0],xmm0[3,0]
+; SSSE3-NEXT:    shufps {{.*#+}} xmm0 = xmm0[0,1],xmm1[0,2]
+; SSSE3-NEXT:    movsd {{.*#+}} xmm1 = mem[0],zero
+; SSSE3-NEXT:    movss {{.*#+}} xmm2 = mem[0],zero,zero,zero
+; SSSE3-NEXT:    shufps {{.*#+}} xmm2 = xmm2[0,0],xmm1[3,0]
+; SSSE3-NEXT:    shufps {{.*#+}} xmm1 = xmm1[0,1],xmm2[0,2]
+; SSSE3-NEXT:    mulps %xmm0, %xmm1
+; SSSE3-NEXT:    movshdup {{.*#+}} xmm0 = xmm1[1,1,3,3]
+; SSSE3-NEXT:    addss %xmm1, %xmm0
+; SSSE3-NEXT:    movhlps {{.*#+}} xmm1 = xmm1[1,1]
+; SSSE3-NEXT:    addss %xmm1, %xmm0
+; SSSE3-NEXT:    retq
+;
+; SSE41-LABEL: dot3_float3:
+; SSE41:       # %bb.0:
+; SSE41-NEXT:    movsd {{.*#+}} xmm0 = mem[0],zero
+; SSE41-NEXT:    insertps {{.*#+}} xmm0 = xmm0[0,1],mem[0],xmm0[3]
+; SSE41-NEXT:    movsd {{.*#+}} xmm1 = mem[0],zero
+; SSE41-NEXT:    insertps {{.*#+}} xmm1 = xmm1[0,1],mem[0],xmm1[3]
+; SSE41-NEXT:    mulps %xmm0, %xmm1
+; SSE41-NEXT:    movshdup {{.*#+}} xmm0 = xmm1[1,1,3,3]
+; SSE41-NEXT:    addss %xmm1, %xmm0
+; SSE41-NEXT:    movhlps {{.*#+}} xmm1 = xmm1[1,1]
+; SSE41-NEXT:    addss %xmm1, %xmm0
+; SSE41-NEXT:    retq
+;
+; AVX-LABEL: dot3_float3:
+; AVX:       # %bb.0:
+; AVX-NEXT:    vmovsd {{.*#+}} xmm0 = mem[0],zero
+; AVX-NEXT:    vinsertps {{.*#+}} xmm0 = xmm0[0,1],mem[0],xmm0[3]
+; AVX-NEXT:    vmovsd {{.*#+}} xmm1 = mem[0],zero
+; AVX-NEXT:    vinsertps {{.*#+}} xmm1 = xmm1[0,1],mem[0],xmm1[3]
+; AVX-NEXT:    vmulps %xmm1, %xmm0, %xmm0
+; AVX-NEXT:    vmovshdup {{.*#+}} xmm1 = xmm0[1,1,3,3]
+; AVX-NEXT:    vpermilpd {{.*#+}} xmm2 = xmm0[1,0]
+; AVX-NEXT:    vaddss %xmm1, %xmm0, %xmm0
+; AVX-NEXT:    vaddss %xmm2, %xmm0, %xmm0
+; AVX-NEXT:    retq
+  %bcx012 = bitcast float* %a0 to <3 x float>*
+  %bcy012 = bitcast float* %a1 to <3 x float>*
+  %x012 = load <3 x float>, <3 x float>* %bcx012, align 4
+  %y012 = load <3 x float>, <3 x float>* %bcy012, align 4
+  %mul012 = fmul <3 x float> %x012, %y012
+  %mul0 = extractelement <3 x float> %mul012, i32 0
+  %mul1 = extractelement <3 x float> %mul012, i32 1
+  %mul2 = extractelement <3 x float> %mul012, i32 2
+  %dot01 = fadd float %mul0, %mul1
+  %dot012 = fadd float %dot01, %mul2
+  ret float %dot012
+}
+
+define float @dot3_float2_float(float* dereferenceable(16) %a0, float* dereferenceable(16) %a1) {
+; SSE2-LABEL: dot3_float2_float:
+; SSE2:       # %bb.0:
+; SSE2-NEXT:    movsd {{.*#+}} xmm0 = mem[0],zero
+; SSE2-NEXT:    movsd {{.*#+}} xmm1 = mem[0],zero
+; SSE2-NEXT:    mulps %xmm0, %xmm1
+; SSE2-NEXT:    movss {{.*#+}} xmm2 = mem[0],zero,zero,zero
+; SSE2-NEXT:    mulss 8(%rsi), %xmm2
+; SSE2-NEXT:    movaps %xmm1, %xmm0
+; SSE2-NEXT:    shufps {{.*#+}} xmm0 = xmm0[1,1],xmm1[1,1]
+; SSE2-NEXT:    addss %xmm1, %xmm0
+; SSE2-NEXT:    addss %xmm2, %xmm0
+; SSE2-NEXT:    retq
+;
+; SSSE3-LABEL: dot3_float2_float:
+; SSSE3:       # %bb.0:
+; SSSE3-NEXT:    movsd {{.*#+}} xmm0 = mem[0],zero
+; SSSE3-NEXT:    movsd {{.*#+}} xmm1 = mem[0],zero
+; SSSE3-NEXT:    mulps %xmm0, %xmm1
+; SSSE3-NEXT:    movss {{.*#+}} xmm2 = mem[0],zero,zero,zero
+; SSSE3-NEXT:    mulss 8(%rsi), %xmm2
+; SSSE3-NEXT:    movshdup {{.*#+}} xmm0 = xmm1[1,1,3,3]
+; SSSE3-NEXT:    addss %xmm1, %xmm0
+; SSSE3-NEXT:    addss %xmm2, %xmm0
+; SSSE3-NEXT:    retq
+;
+; SSE41-LABEL: dot3_float2_float:
+; SSE41:       # %bb.0:
+; SSE41-NEXT:    movsd {{.*#+}} xmm0 = mem[0],zero
+; SSE41-NEXT:    movsd {{.*#+}} xmm1 = mem[0],zero
+; SSE41-NEXT:    mulps %xmm0, %xmm1
+; SSE41-NEXT:    movss {{.*#+}} xmm2 = mem[0],zero,zero,zero
+; SSE41-NEXT:    mulss 8(%rsi), %xmm2
+; SSE41-NEXT:    movshdup {{.*#+}} xmm0 = xmm1[1,1,3,3]
+; SSE41-NEXT:    addss %xmm1, %xmm0
+; SSE41-NEXT:    addss %xmm2, %xmm0
+; SSE41-NEXT:    retq
+;
+; AVX-LABEL: dot3_float2_float:
+; AVX:       # %bb.0:
+; AVX-NEXT:    vmovsd {{.*#+}} xmm0 = mem[0],zero
+; AVX-NEXT:    vmovsd {{.*#+}} xmm1 = mem[0],zero
+; AVX-NEXT:    vmulps %xmm1, %xmm0, %xmm0
+; AVX-NEXT:    vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero
+; AVX-NEXT:    vmulss 8(%rsi), %xmm1, %xmm1
+; AVX-NEXT:    vmovshdup {{.*#+}} xmm2 = xmm0[1,1,3,3]
+; AVX-NEXT:    vaddss %xmm2, %xmm0, %xmm0
+; AVX-NEXT:    vaddss %xmm1, %xmm0, %xmm0
+; AVX-NEXT:    retq
+  %bcx01 = bitcast float* %a0 to <2 x float>*
+  %bcy01 = bitcast float* %a1 to <2 x float>*
+  %x01 = load <2 x float>, <2 x float>* %bcx01, align 4
+  %y01 = load <2 x float>, <2 x float>* %bcy01, align 4
+  %ptrx2 = getelementptr inbounds float, float* %a0, i64 2
+  %ptry2 = getelementptr inbounds float, float* %a1, i64 2
+  %x2 = load float, float* %ptrx2, align 4
+  %y2 = load float, float* %ptry2, align 4
+  %mul01 = fmul <2 x float> %x01, %y01
+  %mul2 = fmul float %x2, %y2
+  %mul0 = extractelement <2 x float> %mul01, i32 0
+  %mul1 = extractelement <2 x float> %mul01, i32 1
+  %dot01 = fadd float %mul0, %mul1
+  %dot012 = fadd float %dot01, %mul2
+  ret float %dot012
+}
+
+define float @dot3_float_float2(float* dereferenceable(16) %a0, float* dereferenceable(16) %a1) {
+; SSE2-LABEL: dot3_float_float2:
+; SSE2:       # %bb.0:
+; SSE2-NEXT:    movss {{.*#+}} xmm1 = mem[0],zero,zero,zero
+; SSE2-NEXT:    movsd {{.*#+}} xmm2 = mem[0],zero
+; SSE2-NEXT:    movsd {{.*#+}} xmm0 = mem[0],zero
+; SSE2-NEXT:    mulps %xmm2, %xmm0
+; SSE2-NEXT:    mulss (%rsi), %xmm1
+; SSE2-NEXT:    addss %xmm0, %xmm1
+; SSE2-NEXT:    shufps {{.*#+}} xmm0 = xmm0[1,1,1,1]
+; SSE2-NEXT:    addss %xmm1, %xmm0
+; SSE2-NEXT:    retq
+;
+; SSSE3-LABEL: dot3_float_float2:
+; SSSE3:       # %bb.0:
+; SSSE3-NEXT:    movss {{.*#+}} xmm0 = mem[0],zero,zero,zero
+; SSSE3-NEXT:    movsd {{.*#+}} xmm1 = mem[0],zero
+; SSSE3-NEXT:    movsd {{.*#+}} xmm2 = mem[0],zero
+; SSSE3-NEXT:    mulps %xmm1, %xmm2
+; SSSE3-NEXT:    mulss (%rsi), %xmm0
+; SSSE3-NEXT:    movshdup {{.*#+}} xmm1 = xmm2[1,1,3,3]
+; SSSE3-NEXT:    addss %xmm2, %xmm0
+; SSSE3-NEXT:    addss %xmm1, %xmm0
+; SSSE3-NEXT:    retq
+;
+; SSE41-LABEL: dot3_float_float2:
+; SSE41:       # %bb.0:
+; SSE41-NEXT:    movss {{.*#+}} xmm0 = mem[0],zero,zero,zero
+; SSE41-NEXT:    movsd {{.*#+}} xmm1 = mem[0],zero
+; SSE41-NEXT:    movsd {{.*#+}} xmm2 = mem[0],zero
+; SSE41-NEXT:    mulps %xmm1, %xmm2
+; SSE41-NEXT:    mulss (%rsi), %xmm0
+; SSE41-NEXT:    movshdup {{.*#+}} xmm1 = xmm2[1,1,3,3]
+; SSE41-NEXT:    addss %xmm2, %xmm0
+; SSE41-NEXT:    addss %xmm1, %xmm0
+; SSE41-NEXT:    retq
+;
+; AVX-LABEL: dot3_float_float2:
+; AVX:       # %bb.0:
+; AVX-NEXT:    vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero
+; AVX-NEXT:    vmovsd {{.*#+}} xmm1 = mem[0],zero
+; AVX-NEXT:    vmovsd {{.*#+}} xmm2 = mem[0],zero
+; AVX-NEXT:    vmulps %xmm2, %xmm1, %xmm1
+; AVX-NEXT:    vmulss (%rsi), %xmm0, %xmm0
+; AVX-NEXT:    vmovshdup {{.*#+}} xmm2 = xmm1[1,1,3,3]
+; AVX-NEXT:    vaddss %xmm1, %xmm0, %xmm0
+; AVX-NEXT:    vaddss %xmm2, %xmm0, %xmm0
+; AVX-NEXT:    retq
+  %x0 = load float, float* %a0, align 4
+  %y0 = load float, float* %a1, align 4
+  %ptrx12 = getelementptr inbounds float, float* %a0, i64 1
+  %ptry12 = getelementptr inbounds float, float* %a1, i64 1
+  %bcx12 = bitcast float* %ptrx12 to <2 x float>*
+  %bcy12 = bitcast float* %ptry12 to <2 x float>*
+  %x12 = load <2 x float>, <2 x float>* %bcx12, align 4
+  %y12 = load <2 x float>, <2 x float>* %bcy12, align 4
+  %mul0 = fmul float %x0, %y0
+  %mul12 = fmul <2 x float> %x12, %y12
+  %mul1 = extractelement <2 x float> %mul12, i32 0
+  %mul2 = extractelement <2 x float> %mul12, i32 1
+  %dot01 = fadd float %mul0, %mul1
+  %dot012 = fadd float %dot01, %mul2
+  ret float %dot012
+}
+
+;
+; dot2(float *x, float *y) - ((x[0]*y[0])+(x[1]*y[1]))
+;
+
+define float @dot2_float4(float* dereferenceable(16) %a0, float* dereferenceable(16) %a1) {
+; SSE2-LABEL: dot2_float4:
+; SSE2:       # %bb.0:
+; SSE2-NEXT:    movups (%rdi), %xmm0
+; SSE2-NEXT:    movups (%rsi), %xmm1
+; SSE2-NEXT:    mulps %xmm0, %xmm1
+; SSE2-NEXT:    movaps %xmm1, %xmm0
+; SSE2-NEXT:    shufps {{.*#+}} xmm0 = xmm0[1,1],xmm1[1,1]
+; SSE2-NEXT:    addss %xmm1, %xmm0
+; SSE2-NEXT:    retq
+;
+; SSSE3-LABEL: dot2_float4:
+; SSSE3:       # %bb.0:
+; SSSE3-NEXT:    movups (%rdi), %xmm0
+; SSSE3-NEXT:    movups (%rsi), %xmm1
+; SSSE3-NEXT:    mulps %xmm0, %xmm1
+; SSSE3-NEXT:    movshdup {{.*#+}} xmm0 = xmm1[1,1,3,3]
+; SSSE3-NEXT:    addss %xmm1, %xmm0
+; SSSE3-NEXT:    retq
+;
+; SSE41-LABEL: dot2_float4:
+; SSE41:       # %bb.0:
+; SSE41-NEXT:    movups (%rdi), %xmm0
+; SSE41-NEXT:    movups (%rsi), %xmm1
+; SSE41-NEXT:    mulps %xmm0, %xmm1
+; SSE41-NEXT:    movshdup {{.*#+}} xmm0 = xmm1[1,1,3,3]
+; SSE41-NEXT:    addss %xmm1, %xmm0
+; SSE41-NEXT:    retq
+;
+; AVX-LABEL: dot2_float4:
+; AVX:       # %bb.0:
+; AVX-NEXT:    vmovups (%rdi), %xmm0
+; AVX-NEXT:    vmulps (%rsi), %xmm0, %xmm0
+; AVX-NEXT:    vmovshdup {{.*#+}} xmm1 = xmm0[1,1,3,3]
+; AVX-NEXT:    vaddss %xmm1, %xmm0, %xmm0
+; AVX-NEXT:    retq
+  %bcx0123 = bitcast float* %a0 to <4 x float>*
+  %bcy0123 = bitcast float* %a1 to <4 x float>*
+  %x0123 = load <4 x float>, <4 x float>* %bcx0123, align 4
+  %y0123 = load <4 x float>, <4 x float>* %bcy0123, align 4
+  %mul0123 = fmul <4 x float> %x0123, %y0123
+  %mul0 = extractelement <4 x float> %mul0123, i32 0
+  %mul1 = extractelement <4 x float> %mul0123, i32 1
+  %dot01 = fadd float %mul0, %mul1
+  ret float %dot01
+}
+
+define float @dot2_float2(float* dereferenceable(16) %a0, float* dereferenceable(16) %a1) {
+; SSE2-LABEL: dot2_float2:
+; SSE2:       # %bb.0:
+; SSE2-NEXT:    movsd {{.*#+}} xmm0 = mem[0],zero
+; SSE2-NEXT:    movsd {{.*#+}} xmm1 = mem[0],zero
+; SSE2-NEXT:    mulps %xmm0, %xmm1
+; SSE2-NEXT:    movaps %xmm1, %xmm0
+; SSE2-NEXT:    shufps {{.*#+}} xmm0 = xmm0[1,1],xmm1[1,1]
+; SSE2-NEXT:    addss %xmm1, %xmm0
+; SSE2-NEXT:    retq
+;
+; SSSE3-LABEL: dot2_float2:
+; SSSE3:       # %bb.0:
+; SSSE3-NEXT:    movsd {{.*#+}} xmm0 = mem[0],zero
+; SSSE3-NEXT:    movsd {{.*#+}} xmm1 = mem[0],zero
+; SSSE3-NEXT:    mulps %xmm0, %xmm1
+; SSSE3-NEXT:    movshdup {{.*#+}} xmm0 = xmm1[1,1,3,3]
+; SSSE3-NEXT:    addss %xmm1, %xmm0
+; SSSE3-NEXT:    retq
+;
+; SSE41-LABEL: dot2_float2:
+; SSE41:       # %bb.0:
+; SSE41-NEXT:    movsd {{.*#+}} xmm0 = mem[0],zero
+; SSE41-NEXT:    movsd {{.*#+}} xmm1 = mem[0],zero
+; SSE41-NEXT:    mulps %xmm0, %xmm1
+; SSE41-NEXT:    movshdup {{.*#+}} xmm0 = xmm1[1,1,3,3]
+; SSE41-NEXT:    addss %xmm1, %xmm0
+; SSE41-NEXT:    retq
+;
+; AVX-LABEL: dot2_float2:
+; AVX:       # %bb.0:
+; AVX-NEXT:    vmovsd {{.*#+}} xmm0 = mem[0],zero
+; AVX-NEXT:    vmovsd {{.*#+}} xmm1 = mem[0],zero
+; AVX-NEXT:    vmulps %xmm1, %xmm0, %xmm0
+; AVX-NEXT:    vmovshdup {{.*#+}} xmm1 = xmm0[1,1,3,3]
+; AVX-NEXT:    vaddss %xmm1, %xmm0, %xmm0
+; AVX-NEXT:    retq
+  %bcx01 = bitcast float* %a0 to <2 x float>*
+  %bcy01 = bitcast float* %a1 to <2 x float>*
+  %x01 = load <2 x float>, <2 x float>* %bcx01, align 4
+  %y01 = load <2 x float>, <2 x float>* %bcy01, align 4
+  %mul01 = fmul <2 x float> %x01, %y01
+  %mul0 = extractelement <2 x float> %mul01, i32 0
+  %mul1 = extractelement <2 x float> %mul01, i32 1
+  %dot01 = fadd float %mul0, %mul1
+  ret float %dot01
+}


        


More information about the llvm-commits mailing list