[llvm] r365055 - [SLP] add tests for bitcasted vector pointer load; NFC

Sanjay Patel via llvm-commits llvm-commits at lists.llvm.org
Wed Jul 3 09:46:14 PDT 2019


Author: spatel
Date: Wed Jul  3 09:46:14 2019
New Revision: 365055

URL: http://llvm.org/viewvc/llvm-project?rev=365055&view=rev
Log:
[SLP] add tests for bitcasted vector pointer load; NFC

I'm not sure if this falls within the scope of SLP,
but we could create vector loads for some of these
patterns.

Added:
    llvm/trunk/test/Transforms/SLPVectorizer/X86/load-bitcast-vec.ll

Added: llvm/trunk/test/Transforms/SLPVectorizer/X86/load-bitcast-vec.ll
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/Transforms/SLPVectorizer/X86/load-bitcast-vec.ll?rev=365055&view=auto
==============================================================================
--- llvm/trunk/test/Transforms/SLPVectorizer/X86/load-bitcast-vec.ll (added)
+++ llvm/trunk/test/Transforms/SLPVectorizer/X86/load-bitcast-vec.ll Wed Jul  3 09:46:14 2019
@@ -0,0 +1,102 @@
+; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
+; RUN: opt < %s -slp-vectorizer -S -mtriple=x86_64-- -mattr=+avx  | FileCheck %s --check-prefixes=CHECK,AVX1
+; RUN: opt < %s -slp-vectorizer -S -mtriple=x86_64-- -mattr=+avx2 | FileCheck %s --check-prefixes=CHECK,AVX2
+
+define float @matching_scalar(<4 x float>* dereferenceable(16) %p) {
+; CHECK-LABEL: @matching_scalar(
+; CHECK-NEXT:    [[BC:%.*]] = bitcast <4 x float>* [[P:%.*]] to float*
+; CHECK-NEXT:    [[R:%.*]] = load float, float* [[BC]], align 16
+; CHECK-NEXT:    ret float [[R]]
+;
+  %bc = bitcast <4 x float>* %p to float*
+  %r = load float, float* %bc, align 16
+  ret float %r
+}
+
+define i32 @nonmatching_scalar(<4 x float>* dereferenceable(16) %p) {
+; CHECK-LABEL: @nonmatching_scalar(
+; CHECK-NEXT:    [[BC:%.*]] = bitcast <4 x float>* [[P:%.*]] to i32*
+; CHECK-NEXT:    [[R:%.*]] = load i32, i32* [[BC]], align 16
+; CHECK-NEXT:    ret i32 [[R]]
+;
+  %bc = bitcast <4 x float>* %p to i32*
+  %r = load i32, i32* %bc, align 16
+  ret i32 %r
+}
+
+define i64 @larger_scalar(<4 x float>* dereferenceable(16) %p) {
+; CHECK-LABEL: @larger_scalar(
+; CHECK-NEXT:    [[BC:%.*]] = bitcast <4 x float>* [[P:%.*]] to i64*
+; CHECK-NEXT:    [[R:%.*]] = load i64, i64* [[BC]], align 16
+; CHECK-NEXT:    ret i64 [[R]]
+;
+  %bc = bitcast <4 x float>* %p to i64*
+  %r = load i64, i64* %bc, align 16
+  ret i64 %r
+}
+
+define i8 @smaller_scalar(<4 x float>* dereferenceable(16) %p) {
+; CHECK-LABEL: @smaller_scalar(
+; CHECK-NEXT:    [[BC:%.*]] = bitcast <4 x float>* [[P:%.*]] to i8*
+; CHECK-NEXT:    [[R:%.*]] = load i8, i8* [[BC]], align 16
+; CHECK-NEXT:    ret i8 [[R]]
+;
+  %bc = bitcast <4 x float>* %p to i8*
+  %r = load i8, i8* %bc, align 16
+  ret i8 %r
+}
+
+define i8 @smaller_scalar_256bit_vec(<8 x float>* dereferenceable(32) %p) {
+; CHECK-LABEL: @smaller_scalar_256bit_vec(
+; CHECK-NEXT:    [[BC:%.*]] = bitcast <8 x float>* [[P:%.*]] to i8*
+; CHECK-NEXT:    [[R:%.*]] = load i8, i8* [[BC]], align 32
+; CHECK-NEXT:    ret i8 [[R]]
+;
+  %bc = bitcast <8 x float>* %p to i8*
+  %r = load i8, i8* %bc, align 32
+  ret i8 %r
+}
+
+define i8 @smaller_scalar_less_aligned(<4 x float>* dereferenceable(16) %p) {
+; CHECK-LABEL: @smaller_scalar_less_aligned(
+; CHECK-NEXT:    [[BC:%.*]] = bitcast <4 x float>* [[P:%.*]] to i8*
+; CHECK-NEXT:    [[R:%.*]] = load i8, i8* [[BC]], align 4
+; CHECK-NEXT:    ret i8 [[R]]
+;
+  %bc = bitcast <4 x float>* %p to i8*
+  %r = load i8, i8* %bc, align 4
+  ret i8 %r
+}
+
+define float @matching_scalar_small_deref(<4 x float>* dereferenceable(15) %p) {
+; CHECK-LABEL: @matching_scalar_small_deref(
+; CHECK-NEXT:    [[BC:%.*]] = bitcast <4 x float>* [[P:%.*]] to float*
+; CHECK-NEXT:    [[R:%.*]] = load float, float* [[BC]], align 16
+; CHECK-NEXT:    ret float [[R]]
+;
+  %bc = bitcast <4 x float>* %p to float*
+  %r = load float, float* %bc, align 16
+  ret float %r
+}
+
+define float @matching_scalar_volatile(<4 x float>* dereferenceable(16) %p) {
+; CHECK-LABEL: @matching_scalar_volatile(
+; CHECK-NEXT:    [[BC:%.*]] = bitcast <4 x float>* [[P:%.*]] to float*
+; CHECK-NEXT:    [[R:%.*]] = load volatile float, float* [[BC]], align 16
+; CHECK-NEXT:    ret float [[R]]
+;
+  %bc = bitcast <4 x float>* %p to float*
+  %r = load volatile float, float* %bc, align 16
+  ret float %r
+}
+
+define float @nonvector(double* dereferenceable(16) %p) {
+; CHECK-LABEL: @nonvector(
+; CHECK-NEXT:    [[BC:%.*]] = bitcast double* [[P:%.*]] to float*
+; CHECK-NEXT:    [[R:%.*]] = load float, float* [[BC]], align 16
+; CHECK-NEXT:    ret float [[R]]
+;
+  %bc = bitcast double* %p to float*
+  %r = load float, float* %bc, align 16
+  ret float %r
+}




More information about the llvm-commits mailing list