[llvm] [LoongArch] Pre-commit for vecreduce_add. (PR #154302)
via llvm-commits
llvm-commits at lists.llvm.org
Tue Aug 19 02:41:36 PDT 2025
llvmbot wrote:
<!--LLVM PR SUMMARY COMMENT-->
@llvm/pr-subscribers-backend-loongarch
Author: None (tangaac)
<details>
<summary>Changes</summary>
---
Full diff: https://github.com/llvm/llvm-project/pull/154302.diff
2 Files Affected:
- (added) llvm/test/CodeGen/LoongArch/lasx/vec-reduce-add.ll (+96)
- (added) llvm/test/CodeGen/LoongArch/lsx/vec-reduce-add.ll (+70)
``````````diff
diff --git a/llvm/test/CodeGen/LoongArch/lasx/vec-reduce-add.ll b/llvm/test/CodeGen/LoongArch/lasx/vec-reduce-add.ll
new file mode 100644
index 0000000000000..bf5effd7b3912
--- /dev/null
+++ b/llvm/test/CodeGen/LoongArch/lasx/vec-reduce-add.ll
@@ -0,0 +1,96 @@
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 5
+
+; RUN: llc --mtriple=loongarch64 --mattr=+lasx %s -o - | FileCheck %s
+
+define void @vec_reduce_add_v32i8(ptr %src, ptr %dst) nounwind {
+; CHECK-LABEL: vec_reduce_add_v32i8:
+; CHECK: # %bb.0:
+; CHECK-NEXT: xvld $xr0, $a0, 0
+; CHECK-NEXT: xvpermi.d $xr1, $xr0, 78
+; CHECK-NEXT: xvshuf4i.b $xr1, $xr1, 228
+; CHECK-NEXT: xvadd.b $xr0, $xr0, $xr1
+; CHECK-NEXT: xvpermi.d $xr1, $xr0, 68
+; CHECK-NEXT: xvbsrl.v $xr1, $xr1, 8
+; CHECK-NEXT: xvadd.b $xr0, $xr0, $xr1
+; CHECK-NEXT: xvpermi.d $xr1, $xr0, 68
+; CHECK-NEXT: xvsrli.d $xr1, $xr1, 32
+; CHECK-NEXT: xvadd.b $xr0, $xr0, $xr1
+; CHECK-NEXT: xvpermi.d $xr1, $xr0, 68
+; CHECK-NEXT: xvshuf4i.b $xr1, $xr1, 14
+; CHECK-NEXT: xvadd.b $xr0, $xr0, $xr1
+; CHECK-NEXT: xvpermi.d $xr1, $xr0, 68
+; CHECK-NEXT: xvrepl128vei.b $xr1, $xr1, 1
+; CHECK-NEXT: xvadd.b $xr0, $xr0, $xr1
+; CHECK-NEXT: xvstelm.b $xr0, $a1, 0, 0
+; CHECK-NEXT: ret
+ %v = load <32 x i8>, ptr %src
+ %res = call i8 @llvm.vector.reduce.add.v32i8(<32 x i8> %v)
+ store i8 %res, ptr %dst
+ ret void
+}
+
+define void @vec_reduce_add_v16i16(ptr %src, ptr %dst) nounwind {
+; CHECK-LABEL: vec_reduce_add_v16i16:
+; CHECK: # %bb.0:
+; CHECK-NEXT: xvld $xr0, $a0, 0
+; CHECK-NEXT: xvpermi.d $xr1, $xr0, 78
+; CHECK-NEXT: xvshuf4i.h $xr1, $xr1, 228
+; CHECK-NEXT: xvadd.h $xr0, $xr0, $xr1
+; CHECK-NEXT: xvpermi.d $xr1, $xr0, 68
+; CHECK-NEXT: xvbsrl.v $xr1, $xr1, 8
+; CHECK-NEXT: xvadd.h $xr0, $xr0, $xr1
+; CHECK-NEXT: xvpermi.d $xr1, $xr0, 68
+; CHECK-NEXT: xvshuf4i.h $xr1, $xr1, 14
+; CHECK-NEXT: xvadd.h $xr0, $xr0, $xr1
+; CHECK-NEXT: xvpermi.d $xr1, $xr0, 68
+; CHECK-NEXT: xvrepl128vei.h $xr1, $xr1, 1
+; CHECK-NEXT: xvadd.h $xr0, $xr0, $xr1
+; CHECK-NEXT: xvstelm.h $xr0, $a1, 0, 0
+; CHECK-NEXT: ret
+ %v = load <16 x i16>, ptr %src
+ %res = call i16 @llvm.vector.reduce.add.v16i16(<16 x i16> %v)
+ store i16 %res, ptr %dst
+ ret void
+}
+
+define void @vec_reduce_add_v8i32(ptr %src, ptr %dst) nounwind {
+; CHECK-LABEL: vec_reduce_add_v8i32:
+; CHECK: # %bb.0:
+; CHECK-NEXT: xvld $xr0, $a0, 0
+; CHECK-NEXT: xvpermi.d $xr1, $xr0, 78
+; CHECK-NEXT: xvshuf4i.w $xr1, $xr1, 228
+; CHECK-NEXT: xvadd.w $xr0, $xr0, $xr1
+; CHECK-NEXT: xvpermi.d $xr1, $xr0, 68
+; CHECK-NEXT: xvshuf4i.w $xr1, $xr1, 14
+; CHECK-NEXT: xvadd.w $xr0, $xr0, $xr1
+; CHECK-NEXT: xvpermi.d $xr1, $xr0, 68
+; CHECK-NEXT: xvrepl128vei.w $xr1, $xr1, 1
+; CHECK-NEXT: xvadd.w $xr0, $xr0, $xr1
+; CHECK-NEXT: xvstelm.w $xr0, $a1, 0, 0
+; CHECK-NEXT: ret
+ %v = load <8 x i32>, ptr %src
+ %res = call i32 @llvm.vector.reduce.add.v8i32(<8 x i32> %v)
+ store i32 %res, ptr %dst
+ ret void
+}
+
+define void @vec_reduce_add_v4i64(ptr %src, ptr %dst) nounwind {
+; CHECK-LABEL: vec_reduce_add_v4i64:
+; CHECK: # %bb.0:
+; CHECK-NEXT: xvld $xr0, $a0, 0
+; CHECK-NEXT: pcalau12i $a0, %pc_hi20(.LCPI3_0)
+; CHECK-NEXT: xvld $xr1, $a0, %pc_lo12(.LCPI3_0)
+; CHECK-NEXT: xvpermi.d $xr2, $xr0, 78
+; CHECK-NEXT: xvshuf.d $xr1, $xr0, $xr2
+; CHECK-NEXT: xvadd.d $xr0, $xr0, $xr1
+; CHECK-NEXT: xvpermi.d $xr1, $xr0, 68
+; CHECK-NEXT: xvrepl128vei.d $xr1, $xr1, 1
+; CHECK-NEXT: xvadd.d $xr0, $xr0, $xr1
+; CHECK-NEXT: xvstelm.d $xr0, $a1, 0, 0
+; CHECK-NEXT: ret
+ %v = load <4 x i64>, ptr %src
+ %res = call i64 @llvm.vector.reduce.add.v4i64(<4 x i64> %v)
+ store i64 %res, ptr %dst
+ ret void
+}
+
diff --git a/llvm/test/CodeGen/LoongArch/lsx/vec-reduce-add.ll b/llvm/test/CodeGen/LoongArch/lsx/vec-reduce-add.ll
new file mode 100644
index 0000000000000..a71bdea917cba
--- /dev/null
+++ b/llvm/test/CodeGen/LoongArch/lsx/vec-reduce-add.ll
@@ -0,0 +1,70 @@
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 5
+; RUN: llc --mtriple=loongarch64 --mattr=+lsx %s -o - | FileCheck %s
+
+define void @vec_reduce_add_v16i8(ptr %src, ptr %dst) nounwind {
+; CHECK-LABEL: vec_reduce_add_v16i8:
+; CHECK: # %bb.0:
+; CHECK-NEXT: vld $vr0, $a0, 0
+; CHECK-NEXT: vbsrl.v $vr1, $vr0, 8
+; CHECK-NEXT: vadd.b $vr0, $vr0, $vr1
+; CHECK-NEXT: vsrli.d $vr1, $vr0, 32
+; CHECK-NEXT: vadd.b $vr0, $vr0, $vr1
+; CHECK-NEXT: vshuf4i.b $vr1, $vr0, 14
+; CHECK-NEXT: vadd.b $vr0, $vr0, $vr1
+; CHECK-NEXT: vreplvei.b $vr1, $vr0, 1
+; CHECK-NEXT: vadd.b $vr0, $vr0, $vr1
+; CHECK-NEXT: vstelm.b $vr0, $a1, 0, 0
+; CHECK-NEXT: ret
+ %v = load <16 x i8>, ptr %src
+ %res = call i8 @llvm.vector.reduce.add.v16i8(<16 x i8> %v)
+ store i8 %res, ptr %dst
+ ret void
+}
+
+define void @vec_reduce_add_v8i16(ptr %src, ptr %dst) nounwind {
+; CHECK-LABEL: vec_reduce_add_v8i16:
+; CHECK: # %bb.0:
+; CHECK-NEXT: vld $vr0, $a0, 0
+; CHECK-NEXT: vbsrl.v $vr1, $vr0, 8
+; CHECK-NEXT: vadd.h $vr0, $vr0, $vr1
+; CHECK-NEXT: vshuf4i.h $vr1, $vr0, 14
+; CHECK-NEXT: vadd.h $vr0, $vr0, $vr1
+; CHECK-NEXT: vreplvei.h $vr1, $vr0, 1
+; CHECK-NEXT: vadd.h $vr0, $vr0, $vr1
+; CHECK-NEXT: vstelm.h $vr0, $a1, 0, 0
+; CHECK-NEXT: ret
+ %v = load <8 x i16>, ptr %src
+ %res = call i16 @llvm.vector.reduce.add.v8i16(<8 x i16> %v)
+ store i16 %res, ptr %dst
+ ret void
+}
+
+define void @vec_reduce_add_v4i32(ptr %src, ptr %dst) nounwind {
+; CHECK-LABEL: vec_reduce_add_v4i32:
+; CHECK: # %bb.0:
+; CHECK-NEXT: vld $vr0, $a0, 0
+; CHECK-NEXT: vshuf4i.w $vr1, $vr0, 14
+; CHECK-NEXT: vadd.w $vr0, $vr0, $vr1
+; CHECK-NEXT: vreplvei.w $vr1, $vr0, 1
+; CHECK-NEXT: vadd.w $vr0, $vr0, $vr1
+; CHECK-NEXT: vstelm.w $vr0, $a1, 0, 0
+; CHECK-NEXT: ret
+ %v = load <4 x i32>, ptr %src
+ %res = call i32 @llvm.vector.reduce.add.v4i32(<4 x i32> %v)
+ store i32 %res, ptr %dst
+ ret void
+}
+
+define void @vec_reduce_add_v2i64(ptr %src, ptr %dst) nounwind {
+; CHECK-LABEL: vec_reduce_add_v2i64:
+; CHECK: # %bb.0:
+; CHECK-NEXT: vld $vr0, $a0, 0
+; CHECK-NEXT: vreplvei.d $vr1, $vr0, 1
+; CHECK-NEXT: vadd.d $vr0, $vr0, $vr1
+; CHECK-NEXT: vstelm.d $vr0, $a1, 0, 0
+; CHECK-NEXT: ret
+ %v = load <2 x i64>, ptr %src
+ %res = call i64 @llvm.vector.reduce.add.v2i64(<2 x i64> %v)
+ store i64 %res, ptr %dst
+ ret void
+}
``````````
</details>
https://github.com/llvm/llvm-project/pull/154302
More information about the llvm-commits
mailing list