[llvm] 75a5feb - [SystemZ] Don't emit PC-relative memory accesses to unaligned symbols.

Jonas Paulsson via llvm-commits llvm-commits at lists.llvm.org
Tue Sep 29 05:53:46 PDT 2020


Author: Jonas Paulsson
Date: 2020-09-29T14:51:13+02:00
New Revision: 75a5febe31cb2660c4f72d9745625704d29946e1

URL: https://github.com/llvm/llvm-project/commit/75a5febe31cb2660c4f72d9745625704d29946e1
DIFF: https://github.com/llvm/llvm-project/commit/75a5febe31cb2660c4f72d9745625704d29946e1.diff

LOG: [SystemZ] Don't emit PC-relative memory accesses to unaligned symbols.

In the presence of packed structures (#pragma pack(1)) where elements are
referenced through pointers, there will be stores/loads with alignment values
matching the default alignments for the element types while the elements are
in fact unaligned. Strictly speaking this is incorrect source code, but is
unfortunately part of existing code and therefore now addressed.

This patch improves the pattern predicate for PC-relative loads and stores by
not only checking the alignment value of the instruction, but also making
sure that the symbol (and element) itself is aligned.

Fixes https://bugs.llvm.org/show_bug.cgi?id=44405

Review: Ulrich Weigand

Differential Revision: https://reviews.llvm.org/D87510

Added: 
    llvm/test/CodeGen/SystemZ/int-move-10.ll

Modified: 
    llvm/lib/Target/SystemZ/SystemZISelDAGToDAG.cpp
    llvm/lib/Target/SystemZ/SystemZOperators.td

Removed: 
    


################################################################################
diff  --git a/llvm/lib/Target/SystemZ/SystemZISelDAGToDAG.cpp b/llvm/lib/Target/SystemZ/SystemZISelDAGToDAG.cpp
index 37328684399b..9d90a4940cba 100644
--- a/llvm/lib/Target/SystemZ/SystemZISelDAGToDAG.cpp
+++ b/llvm/lib/Target/SystemZ/SystemZISelDAGToDAG.cpp
@@ -338,6 +338,10 @@ class SystemZDAGToDAGISel : public SelectionDAGISel {
   // to X.
   bool storeLoadCanUseBlockBinary(SDNode *N, unsigned I) const;
 
+  // Return true if N (a load or a store) fullfills the alignment
+  // requirements for a PC-relative access.
+  bool storeLoadIsAligned(SDNode *N) const;
+
   // Try to expand a boolean SELECT_CCMASK using an IPM sequence.
   SDValue expandSelectBoolean(SDNode *Node);
 
@@ -1460,6 +1464,46 @@ bool SystemZDAGToDAGISel::storeLoadCanUseBlockBinary(SDNode *N,
          canUseBlockOperation(StoreA, LoadB);
 }
 
+bool SystemZDAGToDAGISel::storeLoadIsAligned(SDNode *N) const {
+
+  auto *MemAccess = cast<LSBaseSDNode>(N);
+  TypeSize StoreSize = MemAccess->getMemoryVT().getStoreSize();
+  SDValue BasePtr = MemAccess->getBasePtr();
+  MachineMemOperand *MMO = MemAccess->getMemOperand();
+  assert(MMO && "Expected a memory operand.");
+
+  // The memory access must have a proper alignment and no index register.
+  if (MemAccess->getAlignment() < StoreSize ||
+      !MemAccess->getOffset().isUndef())
+    return false;
+
+  // The MMO must not have an unaligned offset.
+  if (MMO->getOffset() % StoreSize != 0)
+    return false;
+
+  // An access to GOT or the Constant Pool is aligned.
+  if (const PseudoSourceValue *PSV = MMO->getPseudoValue())
+    if ((PSV->isGOT() || PSV->isConstantPool()))
+      return true;
+
+  // Check the alignment of a Global Address.
+  if (BasePtr.getNumOperands())
+    if (GlobalAddressSDNode *GA =
+        dyn_cast<GlobalAddressSDNode>(BasePtr.getOperand(0))) {
+      // The immediate offset must be aligned.
+      if (GA->getOffset() % StoreSize != 0)
+        return false;
+
+      // The alignment of the symbol itself must be at least the store size.
+      const GlobalValue *GV = GA->getGlobal();
+      const DataLayout &DL = GV->getParent()->getDataLayout();
+      if (GV->getPointerAlignment(DL).value() < StoreSize)
+        return false;
+    }
+
+  return true;
+}
+
 void SystemZDAGToDAGISel::Select(SDNode *Node) {
   // If we have a custom node, we already have selected!
   if (Node->isMachineOpcode()) {

diff  --git a/llvm/lib/Target/SystemZ/SystemZOperators.td b/llvm/lib/Target/SystemZ/SystemZOperators.td
index 81af5fd854db..a5f29f29a706 100644
--- a/llvm/lib/Target/SystemZ/SystemZOperators.td
+++ b/llvm/lib/Target/SystemZ/SystemZOperators.td
@@ -572,10 +572,8 @@ def anyextloadi32 : PatFrag<(ops node:$ptr), (anyextload node:$ptr), [{
 
 // Aligned loads.
 class AlignedLoad<SDPatternOperator load>
-  : PatFrag<(ops node:$addr), (load node:$addr), [{
-  auto *Load = cast<LoadSDNode>(N);
-  return Load->getAlignment() >= Load->getMemoryVT().getStoreSize();
-}]>;
+  : PatFrag<(ops node:$addr), (load node:$addr),
+  [{ return storeLoadIsAligned(N); }]>;
 def aligned_load         : AlignedLoad<load>;
 def aligned_asextloadi16 : AlignedLoad<asextloadi16>;
 def aligned_asextloadi32 : AlignedLoad<asextloadi32>;
@@ -584,10 +582,8 @@ def aligned_azextloadi32 : AlignedLoad<azextloadi32>;
 
 // Aligned stores.
 class AlignedStore<SDPatternOperator store>
-  : PatFrag<(ops node:$src, node:$addr), (store node:$src, node:$addr), [{
-  auto *Store = cast<StoreSDNode>(N);
-  return Store->getAlignment() >= Store->getMemoryVT().getStoreSize();
-}]>;
+  : PatFrag<(ops node:$src, node:$addr), (store node:$src, node:$addr),
+  [{ return storeLoadIsAligned(N); }]>;
 def aligned_store         : AlignedStore<store>;
 def aligned_truncstorei16 : AlignedStore<truncstorei16>;
 def aligned_truncstorei32 : AlignedStore<truncstorei32>;

diff  --git a/llvm/test/CodeGen/SystemZ/int-move-10.ll b/llvm/test/CodeGen/SystemZ/int-move-10.ll
new file mode 100644
index 000000000000..8b8b9ed1a94a
--- /dev/null
+++ b/llvm/test/CodeGen/SystemZ/int-move-10.ll
@@ -0,0 +1,209 @@
+; RUN: llc < %s -mtriple=s390x-linux-gnu | FileCheck %s
+;
+; Test PC-relative memory accesses of globals with packed struct types.
+; PC-relative memory accesses cannot be used when the address is not
+; aligned. This can happen with programs like the following (which are not
+; strictly correct):
+;
+; #pragma pack(1)
+; struct  {
+;   short a;
+;   int b;
+; } c;
+;
+; void main()    {
+;   int *e = &c.b;
+;   *e = 0;
+; }
+;
+
+%packed.i16i32 = type <{ i16, i32 }>
+%packed.i16i32i16i32 = type <{ i16, i32, i16, i32 }>
+%packed.i16i64 = type <{ i16, i64 }>
+%packed.i8i16 = type <{ i8, i16 }>
+
+ at A_align2 = global %packed.i16i32 zeroinitializer, align 2
+ at B_align2 = global %packed.i16i32i16i32 zeroinitializer, align 2
+ at C_align2 = global %packed.i16i64 zeroinitializer, align 2
+ at D_align4 = global %packed.i16i32 zeroinitializer, align 4
+ at E_align4 = global %packed.i16i32i16i32 zeroinitializer, align 4
+ at F_align2 = global %packed.i8i16 zeroinitializer, align 2
+
+;;; Stores
+
+; unaligned packed struct + 2  -> unaligned address
+define void @f1() {
+; CHECK-LABEL: f1:
+; CHECK: larl %r1, A_align2
+; CHECK: mvhi 2(%r1), 0
+; CHECK: br %r14
+  store i32 0, i32* getelementptr inbounds (%packed.i16i32, %packed.i16i32* @A_align2, i64 0, i32 1), align 4
+  ret void
+}
+
+; unaligned packed struct  + 8  -> unaligned address
+define void @f2() {
+; CHECK-LABEL: f2:
+; CHECK: larl %r1, B_align2
+; CHECK: mvhi 8(%r1), 0
+; CHECK: br %r14
+  store i32 0, i32* getelementptr inbounds (%packed.i16i32i16i32, %packed.i16i32i16i32* @B_align2, i64 0, i32 3), align 4
+  ret void
+}
+
+; aligned packed struct + 2  -> unaligned address
+define void @f3() {
+; CHECK-LABEL: f3:
+; CHECK: larl %r1, D_align4
+; CHECK: mvhi 2(%r1), 0
+; CHECK: br %r14
+  store i32 0, i32* getelementptr inbounds (%packed.i16i32, %packed.i16i32* @D_align4, i64 0, i32 1), align 4
+  ret void
+}
+
+; aligned packed struct + 8  -> aligned address
+define void @f4() {
+; CHECK-LABEL: f4:
+; CHECK: lhi %r0, 0
+; CHECK: strl %r0, E_align4+8
+; CHECK: br %r14
+  store i32 0, i32* getelementptr inbounds (%packed.i16i32i16i32, %packed.i16i32i16i32* @E_align4, i64 0, i32 3), align 4
+  ret void
+}
+
+define void @f5() {
+; CHECK-LABEL: f5:
+; CHECK: larl %r1, C_align2
+; CHECK: mvghi 2(%r1), 0
+; CHECK: br %r14
+  store i64 0, i64* getelementptr inbounds (%packed.i16i64, %packed.i16i64* @C_align2, i64 0, i32 1), align 8
+  ret void
+}
+
+define void @f6() {
+; CHECK-LABEL: f6:
+; CHECK-NOT: sthrl
+  store i16 0, i16* getelementptr inbounds (%packed.i8i16, %packed.i8i16* @F_align2, i64 0, i32 1), align 2
+  ret void
+}
+
+define void @f7(i64* %Src) {
+; CHECK-LABEL: f7:
+; CHECK: lg %r0, 0(%r2)
+; CHECK: larl %r1, D_align4
+; CHECK: st %r0, 2(%r1)
+; CHECK: br      %r14
+  %L = load i64, i64* %Src
+  %T = trunc i64 %L to i32
+  store i32 %T, i32* getelementptr inbounds (%packed.i16i32, %packed.i16i32* @D_align4, i64 0, i32 1), align 4
+  ret void
+}
+
+define void @f8(i64* %Src) {
+; CHECK-LABEL: f8:
+; CHECK-NOT: sthrl
+  %L = load i64, i64* %Src
+  %T = trunc i64 %L to i16
+  store i16 %T, i16* getelementptr inbounds (%packed.i8i16, %packed.i8i16* @F_align2, i64 0, i32 1), align 2
+  ret void
+}
+
+;;; Loads
+
+; unaligned packed struct + 2  -> unaligned address
+define i32 @f9() {
+; CHECK-LABEL: f9:
+; CHECK: larl %r1, A_align2
+; CHECK: l %r2, 2(%r1)
+; CHECK: br %r14
+  %L = load i32, i32* getelementptr inbounds (%packed.i16i32, %packed.i16i32* @A_align2, i64 0, i32 1), align 4
+  ret i32 %L
+}
+
+; unaligned packed struct  + 8  -> unaligned address
+define i32 @f10() {
+; CHECK-LABEL: f10:
+; CHECK: larl %r1, B_align2
+; CHECK: l %r2, 8(%r1)
+; CHECK: br %r14
+  %L = load i32, i32* getelementptr inbounds (%packed.i16i32i16i32, %packed.i16i32i16i32* @B_align2, i64 0, i32 3), align 4
+  ret i32 %L
+}
+
+; aligned packed struct + 2  -> unaligned address
+define i32 @f11() {
+; CHECK-LABEL: f11:
+; CHECK: larl %r1, D_align4
+; CHECK: l %r2, 2(%r1)
+; CHECK: br %r14
+  %L = load i32, i32* getelementptr inbounds (%packed.i16i32, %packed.i16i32* @D_align4, i64 0, i32 1), align 4
+  ret i32 %L
+}
+
+; aligned packed struct + 8  -> aligned address
+define i32 @f12() {
+; CHECK-LABEL: f12:
+; CHECK: lrl %r2, E_align4+8
+; CHECK: br %r14
+  %L = load i32, i32* getelementptr inbounds (%packed.i16i32i16i32, %packed.i16i32i16i32* @E_align4, i64 0, i32 3), align 4
+  ret i32 %L
+}
+
+define i64 @f13() {
+; CHECK-LABEL: f13:
+; CHECK: larl %r1, C_align2
+; CHECK: lg %r2, 2(%r1)
+; CHECK: br %r14
+  %L = load i64, i64* getelementptr inbounds (%packed.i16i64, %packed.i16i64* @C_align2, i64 0, i32 1), align 8
+  ret i64 %L
+}
+
+define i32 @f14() {
+; CHECK-LABEL: f14:
+; CHECK-NOT: lhrl
+  %L = load i16, i16* getelementptr inbounds (%packed.i8i16, %packed.i8i16* @F_align2, i64 0, i32 1), align 2
+  %ext = sext i16 %L to i32
+  ret i32 %ext
+}
+
+define i64 @f15() {
+; CHECK-LABEL: f15:
+; CHECK-NOT: llghrl
+  %L = load i16, i16* getelementptr inbounds (%packed.i8i16, %packed.i8i16* @F_align2, i64 0, i32 1), align 2
+  %ext = zext i16 %L to i64
+  ret i64 %ext
+}
+
+;;; Loads folded into compare instructions
+
+define i32 @f16(i32 %src1) {
+; CHECK-LABEL: f16:
+; CHECK: larl %r1, A_align2
+; CHECK: c %r2, 2(%r1)
+entry:
+  %src2 = load i32, i32* getelementptr inbounds (%packed.i16i32, %packed.i16i32* @A_align2, i64 0, i32 1), align 4
+  %cond = icmp slt i32 %src1, %src2
+  br i1 %cond, label %exit, label %mulb
+mulb:
+  %mul = mul i32 %src1, %src1
+  br label %exit
+exit:
+  %res = phi i32 [ %src1, %entry ], [ %mul, %mulb ]
+  ret i32 %res
+}
+
+define i64 @f17(i64 %src1) {
+; CHECK-LABEL: f17:
+; CHECK: larl %r1, C_align2
+; CHECK: clg %r2, 2(%r1)
+entry:
+  %src2 = load i64, i64* getelementptr inbounds (%packed.i16i64, %packed.i16i64* @C_align2, i64 0, i32 1), align 8
+  %cond = icmp ult i64 %src1, %src2
+  br i1 %cond, label %exit, label %mulb
+mulb:
+  %mul = mul i64 %src1, %src1
+  br label %exit
+exit:
+  %res = phi i64 [ %src1, %entry ], [ %mul, %mulb ]
+  ret i64 %res
+}


        


More information about the llvm-commits mailing list