[llvm] 76db6d8 - [NFC][LoopIdiom] Add more test case to runtime-determined memset size

via llvm-commits llvm-commits at lists.llvm.org
Thu Oct 21 00:05:24 PDT 2021


Author: eopXD
Date: 2021-10-21T00:05:18-07:00
New Revision: 76db6d80805a70802fa9796ca1a5bb999ff205c7

URL: https://github.com/llvm/llvm-project/commit/76db6d80805a70802fa9796ca1a5bb999ff205c7
DIFF: https://github.com/llvm/llvm-project/commit/76db6d80805a70802fa9796ca1a5bb999ff205c7.diff

LOG: [NFC][LoopIdiom] Add more test case to runtime-determined memset size

This patch supplements missing test case for D107353.
- Fix wrong descriptions in 64-bit mode test case
- Added testcase under 32-bit mode

Reviewed By: bmahjour

Differential Revision: https://reviews.llvm.org/D108507

Added: 
    llvm/test/Transforms/LoopIdiom/memset-runtime-32bit.ll
    llvm/test/Transforms/LoopIdiom/memset-runtime-64bit.ll

Modified: 
    

Removed: 
    llvm/test/Transforms/LoopIdiom/memset-runtime.ll


################################################################################
diff  --git a/llvm/test/Transforms/LoopIdiom/memset-runtime-32bit.ll b/llvm/test/Transforms/LoopIdiom/memset-runtime-32bit.ll
new file mode 100644
index 0000000000000..fe6ad07a1feba
--- /dev/null
+++ b/llvm/test/Transforms/LoopIdiom/memset-runtime-32bit.ll
@@ -0,0 +1,372 @@
+; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
+; RUN: opt -passes="function(loop(indvars,loop-idiom,loop-deletion),simplifycfg)" -S < %s | FileCheck %s
+; Compile command:
+; $ clang -m32 -fno-discard-value-names -O0 -S -emit-llvm -Xclang -disable-O0-optnone Code.c
+; $ bin/opt -S -basic-aa -mem2reg -loop-simplify -lcssa -loop-rotate \
+;   -licm -simple-loop-unswitch -enable-nontrivial-unswitch -loop-simplify \
+;   -loop-deletion -simplifycfg -indvars Code.ll > CodeOpt.ll
+target datalayout = "e-m:e-p:32:32-p270:32:32-p271:32:32-p272:64:64-f64:32:64-f80:32-n8:16:32-S128"
+; void PositiveFor32(int *ar, int n, int m)
+; {
+;   int i;
+;   for (i=0; i<n; ++i) {
+;     int *arr = ar + i * m;
+;     memset(arr, 0, m * sizeof(int));
+;   }
+; }
+define dso_local void @PositiveFor32(i32* %ar, i32 %n, i32 %m) {
+; CHECK-LABEL: @PositiveFor32(
+; CHECK-NEXT:  entry:
+; CHECK-NEXT:    [[AR1:%.*]] = bitcast i32* [[AR:%.*]] to i8*
+; CHECK-NEXT:    [[CMP1:%.*]] = icmp slt i32 0, [[N:%.*]]
+; CHECK-NEXT:    br i1 [[CMP1]], label [[FOR_BODY_LR_PH:%.*]], label [[FOR_END:%.*]]
+; CHECK:       for.body.lr.ph:
+; CHECK-NEXT:    [[MUL1:%.*]] = mul i32 [[M:%.*]], 4
+; CHECK-NEXT:    [[TMP0:%.*]] = mul i32 [[M]], [[N]]
+; CHECK-NEXT:    [[TMP1:%.*]] = shl i32 [[TMP0]], 2
+; CHECK-NEXT:    call void @llvm.memset.p0i8.i32(i8* align 4 [[AR1]], i8 0, i32 [[TMP1]], i1 false)
+; CHECK-NEXT:    br label [[FOR_END]]
+; CHECK:       for.end:
+; CHECK-NEXT:    ret void
+;
+entry:
+  %cmp1 = icmp slt i32 0, %n
+  br i1 %cmp1, label %for.body.lr.ph, label %for.end
+
+for.body.lr.ph:                                   ; preds = %entry
+  %mul1 = mul i32 %m, 4
+  br label %for.body
+
+for.body:                                         ; preds = %for.body.lr.ph, %for.body
+  %i.02 = phi i32 [ 0, %for.body.lr.ph ], [ %inc, %for.body ]
+  %mul = mul nsw i32 %i.02, %m
+  %add.ptr = getelementptr inbounds i32, i32* %ar, i32 %mul
+  %0 = bitcast i32* %add.ptr to i8*
+  call void @llvm.memset.p0i8.i32(i8* align 4 %0, i8 0, i32 %mul1, i1 false)
+  %inc = add nsw i32 %i.02, 1
+  %cmp = icmp slt i32 %inc, %n
+  br i1 %cmp, label %for.body, label %for.end
+
+for.end:                                          ; preds = %for.body, %entry
+  ret void
+}
+; void Negative32(int *ar, int n, int m)
+; {
+;   int i;
+;   for (i=n-1; i>=0; i--) {
+;     int *arr = ar + i * m;
+;     memset(arr, 0, m * sizeof(int));
+;   }
+; }
+define void @NegativeFor32(i32* %ar, i32 %n, i32 %m) {
+; CHECK-LABEL: @NegativeFor32(
+; CHECK-NEXT:  entry:
+; CHECK-NEXT:    [[AR1:%.*]] = bitcast i32* [[AR:%.*]] to i8*
+; CHECK-NEXT:    [[SUB:%.*]] = sub nsw i32 [[N:%.*]], 1
+; CHECK-NEXT:    [[CMP1:%.*]] = icmp sge i32 [[SUB]], 0
+; CHECK-NEXT:    br i1 [[CMP1]], label [[FOR_BODY_LR_PH:%.*]], label [[FOR_END:%.*]]
+; CHECK:       for.body.lr.ph:
+; CHECK-NEXT:    [[MUL1:%.*]] = mul i32 [[M:%.*]], 4
+; CHECK-NEXT:    [[TMP0:%.*]] = mul i32 [[M]], [[N]]
+; CHECK-NEXT:    [[TMP1:%.*]] = shl i32 [[TMP0]], 2
+; CHECK-NEXT:    call void @llvm.memset.p0i8.i32(i8* align 4 [[AR1]], i8 0, i32 [[TMP1]], i1 false)
+; CHECK-NEXT:    br label [[FOR_END]]
+; CHECK:       for.end:
+; CHECK-NEXT:    ret void
+;
+entry:
+  %sub = sub nsw i32 %n, 1
+  %cmp1 = icmp sge i32 %sub, 0
+  br i1 %cmp1, label %for.body.lr.ph, label %for.end
+
+for.body.lr.ph:                                   ; preds = %entry
+  %mul1 = mul i32 %m, 4
+  br label %for.body
+
+for.body:                                         ; preds = %for.body.lr.ph, %for.body
+  %i.02 = phi i32 [ %sub, %for.body.lr.ph ], [ %dec, %for.body ]
+  %mul = mul nsw i32 %i.02, %m
+  %add.ptr = getelementptr inbounds i32, i32* %ar, i32 %mul
+  %0 = bitcast i32* %add.ptr to i8*
+  call void @llvm.memset.p0i8.i32(i8* align 4 %0, i8 0, i32 %mul1, i1 false)
+  %dec = add nsw i32 %i.02, -1
+  %cmp = icmp sge i32 %dec, 0
+  br i1 %cmp, label %for.body, label %for.end
+
+for.end:                                          ; preds = %for.body, %entry
+  ret void
+}
+; void NestedFor32(int *ar, int n, int m, int o)
+; {
+;   int i, j;
+;   for (i=0; i<n; ++i) {
+;     for (j=0; j<m; ++j) {
+;       int *arr = ar + i * m * o + j * o;
+;       memset(arr, 0, o * sizeof(int));
+;     }
+;   }
+; }
+define void @NestedFor32(i32* %ar, i32 %n, i32 %m, i32 %o) {
+; CHECK-LABEL: @NestedFor32(
+; CHECK-NEXT:  entry:
+; CHECK-NEXT:    [[AR2:%.*]] = bitcast i32* [[AR:%.*]] to i8*
+; CHECK-NEXT:    [[CMP3:%.*]] = icmp slt i32 0, [[N:%.*]]
+; CHECK-NEXT:    [[CMP21:%.*]] = icmp slt i32 0, [[M:%.*]]
+; CHECK-NEXT:    [[MUL7:%.*]] = mul i32 [[O:%.*]], 4
+; CHECK-NEXT:    [[OR_COND:%.*]] = select i1 [[CMP3]], i1 [[CMP21]], i1 false
+; CHECK-NEXT:    br i1 [[OR_COND]], label [[FOR_BODY_US_PREHEADER:%.*]], label [[FOR_END10:%.*]]
+; CHECK:       for.body.us.preheader:
+; CHECK-NEXT:    [[TMP0:%.*]] = mul i32 [[O]], [[M]]
+; CHECK-NEXT:    [[TMP1:%.*]] = shl i32 [[TMP0]], 2
+; CHECK-NEXT:    [[TMP2:%.*]] = mul i32 [[TMP0]], [[N]]
+; CHECK-NEXT:    [[TMP3:%.*]] = shl i32 [[TMP2]], 2
+; CHECK-NEXT:    call void @llvm.memset.p0i8.i32(i8* align 4 [[AR2]], i8 0, i32 [[TMP3]], i1 false)
+; CHECK-NEXT:    br label [[FOR_END10]]
+; CHECK:       for.end10:
+; CHECK-NEXT:    ret void
+;
+entry:
+  %cmp3 = icmp slt i32 0, %n
+  br i1 %cmp3, label %for.body.lr.ph, label %for.end10
+
+for.body.lr.ph:                                   ; preds = %entry
+  %cmp21 = icmp slt i32 0, %m
+  %mul7 = mul i32 %o, 4
+  br i1 %cmp21, label %for.body.us.preheader, label %for.end10
+
+for.body.us.preheader:                            ; preds = %for.body.lr.ph
+  br label %for.body.us
+
+for.body.us:                                      ; preds = %for.body.us.preheader, %for.cond1.for.end_crit_edge.us
+  %i.04.us = phi i32 [ %inc9.us, %for.cond1.for.end_crit_edge.us ], [ 0, %for.body.us.preheader ]
+  %mul.us = mul nsw i32 %i.04.us, %m
+  %mul4.us = mul nsw i32 %mul.us, %o
+  %add.ptr.us = getelementptr inbounds i32, i32* %ar, i32 %mul4.us
+  br label %for.body3.us
+
+for.body3.us:                                     ; preds = %for.body.us, %for.body3.us
+  %j.02.us = phi i32 [ 0, %for.body.us ], [ %inc.us, %for.body3.us ]
+  %mul5.us = mul nsw i32 %j.02.us, %o
+  %add.ptr6.us = getelementptr inbounds i32, i32* %add.ptr.us, i32 %mul5.us
+  %0 = bitcast i32* %add.ptr6.us to i8*
+  call void @llvm.memset.p0i8.i32(i8* align 4 %0, i8 0, i32 %mul7, i1 false)
+  %inc.us = add nuw nsw i32 %j.02.us, 1
+  %exitcond = icmp ne i32 %inc.us, %m
+  br i1 %exitcond, label %for.body3.us, label %for.cond1.for.end_crit_edge.us
+
+for.cond1.for.end_crit_edge.us:                   ; preds = %for.body3.us
+  %inc9.us = add nuw nsw i32 %i.04.us, 1
+  %exitcond5 = icmp ne i32 %inc9.us, %n
+  br i1 %exitcond5, label %for.body.us, label %for.end10.loopexit
+
+for.end10.loopexit:                               ; preds = %for.cond1.for.end_crit_edge.us
+  br label %for.end10
+
+for.end10:                                        ; preds = %for.end10.loopexit, %for.body.lr.ph, %entry
+  ret void
+}
+; void PositiveFor64(int *ar, long long n, long long m)
+; {
+;   int i;
+;   for (i=0; i<n; ++i) {
+;     int *arr = ar + i * m;
+;     memset(arr, 0, m * sizeof(int));
+;   }
+; }
+define dso_local void @PositiveFor64(i32* %ar, i64 %n, i64 %m) #0 {
+; CHECK-LABEL: @PositiveFor64(
+; CHECK-NEXT:  entry:
+; CHECK-NEXT:    [[AR1:%.*]] = bitcast i32* [[AR:%.*]] to i8*
+; CHECK-NEXT:    [[CMP1:%.*]] = icmp slt i64 0, [[N:%.*]]
+; CHECK-NEXT:    br i1 [[CMP1]], label [[FOR_BODY_LR_PH:%.*]], label [[FOR_END:%.*]]
+; CHECK:       for.body.lr.ph:
+; CHECK-NEXT:    [[MUL3:%.*]] = mul nsw i64 [[M:%.*]], 4
+; CHECK-NEXT:    [[CONV4:%.*]] = trunc i64 [[MUL3]] to i32
+; CHECK-NEXT:    [[TMP0:%.*]] = trunc i64 [[M]] to i32
+; CHECK-NEXT:    [[TMP1:%.*]] = trunc i64 [[N]] to i32
+; CHECK-NEXT:    [[TMP2:%.*]] = mul i32 [[TMP0]], [[TMP1]]
+; CHECK-NEXT:    [[TMP3:%.*]] = shl i32 [[TMP2]], 2
+; CHECK-NEXT:    call void @llvm.memset.p0i8.i32(i8* align 4 [[AR1]], i8 0, i32 [[TMP3]], i1 false)
+; CHECK-NEXT:    br label [[FOR_END]]
+; CHECK:       for.end:
+; CHECK-NEXT:    ret void
+;
+entry:
+  %cmp1 = icmp slt i64 0, %n
+  br i1 %cmp1, label %for.body.lr.ph, label %for.end
+
+for.body.lr.ph:                                   ; preds = %entry
+  %mul3 = mul nsw i64 %m, 4
+  %conv4 = trunc i64 %mul3 to i32
+  br label %for.body
+
+for.body:                                         ; preds = %for.body.lr.ph, %for.body
+  %i.02 = phi i32 [ 0, %for.body.lr.ph ], [ %inc, %for.body ]
+  %conv2 = sext i32 %i.02 to i64
+  %mul = mul nsw i64 %conv2, %m
+  %idx.ext = trunc i64 %mul to i32
+  %add.ptr = getelementptr inbounds i32, i32* %ar, i32 %idx.ext
+  %0 = bitcast i32* %add.ptr to i8*
+  call void @llvm.memset.p0i8.i32(i8* align 4 %0, i8 0, i32 %conv4, i1 false)
+  %inc = add nuw nsw i32 %i.02, 1
+  %conv = sext i32 %inc to i64
+  %cmp = icmp slt i64 %conv, %n
+  br i1 %cmp, label %for.body, label %for.end.loopexit
+
+for.end.loopexit:                                 ; preds = %for.body
+  br label %for.end
+
+for.end:                                          ; preds = %for.end.loopexit, %entry
+  ret void
+}
+; void NegativeFor64(int *ar, long long n, long long m)
+; {
+;   int i;
+;   for (i=n-1; i>=0; --i) {
+;     int *arr = ar + i * m;
+;     memset(arr, 0, m * sizeof(int));
+;   }
+; }
+define dso_local void @NegativeFor64(i32* %ar, i64 %n, i64 %m) #0 {
+; CHECK-LABEL: @NegativeFor64(
+; CHECK-NEXT:  entry:
+; CHECK-NEXT:    [[AR1:%.*]] = bitcast i32* [[AR:%.*]] to i8*
+; CHECK-NEXT:    [[SUB:%.*]] = sub nsw i64 [[N:%.*]], 1
+; CHECK-NEXT:    [[CONV:%.*]] = trunc i64 [[SUB]] to i32
+; CHECK-NEXT:    [[CMP1:%.*]] = icmp sge i32 [[CONV]], 0
+; CHECK-NEXT:    br i1 [[CMP1]], label [[FOR_BODY_LR_PH:%.*]], label [[FOR_END:%.*]]
+; CHECK:       for.body.lr.ph:
+; CHECK-NEXT:    [[MUL3:%.*]] = mul nsw i64 [[M:%.*]], 4
+; CHECK-NEXT:    [[CONV4:%.*]] = trunc i64 [[MUL3]] to i32
+; CHECK-NEXT:    [[TMP0:%.*]] = trunc i64 [[M]] to i32
+; CHECK-NEXT:    [[TMP1:%.*]] = sub i32 [[CONV]], -1
+; CHECK-NEXT:    [[TMP2:%.*]] = mul i32 [[TMP0]], [[TMP1]]
+; CHECK-NEXT:    [[TMP3:%.*]] = shl i32 [[TMP2]], 2
+; CHECK-NEXT:    call void @llvm.memset.p0i8.i32(i8* align 4 [[AR1]], i8 0, i32 [[TMP3]], i1 false)
+; CHECK-NEXT:    br label [[FOR_END]]
+; CHECK:       for.end:
+; CHECK-NEXT:    ret void
+;
+entry:
+  %sub = sub nsw i64 %n, 1
+  %conv = trunc i64 %sub to i32
+  %cmp1 = icmp sge i32 %conv, 0
+  br i1 %cmp1, label %for.body.lr.ph, label %for.end
+
+for.body.lr.ph:                                   ; preds = %entry
+  %mul3 = mul nsw i64 %m, 4
+  %conv4 = trunc i64 %mul3 to i32
+  br label %for.body
+
+for.body:                                         ; preds = %for.body.lr.ph, %for.body
+  %i.02 = phi i32 [ %conv, %for.body.lr.ph ], [ %dec, %for.body ]
+  %conv2 = sext i32 %i.02 to i64
+  %mul = mul nsw i64 %conv2, %m
+  %idx.ext = trunc i64 %mul to i32
+  %add.ptr = getelementptr inbounds i32, i32* %ar, i32 %idx.ext
+  %0 = bitcast i32* %add.ptr to i8*
+  call void @llvm.memset.p0i8.i32(i8* align 4 %0, i8 0, i32 %conv4, i1 false)
+  %dec = add nsw i32 %i.02, -1
+  %cmp = icmp sge i32 %dec, 0
+  br i1 %cmp, label %for.body, label %for.end.loopexit
+
+for.end.loopexit:                                 ; preds = %for.body
+  br label %for.end
+
+for.end:                                          ; preds = %for.end.loopexit, %entry
+  ret void
+}
+; void NestedFor64(int *ar, long long n, long long m, long long o)
+; {
+;   int i, j;
+;   for (i=0; i<n; ++i) {
+;     for (j=0; j<m; j++) {
+;       int *arr = ar + i * m * o + j * o;
+;       memset(arr, 0, o * sizeof(int));
+;     }
+;   }
+; }
+define dso_local void @NestedFor64(i32* %ar, i64 %n, i64 %m, i64 %o) #0 {
+; CHECK-LABEL: @NestedFor64(
+; CHECK-NEXT:  entry:
+; CHECK-NEXT:    [[CMP3:%.*]] = icmp slt i64 0, [[N:%.*]]
+; CHECK-NEXT:    [[CMP41:%.*]] = icmp slt i64 0, [[M:%.*]]
+; CHECK-NEXT:    [[MUL13:%.*]] = mul nsw i64 [[O:%.*]], 4
+; CHECK-NEXT:    [[CONV14:%.*]] = trunc i64 [[MUL13]] to i32
+; CHECK-NEXT:    [[OR_COND:%.*]] = select i1 [[CMP3]], i1 [[CMP41]], i1 false
+; CHECK-NEXT:    br i1 [[OR_COND]], label [[FOR_BODY_US_PREHEADER:%.*]], label [[FOR_END17:%.*]]
+; CHECK:       for.body.us.preheader:
+; CHECK-NEXT:    [[TMP0:%.*]] = mul i64 [[O]], [[M]]
+; CHECK-NEXT:    [[TMP1:%.*]] = trunc i64 [[TMP0]] to i32
+; CHECK-NEXT:    [[TMP2:%.*]] = trunc i64 [[O]] to i32
+; CHECK-NEXT:    [[TMP3:%.*]] = trunc i64 [[M]] to i32
+; CHECK-NEXT:    [[TMP4:%.*]] = mul i32 [[TMP2]], [[TMP3]]
+; CHECK-NEXT:    [[TMP5:%.*]] = shl i32 [[TMP4]], 2
+; CHECK-NEXT:    br label [[FOR_BODY_US:%.*]]
+; CHECK:       for.body.us:
+; CHECK-NEXT:    [[I_04_US:%.*]] = phi i32 [ [[INC16_US:%.*]], [[FOR_BODY_US]] ], [ 0, [[FOR_BODY_US_PREHEADER]] ]
+; CHECK-NEXT:    [[TMP6:%.*]] = mul i32 [[TMP1]], [[I_04_US]]
+; CHECK-NEXT:    [[SCEVGEP:%.*]] = getelementptr i32, i32* [[AR:%.*]], i32 [[TMP6]]
+; CHECK-NEXT:    [[SCEVGEP1:%.*]] = bitcast i32* [[SCEVGEP]] to i8*
+; CHECK-NEXT:    [[CONV7_US:%.*]] = sext i32 [[I_04_US]] to i64
+; CHECK-NEXT:    [[MUL_US:%.*]] = mul nsw i64 [[CONV7_US]], [[M]]
+; CHECK-NEXT:    [[MUL8_US:%.*]] = mul nsw i64 [[MUL_US]], [[O]]
+; CHECK-NEXT:    [[IDX_EXT_US:%.*]] = trunc i64 [[MUL8_US]] to i32
+; CHECK-NEXT:    [[ADD_PTR_US:%.*]] = getelementptr inbounds i32, i32* [[AR]], i32 [[IDX_EXT_US]]
+; CHECK-NEXT:    call void @llvm.memset.p0i8.i32(i8* align 4 [[SCEVGEP1]], i8 0, i32 [[TMP5]], i1 false)
+; CHECK-NEXT:    [[INC16_US]] = add nuw nsw i32 [[I_04_US]], 1
+; CHECK-NEXT:    [[CONV_US:%.*]] = sext i32 [[INC16_US]] to i64
+; CHECK-NEXT:    [[CMP_US:%.*]] = icmp slt i64 [[CONV_US]], [[N]]
+; CHECK-NEXT:    br i1 [[CMP_US]], label [[FOR_BODY_US]], label [[FOR_END17]]
+; CHECK:       for.end17:
+; CHECK-NEXT:    ret void
+;
+entry:
+  %cmp3 = icmp slt i64 0, %n
+  br i1 %cmp3, label %for.body.lr.ph, label %for.end17
+
+for.body.lr.ph:                                   ; preds = %entry
+  %cmp41 = icmp slt i64 0, %m
+  %mul13 = mul nsw i64 %o, 4
+  %conv14 = trunc i64 %mul13 to i32
+  br i1 %cmp41, label %for.body.us.preheader, label %for.end17
+
+for.body.us.preheader:                            ; preds = %for.body.lr.ph
+  br label %for.body.us
+
+for.body.us:                                      ; preds = %for.body.us.preheader, %for.cond2.for.end_crit_edge.us
+  %i.04.us = phi i32 [ %inc16.us, %for.cond2.for.end_crit_edge.us ], [ 0, %for.body.us.preheader ]
+  %conv7.us = sext i32 %i.04.us to i64
+  %mul.us = mul nsw i64 %conv7.us, %m
+  %mul8.us = mul nsw i64 %mul.us, %o
+  %idx.ext.us = trunc i64 %mul8.us to i32
+  %add.ptr.us = getelementptr inbounds i32, i32* %ar, i32 %idx.ext.us
+  br label %for.body6.us
+
+for.body6.us:                                     ; preds = %for.body.us, %for.body6.us
+  %j.02.us = phi i32 [ 0, %for.body.us ], [ %inc.us, %for.body6.us ]
+  %conv9.us = sext i32 %j.02.us to i64
+  %mul10.us = mul nsw i64 %conv9.us, %o
+  %idx.ext11.us = trunc i64 %mul10.us to i32
+  %add.ptr12.us = getelementptr inbounds i32, i32* %add.ptr.us, i32 %idx.ext11.us
+  %0 = bitcast i32* %add.ptr12.us to i8*
+  call void @llvm.memset.p0i8.i32(i8* align 4 %0, i8 0, i32 %conv14, i1 false)
+  %inc.us = add nuw nsw i32 %j.02.us, 1
+  %conv3.us = sext i32 %inc.us to i64
+  %cmp4.us = icmp slt i64 %conv3.us, %m
+  br i1 %cmp4.us, label %for.body6.us, label %for.cond2.for.end_crit_edge.us
+
+for.cond2.for.end_crit_edge.us:                   ; preds = %for.body6.us
+  %inc16.us = add nuw nsw i32 %i.04.us, 1
+  %conv.us = sext i32 %inc16.us to i64
+  %cmp.us = icmp slt i64 %conv.us, %n
+  br i1 %cmp.us, label %for.body.us, label %for.end17.loopexit
+
+for.end17.loopexit:                               ; preds = %for.cond2.for.end_crit_edge.us
+  br label %for.end17
+
+for.end17:                                        ; preds = %for.end17.loopexit, %for.body.lr.ph, %entry
+  ret void
+}
+
+declare void @llvm.memset.p0i8.i32(i8* nocapture writeonly, i8, i32, i1 immarg)

diff  --git a/llvm/test/Transforms/LoopIdiom/memset-runtime-64bit.ll b/llvm/test/Transforms/LoopIdiom/memset-runtime-64bit.ll
new file mode 100644
index 0000000000000..f6d2441a85cc1
--- /dev/null
+++ b/llvm/test/Transforms/LoopIdiom/memset-runtime-64bit.ll
@@ -0,0 +1,360 @@
+; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
+; RUN: opt -passes="function(loop(indvars,loop-idiom,loop-deletion),simplifycfg)" -S < %s | FileCheck %s
+; Compile command:
+; $ clang -m64 -fno-discard-value-names -O0 -S -emit-llvm -Xclang -disable-O0-optnone Code.c
+; $ bin/opt -S -basic-aa -mem2reg -loop-simplify -lcssa -loop-rotate \
+;   -licm -simple-loop-unswitch -enable-nontrivial-unswitch -loop-simplify \
+;   -loop-deletion -simplifycfg -indvars Code.ll > CodeOpt.ll
+target datalayout = "e-m:e-p270:32:32-p271:32:32-p272:64:64-i64:64-f80:128-n8:16:32:64-S128"
+; void PositiveFor64(int *ar, long long n, long long m)
+; {
+;   long long i;
+;   for (i=0; i<n; ++i) {
+;     int *arr = ar + i * m;
+;     memset(arr, 0, m * sizeof(int));
+;   }
+; }
+define dso_local void @PositiveFor64(i32* %ar, i64 %n, i64 %m) {
+; CHECK-LABEL: @PositiveFor64(
+; CHECK-NEXT:  entry:
+; CHECK-NEXT:    [[AR1:%.*]] = bitcast i32* [[AR:%.*]] to i8*
+; CHECK-NEXT:    [[CMP1:%.*]] = icmp slt i64 0, [[N:%.*]]
+; CHECK-NEXT:    br i1 [[CMP1]], label [[FOR_BODY_LR_PH:%.*]], label [[FOR_END:%.*]]
+; CHECK:       for.body.lr.ph:
+; CHECK-NEXT:    [[MUL1:%.*]] = mul i64 [[M:%.*]], 4
+; CHECK-NEXT:    [[TMP0:%.*]] = mul i64 [[M]], [[N]]
+; CHECK-NEXT:    [[TMP1:%.*]] = shl i64 [[TMP0]], 2
+; CHECK-NEXT:    call void @llvm.memset.p0i8.i64(i8* align 4 [[AR1]], i8 0, i64 [[TMP1]], i1 false)
+; CHECK-NEXT:    br label [[FOR_END]]
+; CHECK:       for.end:
+; CHECK-NEXT:    ret void
+;
+entry:
+  %cmp1 = icmp slt i64 0, %n
+  br i1 %cmp1, label %for.body.lr.ph, label %for.end
+
+for.body.lr.ph:                                   ; preds = %entry
+  %mul1 = mul i64 %m, 4
+  br label %for.body
+
+for.body:                                         ; preds = %for.body.lr.ph, %for.body
+  %i.02 = phi i64 [ 0, %for.body.lr.ph ], [ %inc, %for.body ]
+  %mul = mul nsw i64 %i.02, %m
+  %add.ptr = getelementptr inbounds i32, i32* %ar, i64 %mul
+  %0 = bitcast i32* %add.ptr to i8*
+  call void @llvm.memset.p0i8.i64(i8* align 4 %0, i8 0, i64 %mul1, i1 false)
+  %inc = add nsw i64 %i.02, 1
+  %cmp = icmp slt i64 %inc, %n
+  br i1 %cmp, label %for.body, label %for.end
+
+for.end:                                          ; preds = %for.body, %entry
+  ret void
+}
+; void NegativeFor64(int *ar, long long n, long long m)
+; {
+;   long long i;
+;   for (i=n-1; i>=0; --i) {
+;     int *arr = ar + i * m;
+;     memset(arr, 0, m * sizeof(int));
+;   }
+; }
+define dso_local void @NegativeFor64(i32* %ar, i64 %n, i64 %m) {
+; CHECK-LABEL: @NegativeFor64(
+; CHECK-NEXT:  entry:
+; CHECK-NEXT:    [[AR1:%.*]] = bitcast i32* [[AR:%.*]] to i8*
+; CHECK-NEXT:    [[SUB:%.*]] = sub nsw i64 [[N:%.*]], 1
+; CHECK-NEXT:    [[CMP1:%.*]] = icmp sge i64 [[SUB]], 0
+; CHECK-NEXT:    br i1 [[CMP1]], label [[FOR_BODY_LR_PH:%.*]], label [[FOR_END:%.*]]
+; CHECK:       for.body.lr.ph:
+; CHECK-NEXT:    [[MUL1:%.*]] = mul i64 [[M:%.*]], 4
+; CHECK-NEXT:    [[TMP0:%.*]] = mul i64 [[M]], [[N]]
+; CHECK-NEXT:    [[TMP1:%.*]] = shl i64 [[TMP0]], 2
+; CHECK-NEXT:    call void @llvm.memset.p0i8.i64(i8* align 4 [[AR1]], i8 0, i64 [[TMP1]], i1 false)
+; CHECK-NEXT:    br label [[FOR_END]]
+; CHECK:       for.end:
+; CHECK-NEXT:    ret void
+;
+entry:
+  %sub = sub nsw i64 %n, 1
+  %cmp1 = icmp sge i64 %sub, 0
+  br i1 %cmp1, label %for.body.lr.ph, label %for.end
+
+for.body.lr.ph:                                   ; preds = %entry
+  %mul1 = mul i64 %m, 4
+  br label %for.body
+
+for.body:                                         ; preds = %for.body.lr.ph, %for.body
+  %i.02 = phi i64 [ %sub, %for.body.lr.ph ], [ %dec, %for.body ]
+  %mul = mul nsw i64 %i.02, %m
+  %add.ptr = getelementptr inbounds i32, i32* %ar, i64 %mul
+  %0 = bitcast i32* %add.ptr to i8*
+  call void @llvm.memset.p0i8.i64(i8* align 4 %0, i8 0, i64 %mul1, i1 false)
+  %dec = add nsw i64 %i.02, -1
+  %cmp = icmp sge i64 %dec, 0
+  br i1 %cmp, label %for.body, label %for.end
+
+for.end:                                          ; preds = %for.body, %entry
+  ret void
+}
+; void NestedFor64(int *ar, long long n, long long m, long long o)
+; {
+;   long long i, j;
+;   for (i=0; i<n; ++i) {
+;     for (j=0; j<m; j++) {
+;       int *arr = ar + i * m * o + j * o;
+;       memset(arr, 0, o * sizeof(int));
+;     }
+;   }
+; }
+define void @NestedFor64(i32* %ar, i64 %n, i64 %m, i64 %o) {
+; CHECK-LABEL: @NestedFor64(
+; CHECK-NEXT:  entry:
+; CHECK-NEXT:    [[AR2:%.*]] = bitcast i32* [[AR:%.*]] to i8*
+; CHECK-NEXT:    [[CMP3:%.*]] = icmp slt i64 0, [[N:%.*]]
+; CHECK-NEXT:    [[CMP21:%.*]] = icmp slt i64 0, [[M:%.*]]
+; CHECK-NEXT:    [[MUL7:%.*]] = mul i64 [[O:%.*]], 4
+; CHECK-NEXT:    [[OR_COND:%.*]] = select i1 [[CMP3]], i1 [[CMP21]], i1 false
+; CHECK-NEXT:    br i1 [[OR_COND]], label [[FOR_BODY_US_PREHEADER:%.*]], label [[FOR_END10:%.*]]
+; CHECK:       for.body.us.preheader:
+; CHECK-NEXT:    [[TMP0:%.*]] = mul i64 [[O]], [[M]]
+; CHECK-NEXT:    [[TMP1:%.*]] = shl i64 [[TMP0]], 2
+; CHECK-NEXT:    [[TMP2:%.*]] = mul i64 [[TMP0]], [[N]]
+; CHECK-NEXT:    [[TMP3:%.*]] = shl i64 [[TMP2]], 2
+; CHECK-NEXT:    call void @llvm.memset.p0i8.i64(i8* align 4 [[AR2]], i8 0, i64 [[TMP3]], i1 false)
+; CHECK-NEXT:    br label [[FOR_END10]]
+; CHECK:       for.end10:
+; CHECK-NEXT:    ret void
+;
+entry:
+  %cmp3 = icmp slt i64 0, %n
+  br i1 %cmp3, label %for.body.lr.ph, label %for.end10
+
+for.body.lr.ph:                                   ; preds = %entry
+  %cmp21 = icmp slt i64 0, %m
+  %mul7 = mul i64 %o, 4
+  br i1 %cmp21, label %for.body.us.preheader, label %for.end10
+
+for.body.us.preheader:                            ; preds = %for.body.lr.ph
+  br label %for.body.us
+
+for.body.us:                                      ; preds = %for.body.us.preheader, %for.cond1.for.end_crit_edge.us
+  %i.04.us = phi i64 [ %inc9.us, %for.cond1.for.end_crit_edge.us ], [ 0, %for.body.us.preheader ]
+  %mul.us = mul nsw i64 %i.04.us, %m
+  %mul4.us = mul nsw i64 %mul.us, %o
+  %add.ptr.us = getelementptr inbounds i32, i32* %ar, i64 %mul4.us
+  br label %for.body3.us
+
+for.body3.us:                                     ; preds = %for.body.us, %for.body3.us
+  %j.02.us = phi i64 [ 0, %for.body.us ], [ %inc.us, %for.body3.us ]
+  %mul5.us = mul nsw i64 %j.02.us, %o
+  %add.ptr6.us = getelementptr inbounds i32, i32* %add.ptr.us, i64 %mul5.us
+  %0 = bitcast i32* %add.ptr6.us to i8*
+  call void @llvm.memset.p0i8.i64(i8* align 4 %0, i8 0, i64 %mul7, i1 false)
+  %inc.us = add nuw nsw i64 %j.02.us, 1
+  %exitcond = icmp ne i64 %inc.us, %m
+  br i1 %exitcond, label %for.body3.us, label %for.cond1.for.end_crit_edge.us
+
+for.cond1.for.end_crit_edge.us:                   ; preds = %for.body3.us
+  %inc9.us = add nuw nsw i64 %i.04.us, 1
+  %exitcond5 = icmp ne i64 %inc9.us, %n
+  br i1 %exitcond5, label %for.body.us, label %for.end10.loopexit
+
+for.end10.loopexit:                               ; preds = %for.cond1.for.end_crit_edge.us
+  br label %for.end10
+
+for.end10:                                        ; preds = %for.end10.loopexit, %for.body.lr.ph, %entry
+  ret void
+}
+; void PositiveFor32(int *ar, int n, int m)
+; {
+;   int i;
+;   for (i=0; i<n; ++i) {
+;     int *arr = ar + i * m;
+;     memset(arr, 0, m * sizeof(int));
+;   }
+; }
+define void @PositiveFor32(i32* %ar, i32 %n, i32 %m) {
+; CHECK-LABEL: @PositiveFor32(
+; CHECK-NEXT:  entry:
+; CHECK-NEXT:    [[AR1:%.*]] = bitcast i32* [[AR:%.*]] to i8*
+; CHECK-NEXT:    [[CONV:%.*]] = sext i32 [[N:%.*]] to i64
+; CHECK-NEXT:    [[CMP1:%.*]] = icmp slt i64 0, [[CONV]]
+; CHECK-NEXT:    br i1 [[CMP1]], label [[FOR_BODY_LR_PH:%.*]], label [[FOR_END:%.*]]
+; CHECK:       for.body.lr.ph:
+; CHECK-NEXT:    [[CONV1:%.*]] = sext i32 [[M:%.*]] to i64
+; CHECK-NEXT:    [[CONV2:%.*]] = sext i32 [[M]] to i64
+; CHECK-NEXT:    [[MUL3:%.*]] = mul i64 [[CONV2]], 4
+; CHECK-NEXT:    [[TMP0:%.*]] = mul i64 [[CONV1]], [[CONV]]
+; CHECK-NEXT:    [[TMP1:%.*]] = shl i64 [[TMP0]], 2
+; CHECK-NEXT:    call void @llvm.memset.p0i8.i64(i8* align 4 [[AR1]], i8 0, i64 [[TMP1]], i1 false)
+; CHECK-NEXT:    br label [[FOR_END]]
+; CHECK:       for.end:
+; CHECK-NEXT:    ret void
+;
+entry:
+  %conv = sext i32 %n to i64
+  %cmp1 = icmp slt i64 0, %conv
+  br i1 %cmp1, label %for.body.lr.ph, label %for.end
+
+for.body.lr.ph:                                   ; preds = %entry
+  %conv1 = sext i32 %m to i64
+  %conv2 = sext i32 %m to i64
+  %mul3 = mul i64 %conv2, 4
+  br label %for.body
+
+for.body:                                         ; preds = %for.body.lr.ph, %for.body
+  %i.02 = phi i64 [ 0, %for.body.lr.ph ], [ %inc, %for.body ]
+  %mul = mul nsw i64 %i.02, %conv1
+  %add.ptr = getelementptr inbounds i32, i32* %ar, i64 %mul
+  %0 = bitcast i32* %add.ptr to i8*
+  call void @llvm.memset.p0i8.i64(i8* align 4 %0, i8 0, i64 %mul3, i1 false)
+  %inc = add nsw i64 %i.02, 1
+  %cmp = icmp slt i64 %inc, %conv
+  br i1 %cmp, label %for.body, label %for.end
+
+for.end:                                          ; preds = %for.body, %entry
+  ret void
+}
+; void Negative32(int *ar, int n, int m)
+; {
+;   long long i;
+;   for (i=n-1; i>=0; i--) {
+;     int *arr = ar + i * m;
+;     memset(arr, 0, m * sizeof(int));
+;   }
+; }
+define void @Negative32(i32* %ar, i32 %n, i32 %m) {
+; CHECK-LABEL: @Negative32(
+; CHECK-NEXT:  entry:
+; CHECK-NEXT:    [[AR1:%.*]] = bitcast i32* [[AR:%.*]] to i8*
+; CHECK-NEXT:    [[SUB:%.*]] = sub nsw i32 [[N:%.*]], 1
+; CHECK-NEXT:    [[CONV:%.*]] = sext i32 [[SUB]] to i64
+; CHECK-NEXT:    [[CMP1:%.*]] = icmp sge i64 [[CONV]], 0
+; CHECK-NEXT:    br i1 [[CMP1]], label [[FOR_BODY_LR_PH:%.*]], label [[FOR_END:%.*]]
+; CHECK:       for.body.lr.ph:
+; CHECK-NEXT:    [[CONV1:%.*]] = sext i32 [[M:%.*]] to i64
+; CHECK-NEXT:    [[CONV2:%.*]] = sext i32 [[M]] to i64
+; CHECK-NEXT:    [[MUL3:%.*]] = mul i64 [[CONV2]], 4
+; CHECK-NEXT:    [[TMP0:%.*]] = sub i64 [[CONV]], -1
+; CHECK-NEXT:    [[TMP1:%.*]] = mul i64 [[CONV1]], [[TMP0]]
+; CHECK-NEXT:    [[TMP2:%.*]] = shl i64 [[TMP1]], 2
+; CHECK-NEXT:    call void @llvm.memset.p0i8.i64(i8* align 4 [[AR1]], i8 0, i64 [[TMP2]], i1 false)
+; CHECK-NEXT:    br label [[FOR_END]]
+; CHECK:       for.end:
+; CHECK-NEXT:    ret void
+;
+entry:
+  %sub = sub nsw i32 %n, 1
+  %conv = sext i32 %sub to i64
+  %cmp1 = icmp sge i64 %conv, 0
+  br i1 %cmp1, label %for.body.lr.ph, label %for.end
+
+for.body.lr.ph:                                   ; preds = %entry
+  %conv1 = sext i32 %m to i64
+  %conv2 = sext i32 %m to i64
+  %mul3 = mul i64 %conv2, 4
+  br label %for.body
+
+for.body:                                         ; preds = %for.body.lr.ph, %for.body
+  %i.02 = phi i64 [ %conv, %for.body.lr.ph ], [ %dec, %for.body ]
+  %mul = mul nsw i64 %i.02, %conv1
+  %add.ptr = getelementptr inbounds i32, i32* %ar, i64 %mul
+  %0 = bitcast i32* %add.ptr to i8*
+  call void @llvm.memset.p0i8.i64(i8* align 4 %0, i8 0, i64 %mul3, i1 false)
+  %dec = add nsw i64 %i.02, -1
+  %cmp = icmp sge i64 %dec, 0
+  br i1 %cmp, label %for.body, label %for.end
+
+for.end:                                          ; preds = %for.body, %entry
+  ret void
+}
+; void NestedFor32(int *ar, int n, int m, int o)
+; {
+;   int i, j;
+;   for (i=0; i<n; ++i) {
+;     for (j=0; j<m; j++) {
+;       int *arr = ar + i * m * o + j * o;
+;       memset(arr, 0, o * sizeof(int));
+;     }
+;   }
+; }
+define void @NestedFor32(i32* %ar, i32 %n, i32 %m, i32 %o) {
+; CHECK-LABEL: @NestedFor32(
+; CHECK-NEXT:  entry:
+; CHECK-NEXT:    [[CMP3:%.*]] = icmp slt i32 0, [[N:%.*]]
+; CHECK-NEXT:    br i1 [[CMP3]], label [[FOR_BODY_LR_PH:%.*]], label [[FOR_END11:%.*]]
+; CHECK:       for.body.lr.ph:
+; CHECK-NEXT:    [[CMP21:%.*]] = icmp slt i32 0, [[M:%.*]]
+; CHECK-NEXT:    [[CONV:%.*]] = sext i32 [[O:%.*]] to i64
+; CHECK-NEXT:    [[MUL8:%.*]] = mul i64 [[CONV]], 4
+; CHECK-NEXT:    br i1 [[CMP21]], label [[FOR_BODY_US_PREHEADER:%.*]], label [[FOR_END11]]
+; CHECK:       for.body.us.preheader:
+; CHECK-NEXT:    [[TMP0:%.*]] = sext i32 [[O]] to i64
+; CHECK-NEXT:    [[TMP1:%.*]] = sext i32 [[M]] to i64
+; CHECK-NEXT:    [[WIDE_TRIP_COUNT10:%.*]] = zext i32 [[N]] to i64
+; CHECK-NEXT:    [[TMP2:%.*]] = mul i64 [[TMP0]], [[TMP1]]
+; CHECK-NEXT:    [[TMP3:%.*]] = zext i32 [[M]] to i64
+; CHECK-NEXT:    [[TMP4:%.*]] = mul i64 [[TMP0]], [[TMP3]]
+; CHECK-NEXT:    [[TMP5:%.*]] = shl i64 [[TMP4]], 2
+; CHECK-NEXT:    br label [[FOR_BODY_US:%.*]]
+; CHECK:       for.body.us:
+; CHECK-NEXT:    [[INDVARS_IV6:%.*]] = phi i64 [ 0, [[FOR_BODY_US_PREHEADER]] ], [ [[INDVARS_IV_NEXT7:%.*]], [[FOR_BODY_US]] ]
+; CHECK-NEXT:    [[TMP6:%.*]] = mul i64 [[TMP2]], [[INDVARS_IV6]]
+; CHECK-NEXT:    [[SCEVGEP:%.*]] = getelementptr i32, i32* [[AR:%.*]], i64 [[TMP6]]
+; CHECK-NEXT:    [[SCEVGEP1:%.*]] = bitcast i32* [[SCEVGEP]] to i8*
+; CHECK-NEXT:    [[WIDE_TRIP_COUNT:%.*]] = zext i32 [[M]] to i64
+; CHECK-NEXT:    call void @llvm.memset.p0i8.i64(i8* align 4 [[SCEVGEP1]], i8 0, i64 [[TMP5]], i1 false)
+; CHECK-NEXT:    [[INDVARS_IV_NEXT7]] = add nuw nsw i64 [[INDVARS_IV6]], 1
+; CHECK-NEXT:    [[EXITCOND11:%.*]] = icmp ne i64 [[INDVARS_IV_NEXT7]], [[WIDE_TRIP_COUNT10]]
+; CHECK-NEXT:    br i1 [[EXITCOND11]], label [[FOR_BODY_US]], label [[FOR_END11]]
+; CHECK:       for.end11:
+; CHECK-NEXT:    ret void
+;
+entry:
+  %cmp3 = icmp slt i32 0, %n
+  br i1 %cmp3, label %for.body.lr.ph, label %for.end11
+
+for.body.lr.ph:                                   ; preds = %entry
+  %cmp21 = icmp slt i32 0, %m
+  %conv = sext i32 %o to i64
+  %mul8 = mul i64 %conv, 4
+  br i1 %cmp21, label %for.body.us.preheader, label %for.end11
+
+for.body.us.preheader:                            ; preds = %for.body.lr.ph
+  %0 = sext i32 %o to i64
+  %1 = sext i32 %m to i64
+  %2 = sext i32 %o to i64
+  %wide.trip.count10 = zext i32 %n to i64
+  br label %for.body.us
+
+for.body.us:                                      ; preds = %for.body.us.preheader, %for.cond1.for.end_crit_edge.us
+  %indvars.iv6 = phi i64 [ 0, %for.body.us.preheader ], [ %indvars.iv.next7, %for.cond1.for.end_crit_edge.us ]
+  %3 = mul nsw i64 %indvars.iv6, %1
+  %4 = mul nsw i64 %3, %2
+  %add.ptr.us = getelementptr inbounds i32, i32* %ar, i64 %4
+  %wide.trip.count = zext i32 %m to i64
+  br label %for.body3.us
+
+for.body3.us:                                     ; preds = %for.body.us, %for.body3.us
+  %indvars.iv = phi i64 [ 0, %for.body.us ], [ %indvars.iv.next, %for.body3.us ]
+  %5 = mul nsw i64 %indvars.iv, %0
+  %add.ptr7.us = getelementptr inbounds i32, i32* %add.ptr.us, i64 %5
+  %6 = bitcast i32* %add.ptr7.us to i8*
+  call void @llvm.memset.p0i8.i64(i8* align 4 %6, i8 0, i64 %mul8, i1 false)
+  %indvars.iv.next = add nuw nsw i64 %indvars.iv, 1
+  %exitcond = icmp ne i64 %indvars.iv.next, %wide.trip.count
+  br i1 %exitcond, label %for.body3.us, label %for.cond1.for.end_crit_edge.us
+
+for.cond1.for.end_crit_edge.us:                   ; preds = %for.body3.us
+  %indvars.iv.next7 = add nuw nsw i64 %indvars.iv6, 1
+  %exitcond11 = icmp ne i64 %indvars.iv.next7, %wide.trip.count10
+  br i1 %exitcond11, label %for.body.us, label %for.end11.loopexit
+
+for.end11.loopexit:                               ; preds = %for.cond1.for.end_crit_edge.us
+  br label %for.end11
+
+for.end11:                                        ; preds = %for.end11.loopexit, %for.body.lr.ph, %entry
+  ret void
+}
+
+declare void @llvm.memset.p0i8.i64(i8* nocapture writeonly, i8, i64, i1 immarg)

diff  --git a/llvm/test/Transforms/LoopIdiom/memset-runtime.ll b/llvm/test/Transforms/LoopIdiom/memset-runtime.ll
deleted file mode 100644
index c5eba6cea7d22..0000000000000
--- a/llvm/test/Transforms/LoopIdiom/memset-runtime.ll
+++ /dev/null
@@ -1,110 +0,0 @@
-; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
-; RUN: opt -passes="function(loop(loop-idiom,loop-deletion),simplifycfg)" -S < %s | FileCheck %s
-; The C code to generate this testcase:
-; void test(int n, int m, int *ar)
-; {
-;	  long i;
-;	  for (i=0; i<n; ++i) {
-;     int *arr = ar + i * m; // ar[i];
-;     memset(arr, 0, m * sizeof(int));
-;   }
-; }
-; The optimized IR should be similar to the following:
-; void test(int n, int m, int *ar)
-; {
-;   memset(ar, 0, m * n * sizeof(int));
-; }
-define void @For_PositiveStride(i32* nocapture %ar, i64 %n, i64 %m) {
-; CHECK-LABEL: @For_PositiveStride(
-; CHECK-NEXT:  entry:
-; CHECK-NEXT:    [[AR1:%.*]] = bitcast i32* [[AR:%.*]] to i8*
-; CHECK-NEXT:    [[TMP0:%.*]] = shl nuw i64 [[M:%.*]], 2
-; CHECK-NEXT:    [[TMP1:%.*]] = mul i64 [[M]], [[N:%.*]]
-; CHECK-NEXT:    [[TMP2:%.*]] = shl i64 [[TMP1]], 2
-; CHECK-NEXT:    call void @llvm.memset.p0i8.i64(i8* align 4 [[AR1]], i8 0, i64 [[TMP2]], i1 false)
-; CHECK-NEXT:    ret void
-;
-entry:
-  %0 = shl nuw i64 %m, 2
-  br label %for.cond1.preheader
-
-for.cond1.preheader:                              ; preds = %for.inc4, %entry
-  %i.017 = phi i64 [ 0, %entry ], [ %inc5, %for.inc4 ]
-  %1 = mul i64 %m, %i.017
-  %scevgep = getelementptr i32, i32* %ar, i64 %1
-  %scevgep1 = bitcast i32* %scevgep to i8*
-  %mul = mul nsw i64 %i.017, %m
-  call void @llvm.memset.p0i8.i64(i8* align 4 %scevgep1, i8 0, i64 %0, i1 false)
-  br label %for.inc4
-
-for.inc4:                                         ; preds = %for.cond1.preheader
-  %inc5 = add nuw nsw i64 %i.017, 1
-  %exitcond18.not = icmp eq i64 %inc5, %n
-  br i1 %exitcond18.not, label %for.end6, label %for.cond1.preheader
-
-for.end6:                                         ; preds = %for.inc4
-  ret void
-}
-
-; The C code to generate this testcase:
-; void test(int n, int m, int *ar)
-; {
-;	  long i;
-;	  for (i=n-1; i>=0; i--) {
-;     int *arr = ar + i * m;
-;     memset(arr, 0, m * sizeof(int));
-;   }
-; }
-define void @For_NegativeStride(i32* %ar, i32 %n, i32 %m) {
-; CHECK-LABEL: @For_NegativeStride(
-; CHECK-NEXT:  entry:
-; CHECK-NEXT:    [[AR1:%.*]] = bitcast i32* [[AR:%.*]] to i8*
-; CHECK-NEXT:    [[SUB:%.*]] = sub nsw i32 [[N:%.*]], 1
-; CHECK-NEXT:    [[CONV:%.*]] = sext i32 [[SUB]] to i64
-; CHECK-NEXT:    [[CMP1:%.*]] = icmp sge i64 [[CONV]], 0
-; CHECK-NEXT:    br i1 [[CMP1]], label [[FOR_BODY_LR_PH:%.*]], label [[FOR_END:%.*]]
-; CHECK:       for.body.lr.ph:
-; CHECK-NEXT:    [[CONV1:%.*]] = sext i32 [[M:%.*]] to i64
-; CHECK-NEXT:    [[CONV2:%.*]] = sext i32 [[M]] to i64
-; CHECK-NEXT:    [[MUL3:%.*]] = mul i64 [[CONV2]], 4
-; CHECK-NEXT:    [[TMP0:%.*]] = sub i64 [[CONV]], -1
-; CHECK-NEXT:    [[TMP1:%.*]] = mul i64 [[CONV1]], [[TMP0]]
-; CHECK-NEXT:    [[TMP2:%.*]] = shl i64 [[TMP1]], 2
-; CHECK-NEXT:    call void @llvm.memset.p0i8.i64(i8* align 4 [[AR1]], i8 0, i64 [[TMP2]], i1 false)
-; CHECK-NEXT:    br label [[FOR_END]]
-; CHECK:       for.end:
-; CHECK-NEXT:    ret void
-;
-entry:
-  %sub = sub nsw i32 %n, 1
-  %conv = sext i32 %sub to i64
-  %cmp1 = icmp sge i64 %conv, 0
-  br i1 %cmp1, label %for.body.lr.ph, label %for.end
-
-for.body.lr.ph:                                   ; preds = %entry
-  %conv1 = sext i32 %m to i64
-  %conv2 = sext i32 %m to i64
-  %mul3 = mul i64 %conv2, 4
-  br label %for.body
-
-for.body:                                         ; preds = %for.body.lr.ph, %for.inc
-  %i.02 = phi i64 [ %conv, %for.body.lr.ph ], [ %dec, %for.inc ]
-  %mul = mul nsw i64 %i.02, %conv1
-  %add.ptr = getelementptr inbounds i32, i32* %ar, i64 %mul
-  %0 = bitcast i32* %add.ptr to i8*
-  call void @llvm.memset.p0i8.i64(i8* align 4 %0, i8 0, i64 %mul3, i1 false)
-  br label %for.inc
-
-for.inc:                                          ; preds = %for.body
-  %dec = add nsw i64 %i.02, -1
-  %cmp = icmp sge i64 %dec, 0
-  br i1 %cmp, label %for.body, label %for.cond.for.end_crit_edge
-
-for.cond.for.end_crit_edge:                       ; preds = %for.inc
-  br label %for.end
-
-for.end:                                          ; preds = %for.cond.for.end_crit_edge, %entry
-  ret void
-}
-
-declare void @llvm.memset.p0i8.i64(i8* nocapture writeonly, i8, i64, i1 immarg)


        


More information about the llvm-commits mailing list