[clang] 0b9922e - [CodeGen] Emit IR for fixed-point multiplication and division.

Bevin Hansson via cfe-commits cfe-commits at lists.llvm.org
Wed Apr 8 05:42:13 PDT 2020


Author: Bevin Hansson
Date: 2020-04-08T14:33:04+02:00
New Revision: 0b9922e67a0b5520d76c293e9aef13a7ad4f3a8d

URL: https://github.com/llvm/llvm-project/commit/0b9922e67a0b5520d76c293e9aef13a7ad4f3a8d
DIFF: https://github.com/llvm/llvm-project/commit/0b9922e67a0b5520d76c293e9aef13a7ad4f3a8d.diff

LOG: [CodeGen] Emit IR for fixed-point multiplication and division.

Reviewers: rjmccall, leonardchan

Subscribers: cfe-commits

Tags: #clang

Differential Revision: https://reviews.llvm.org/D73182

Added: 
    clang/test/Frontend/fixed_point_div.c
    clang/test/Frontend/fixed_point_mul.c

Modified: 
    clang/lib/CodeGen/CGExprScalar.cpp

Removed: 
    


################################################################################
diff  --git a/clang/lib/CodeGen/CGExprScalar.cpp b/clang/lib/CodeGen/CGExprScalar.cpp
index a0d1fe24da92..63eb4b0fe932 100644
--- a/clang/lib/CodeGen/CGExprScalar.cpp
+++ b/clang/lib/CodeGen/CGExprScalar.cpp
@@ -746,6 +746,8 @@ class ScalarExprEmitter
       Value *V = Builder.CreateFMul(Ops.LHS, Ops.RHS, "mul");
       return propagateFMFlags(V, Ops);
     }
+    if (Ops.isFixedPointBinOp())
+      return EmitFixedPointBinOp(Ops);
     return Builder.CreateMul(Ops.LHS, Ops.RHS, "mul");
   }
   /// Create a binary op that checks for overflow.
@@ -3121,6 +3123,8 @@ Value *ScalarExprEmitter::EmitDiv(const BinOpInfo &Ops) {
     }
     return Val;
   }
+  else if (Ops.isFixedPointBinOp())
+    return EmitFixedPointBinOp(Ops);
   else if (Ops.Ty->hasUnsignedIntegerRepresentation())
     return Builder.CreateUDiv(Ops.LHS, Ops.RHS, "div");
   else
@@ -3543,6 +3547,32 @@ Value *ScalarExprEmitter::EmitFixedPointBinOp(const BinOpInfo &op) {
     }
     break;
   }
+  case BO_Mul: {
+    llvm::Intrinsic::ID IID;
+    if (ResultFixedSema.isSaturated())
+      IID = ResultFixedSema.isSigned()
+                ? llvm::Intrinsic::smul_fix_sat
+                : llvm::Intrinsic::umul_fix_sat;
+    else
+      IID = ResultFixedSema.isSigned()
+                ? llvm::Intrinsic::smul_fix
+                : llvm::Intrinsic::umul_fix;
+    Result = Builder.CreateIntrinsic(IID, {FullLHS->getType()},
+        {FullLHS, FullRHS, Builder.getInt32(CommonFixedSema.getScale())});
+    break;
+  }
+  case BO_Div: {
+    llvm::Intrinsic::ID IID;
+    if (ResultFixedSema.isSaturated())
+      IID = ResultFixedSema.isSigned() ? llvm::Intrinsic::sdiv_fix_sat
+                                       : llvm::Intrinsic::udiv_fix_sat;
+    else
+      IID = ResultFixedSema.isSigned() ? llvm::Intrinsic::sdiv_fix
+                                       : llvm::Intrinsic::udiv_fix;
+    Result = Builder.CreateIntrinsic(IID, {FullLHS->getType()},
+        {FullLHS, FullRHS, Builder.getInt32(CommonFixedSema.getScale())});
+    break;    
+  }
   case BO_LT:
     return CommonFixedSema.isSigned() ? Builder.CreateICmpSLT(FullLHS, FullRHS)
                                       : Builder.CreateICmpULT(FullLHS, FullRHS);
@@ -3562,8 +3592,6 @@ Value *ScalarExprEmitter::EmitFixedPointBinOp(const BinOpInfo &op) {
     return Builder.CreateICmpEQ(FullLHS, FullRHS);
   case BO_NE:
     return Builder.CreateICmpNE(FullLHS, FullRHS);
-  case BO_Mul:
-  case BO_Div:
   case BO_Shl:
   case BO_Shr:
   case BO_Cmp:

diff  --git a/clang/test/Frontend/fixed_point_div.c b/clang/test/Frontend/fixed_point_div.c
new file mode 100644
index 000000000000..ec17e082558c
--- /dev/null
+++ b/clang/test/Frontend/fixed_point_div.c
@@ -0,0 +1,431 @@
+// RUN: %clang_cc1 -ffixed-point -triple x86_64-unknown-linux-gnu -S -emit-llvm %s -o - | FileCheck %s --check-prefixes=CHECK,SIGNED
+// RUN: %clang_cc1 -ffixed-point -triple x86_64-unknown-linux-gnu -fpadding-on-unsigned-fixed-point -S -emit-llvm %s -o - | FileCheck %s --check-prefixes=CHECK,UNSIGNED
+
+void SignedDivision() {
+  // CHECK-LABEL: SignedDivision
+  short _Accum sa;
+  _Accum a, b, c, d;
+  long _Accum la;
+  unsigned short _Accum usa;
+  unsigned _Accum ua;
+  unsigned long _Accum ula;
+
+  short _Fract sf;
+  _Fract f;
+  long _Fract lf;
+  unsigned short _Fract usf;
+  unsigned _Fract uf;
+  unsigned long _Fract ulf;
+
+  // Same type
+  // CHECK:       [[TMP0:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:  [[TMP1:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:  [[TMP2:%.*]] = call i16 @llvm.sdiv.fix.i16(i16 [[TMP0]], i16 [[TMP1]], i32 7)
+  // CHECK-NEXT:  store i16 [[TMP2]], i16* %sa, align 2
+  sa = sa / sa;
+
+  // To larger scale and larger width
+  // CHECK:       [[TMP3:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:  [[TMP4:%.*]] = load i32, i32* %a, align 4
+  // CHECK-NEXT:  [[RESIZE:%.*]] = sext i16 [[TMP3]] to i32
+  // CHECK-NEXT:  [[UPSCALE:%.*]] = shl i32 [[RESIZE]], 8
+  // CHECK-NEXT:  [[TMP5:%.*]] = call i32 @llvm.sdiv.fix.i32(i32 [[UPSCALE]], i32 [[TMP4]], i32 15)
+  // CHECK-NEXT:  store i32 [[TMP5]], i32* %a, align 4
+  a = sa / a;
+
+  // To same scale and smaller width
+  // CHECK:       [[TMP6:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:  [[TMP7:%.*]] = load i8, i8* %sf, align 1
+  // CHECK-NEXT:  [[RESIZE1:%.*]] = sext i8 [[TMP7]] to i16
+  // CHECK-NEXT:  [[TMP8:%.*]] = call i16 @llvm.sdiv.fix.i16(i16 [[TMP6]], i16 [[RESIZE1]], i32 7)
+  // CHECK-NEXT:  store i16 [[TMP8]], i16* %sa, align 2
+  sa = sa / sf;
+
+  // To smaller scale and same width.
+  // CHECK:       [[TMP9:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:  [[TMP10:%.*]] = load i16, i16* %f, align 2
+  // CHECK-NEXT:  [[RESIZE2:%.*]] = sext i16 [[TMP9]] to i24
+  // CHECK-NEXT:  [[UPSCALE3:%.*]] = shl i24 [[RESIZE2]], 8
+  // CHECK-NEXT:  [[RESIZE4:%.*]] = sext i16 [[TMP10]] to i24
+  // CHECK-NEXT:  [[TMP11:%.*]] = call i24 @llvm.sdiv.fix.i24(i24 [[UPSCALE3]], i24 [[RESIZE4]], i32 15)
+  // CHECK-NEXT:  [[DOWNSCALE:%.*]] = ashr i24 [[TMP11]], 8
+  // CHECK-NEXT:  [[RESIZE5:%.*]] = trunc i24 [[DOWNSCALE]] to i16
+  // CHECK-NEXT:  store i16 [[RESIZE5]], i16* %sa, align 2
+  sa = sa / f;
+
+  // To smaller scale and smaller width
+  // CHECK:       [[TMP12:%.*]] = load i32, i32* %a, align 4
+  // CHECK-NEXT:  [[TMP13:%.*]] = load i8, i8* %sf, align 1
+  // CHECK-NEXT:  [[RESIZE6:%.*]] = sext i8 [[TMP13]] to i32
+  // CHECK-NEXT:  [[UPSCALE7:%.*]] = shl i32 [[RESIZE6]], 8
+  // CHECK-NEXT:  [[TMP14:%.*]] = call i32 @llvm.sdiv.fix.i32(i32 [[TMP12]], i32 [[UPSCALE7]], i32 15)
+  // CHECK-NEXT:  store i32 [[TMP14]], i32* %a, align 4
+  a = a / sf;
+
+  // To larger scale and same width
+  // CHECK:       [[TMP15:%.*]] = load i32, i32* %a, align 4
+  // CHECK-NEXT:  [[TMP16:%.*]] = load i32, i32* %lf, align 4
+  // CHECK-NEXT:  [[RESIZE8:%.*]] = sext i32 [[TMP15]] to i48
+  // CHECK-NEXT:  [[UPSCALE9:%.*]] = shl i48 [[RESIZE8]], 16
+  // CHECK-NEXT:  [[RESIZE10:%.*]] = sext i32 [[TMP16]] to i48
+  // CHECK-NEXT:  [[TMP17:%.*]] = call i48 @llvm.sdiv.fix.i48(i48 [[UPSCALE9]], i48 [[RESIZE10]], i32 31)
+  // CHECK-NEXT:  [[DOWNSCALE11:%.*]] = ashr i48 [[TMP17]], 16
+  // CHECK-NEXT:  [[RESIZE12:%.*]] = trunc i48 [[DOWNSCALE11]] to i32
+  // CHECK-NEXT:  store i32 [[RESIZE12]], i32* %a, align 4
+  a = a / lf;
+
+  // With corresponding unsigned type
+  // CHECK:        [[TMP18:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:   [[TMP19:%.*]] = load i16, i16* %usa, align 2
+  // SIGNED-NEXT:  [[RESIZE13:%.*]] = sext i16 [[TMP18]] to i17
+  // SIGNED-NEXT:  [[UPSCALE14:%.*]] = shl i17 [[RESIZE13]], 1
+  // SIGNED-NEXT:  [[RESIZE15:%.*]] = zext i16 [[TMP19]] to i17
+  // SIGNED-NEXT:  [[TMP20:%.*]] = call i17 @llvm.sdiv.fix.i17(i17 [[UPSCALE14]], i17 [[RESIZE15]], i32 8)
+  // SIGNED-NEXT:  [[DOWNSCALE16:%.*]] = ashr i17 [[TMP20]], 1
+  // SIGNED-NEXT:  [[RESIZE17:%.*]] = trunc i17 [[DOWNSCALE16]] to i16
+  // SIGNED-NEXT:  store i16 [[RESIZE17]], i16* %sa, align 2
+  // UNSIGNED-NEXT:[[TMP20:%.*]] = call i16 @llvm.sdiv.fix.i16(i16 [[TMP18]], i16 [[TMP19]], i32 7)
+  // UNSIGNED-NEXT:store i16 [[TMP20]], i16* %sa, align 2
+  sa = sa / usa;
+
+  // With unsigned of larger scale
+  // CHECK:        [[TMP21:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:   [[TMP22:%.*]] = load i32, i32* %ua, align 4
+  // SIGNED-NEXT:  [[RESIZE18:%.*]] = sext i16 [[TMP21]] to i33
+  // SIGNED-NEXT:  [[UPSCALE19:%.*]] = shl i33 [[RESIZE18]], 9
+  // SIGNED-NEXT:  [[RESIZE20:%.*]] = zext i32 [[TMP22]] to i33
+  // SIGNED-NEXT:  [[TMP23:%.*]] = call i33 @llvm.sdiv.fix.i33(i33 [[UPSCALE19]], i33 [[RESIZE20]], i32 16)
+  // SIGNED-NEXT:  [[DOWNSCALE21:%.*]] = ashr i33 [[TMP23]], 1
+  // SIGNED-NEXT:  [[RESIZE22:%.*]] = trunc i33 [[DOWNSCALE21]] to i32
+  // SIGNED-NEXT:  store i32 [[RESIZE22]], i32* %a, align 4
+  // UNSIGNED-NEXT:[[RESIZE13:%.*]] = sext i16 [[TMP21]] to i32
+  // UNSIGNED-NEXT:[[UPSCALE14:%.*]] = shl i32 [[RESIZE13]], 8
+  // UNSIGNED-NEXT:[[TMP23:%.*]] = call i32 @llvm.sdiv.fix.i32(i32 [[UPSCALE14]], i32 [[TMP22]], i32 15)
+  // UNSIGNED-NEXT:store i32 [[TMP23]], i32* %a, align 4
+  a = sa / ua;
+
+  // With unsigned of smaller width
+  // CHECK:        [[TMP24:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:   [[TMP25:%.*]] = load i8, i8* %usf, align 1
+  // SIGNED-NEXT:  [[RESIZE23:%.*]] = sext i16 [[TMP24]] to i17
+  // SIGNED-NEXT:  [[UPSCALE24:%.*]] = shl i17 [[RESIZE23]], 1
+  // SIGNED-NEXT:  [[RESIZE25:%.*]] = zext i8 [[TMP25]] to i17
+  // SIGNED-NEXT:  [[TMP26:%.*]] = call i17 @llvm.sdiv.fix.i17(i17 [[UPSCALE24]], i17 [[RESIZE25]], i32 8)
+  // SIGNED-NEXT:  [[DOWNSCALE26:%.*]] = ashr i17 [[TMP26]], 1
+  // SIGNED-NEXT:  [[RESIZE27:%.*]] = trunc i17 [[DOWNSCALE26]] to i16
+  // SIGNED-NEXT:  store i16 [[RESIZE27]], i16* %sa, align 2
+  // UNSIGNED-NEXT:[[RESIZE15:%.*]] = zext i8 [[TMP25]] to i16
+  // UNSIGNED-NEXT:[[TMP26:%.*]] = call i16 @llvm.sdiv.fix.i16(i16 [[TMP24]], i16 [[RESIZE15]], i32 7)
+  // UNSIGNED-NEXT:store i16 [[TMP26]], i16* %sa, align 2
+  sa = sa / usf;
+
+  // With unsigned of larger width and smaller scale
+  // CHECK:        [[TMP27:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:   [[TMP28:%.*]] = load i32, i32* %ulf, align 4
+  // SIGNED-NEXT:  [[RESIZE28:%.*]] = sext i16 [[TMP27]] to i41
+  // SIGNED-NEXT:  [[UPSCALE29:%.*]] = shl i41 [[RESIZE28]], 25
+  // SIGNED-NEXT:  [[RESIZE30:%.*]] = zext i32 [[TMP28]] to i41
+  // SIGNED-NEXT:  [[TMP29:%.*]] = call i41 @llvm.sdiv.fix.i41(i41 [[UPSCALE29]], i41 [[RESIZE30]], i32 32)
+  // SIGNED-NEXT:  [[DOWNSCALE31:%.*]] = ashr i41 [[TMP29]], 25
+  // SIGNED-NEXT:  [[RESIZE32:%.*]] = trunc i41 [[DOWNSCALE31]] to i16
+  // SIGNED-NEXT:  store i16 [[RESIZE32]], i16* %sa, align 2
+  // UNSIGNED-NEXT:[[RESIZE16:%.*]] = sext i16 [[TMP27]] to i40
+  // UNSIGNED-NEXT:[[UPSCALE17:%.*]] = shl i40 [[RESIZE16]], 24
+  // UNSIGNED-NEXT:[[RESIZE18:%.*]] = zext i32 [[TMP28]] to i40
+  // UNSIGNED-NEXT:[[TMP29:%.*]] = call i40 @llvm.sdiv.fix.i40(i40 [[UPSCALE17]], i40 [[RESIZE18]], i32 31)
+  // UNSIGNED-NEXT:[[DOWNSCALE19:%.*]] = ashr i40 [[TMP29]], 24
+  // UNSIGNED-NEXT:[[RESIZE20:%.*]] = trunc i40 [[DOWNSCALE19]] to i16
+  // UNSIGNED-NEXT:store i16 [[RESIZE20]], i16* %sa, align 2
+  sa = sa / ulf;
+
+  // Chained divisions of the same signed type should result in the same
+  // CHECK:       [[TMP30:%.*]] = load i32, i32* %a, align 4
+  // CHECK-NEXT:  [[TMP31:%.*]] = load i32, i32* %b, align 4
+  // CHECK-NEXT:  [[TMP32:%.*]] = call i32 @llvm.sdiv.fix.i32(i32 [[TMP30]], i32 [[TMP31]], i32 15)
+  // CHECK-NEXT:  [[TMP33:%.*]] = load i32, i32* %c, align 4
+  // CHECK-NEXT:  [[TMP34:%.*]] = call i32 @llvm.sdiv.fix.i32(i32 [[TMP32]], i32 [[TMP33]], i32 15)
+  // CHECK-NEXT:  [[TMP35:%.*]] = load i32, i32* %d, align 4
+  // CHECK-NEXT:  [[TMP36:%.*]] = call i32 @llvm.sdiv.fix.i32(i32 [[TMP34]], i32 [[TMP35]], i32 15)
+  // CHECK-NEXT:  store i32 [[TMP36]], i32* %a, align 4
+  a = a / b / c / d;
+}
+
+
+void UnsignedDivision() {
+  // CHECK-LABEL: UnsignedDivision
+  unsigned short _Accum usa;
+  unsigned _Accum ua;
+  unsigned long _Accum ula;
+
+  unsigned short _Fract usf;
+  unsigned _Fract uf;
+  unsigned long _Fract ulf;
+
+  // CHECK:         [[TMP0:%.*]] = load i16, i16* %usa, align 2
+  // CHECK-NEXT:    [[TMP1:%.*]] = load i16, i16* %usa, align 2
+  // SIGNED-NEXT:   [[TMP2:%.*]] = call i16 @llvm.udiv.fix.i16(i16 [[TMP0]], i16 [[TMP1]], i32 8)
+  // UNSIGNED-NEXT: [[TMP2:%.*]] = call i16 @llvm.udiv.fix.i16(i16 [[TMP0]], i16 [[TMP1]], i32 7)
+  // CHECK-NEXT:    store i16 [[TMP2]], i16* %usa, align 2
+  usa = usa / usa;
+
+  // CHECK:         [[TMP3:%.*]] = load i16, i16* %usa, align 2
+  // CHECK-NEXT:    [[TMP4:%.*]] = load i32, i32* %ua, align 4
+  // CHECK-NEXT:    [[RESIZE:%.*]] = zext i16 [[TMP3]] to i32
+  // CHECK-NEXT:    [[UPSCALE:%.*]] = shl i32 [[RESIZE]], 8
+  // SIGNED-NEXT:   [[TMP5:%.*]] = call i32 @llvm.udiv.fix.i32(i32 [[UPSCALE]], i32 [[TMP4]], i32 16)
+  // UNSIGNED-NEXT: [[TMP5:%.*]] = call i32 @llvm.udiv.fix.i32(i32 [[UPSCALE]], i32 [[TMP4]], i32 15)
+  // CHECK-NEXT:    store i32 [[TMP5]], i32* %ua, align 4
+  ua = usa / ua;
+
+  // CHECK:         [[TMP6:%.*]] = load i16, i16* %usa, align 2
+  // CHECK-NEXT:    [[TMP7:%.*]] = load i8, i8* %usf, align 1
+  // CHECK-NEXT:    [[RESIZE1:%.*]] = zext i8 [[TMP7]] to i16
+  // SIGNED-NEXT:   [[TMP8:%.*]] = call i16 @llvm.udiv.fix.i16(i16 [[TMP6]], i16 [[RESIZE1]], i32 8)
+  // UNSIGNED-NEXT: [[TMP8:%.*]] = call i16 @llvm.udiv.fix.i16(i16 [[TMP6]], i16 [[RESIZE1]], i32 7)
+  // CHECK-NEXT:    store i16 [[TMP8]], i16* %usa, align 2
+  usa = usa / usf;
+
+  // CHECK:         [[TMP9:%.*]] = load i16, i16* %usa, align 2
+  // CHECK-NEXT:    [[TMP10:%.*]] = load i16, i16* %uf, align 2
+  // CHECK-NEXT:    [[RESIZE2:%.*]] = zext i16 [[TMP9]] to i24
+  // CHECK-NEXT:    [[UPSCALE3:%.*]] = shl i24 [[RESIZE2]], 8
+  // CHECK-NEXT:    [[RESIZE4:%.*]] = zext i16 [[TMP10]] to i24
+  // SIGNED-NEXT:   [[TMP11:%.*]] = call i24 @llvm.udiv.fix.i24(i24 [[UPSCALE3]], i24 [[RESIZE4]], i32 16)
+  // UNSIGNED-NEXT: [[TMP11:%.*]] = call i24 @llvm.udiv.fix.i24(i24 [[UPSCALE3]], i24 [[RESIZE4]], i32 15)
+  // CHECK-NEXT:    [[DOWNSCALE:%.*]] = lshr i24 [[TMP11]], 8
+  // CHECK-NEXT:    [[RESIZE5:%.*]] = trunc i24 [[DOWNSCALE]] to i16
+  // CHECK-NEXT:    store i16 [[RESIZE5]], i16* %usa, align 2
+  usa = usa / uf;
+}
+
+void IntDivision() {
+  // CHECK-LABEL: IntDivision
+  short _Accum sa;
+  _Accum a;
+  unsigned short _Accum usa;
+  int i;
+  unsigned int ui;
+  long _Fract lf;
+  _Bool b;
+
+  // CHECK:         [[TMP0:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:    [[TMP1:%.*]] = load i32, i32* %i, align 4
+  // CHECK-NEXT:    [[RESIZE:%.*]] = sext i16 [[TMP0]] to i39
+  // CHECK-NEXT:    [[RESIZE1:%.*]] = sext i32 [[TMP1]] to i39
+  // CHECK-NEXT:    [[UPSCALE:%.*]] = shl i39 [[RESIZE1]], 7
+  // CHECK-NEXT:    [[TMP2:%.*]] = call i39 @llvm.sdiv.fix.i39(i39 [[RESIZE]], i39 [[UPSCALE]], i32 7)
+  // CHECK-NEXT:    [[RESIZE2:%.*]] = trunc i39 [[TMP2]] to i16
+  // CHECK-NEXT:    store i16 [[RESIZE2]], i16* %sa, align 2
+  sa = sa / i;
+
+  // CHECK:         [[TMP3:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:    [[TMP4:%.*]] = load i32, i32* %ui, align 4
+  // CHECK-NEXT:    [[RESIZE3:%.*]] = sext i16 [[TMP3]] to i40
+  // CHECK-NEXT:    [[RESIZE4:%.*]] = zext i32 [[TMP4]] to i40
+  // CHECK-NEXT:    [[UPSCALE5:%.*]] = shl i40 [[RESIZE4]], 7
+  // CHECK-NEXT:    [[TMP5:%.*]] = call i40 @llvm.sdiv.fix.i40(i40 [[RESIZE3]], i40 [[UPSCALE5]], i32 7)
+  // CHECK-NEXT:    [[RESIZE6:%.*]] = trunc i40 [[TMP5]] to i16
+  // CHECK-NEXT:    store i16 [[RESIZE6]], i16* %sa, align 2
+  sa = sa / ui;
+
+  // CHECK:         [[TMP6:%.*]] = load i16, i16* %usa, align 2
+  // CHECK-NEXT:    [[TMP7:%.*]] = load i32, i32* %i, align 4
+  // SIGNED-NEXT:   [[RESIZE7:%.*]] = zext i16 [[TMP6]] to i40
+  // SIGNED-NEXT:   [[RESIZE8:%.*]] = sext i32 [[TMP7]] to i40
+  // SIGNED-NEXT:   [[UPSCALE9:%.*]] = shl i40 [[RESIZE8]], 8
+  // SIGNED-NEXT:   [[TMP8:%.*]] = call i40 @llvm.udiv.fix.i40(i40 [[RESIZE7]], i40 [[UPSCALE9]], i32 8)
+  // SIGNED-NEXT:   [[RESIZE10:%.*]] = trunc i40 [[TMP8]] to i16
+  // UNSIGNED-NEXT: [[RESIZE7:%.*]] = zext i16 [[TMP6]] to i39
+  // UNSIGNED-NEXT: [[RESIZE8:%.*]] = sext i32 [[TMP7]] to i39
+  // UNSIGNED-NEXT: [[UPSCALE9:%.*]] = shl i39 [[RESIZE8]], 7
+  // UNSIGNED-NEXT: [[TMP8:%.*]] = call i39 @llvm.udiv.fix.i39(i39 [[RESIZE7]], i39 [[UPSCALE9]], i32 7)
+  // UNSIGNED-NEXT: [[RESIZE10:%.*]] = trunc i39 [[TMP8]] to i16
+  // CHECK-NEXT:    store i16 [[RESIZE10]], i16* %usa, align 2
+  usa = usa / i;
+
+  // CHECK:         [[TMP9:%.*]] = load i16, i16* %usa, align 2
+  // CHECK-NEXT:    [[TMP10:%.*]] = load i32, i32* %ui, align 4
+  // SIGNED-NEXT:   [[RESIZE11:%.*]] = zext i16 [[TMP9]] to i40
+  // SIGNED-NEXT:   [[RESIZE12:%.*]] = zext i32 [[TMP10]] to i40
+  // SIGNED-NEXT:   [[UPSCALE13:%.*]] = shl i40 [[RESIZE12]], 8
+  // SIGNED-NEXT:   [[TMP11:%.*]] = call i40 @llvm.udiv.fix.i40(i40 [[RESIZE11]], i40 [[UPSCALE13]], i32 8)
+  // SIGNED-NEXT:   [[RESIZE14:%.*]] = trunc i40 [[TMP11]] to i16
+  // UNSIGNED-NEXT: [[RESIZE11:%.*]] = zext i16 [[TMP9]] to i39
+  // UNSIGNED-NEXT: [[RESIZE12:%.*]] = zext i32 [[TMP10]] to i39
+  // UNSIGNED-NEXT: [[UPSCALE13:%.*]] = shl i39 [[RESIZE12]], 7
+  // UNSIGNED-NEXT: [[TMP11:%.*]] = call i39 @llvm.udiv.fix.i39(i39 [[RESIZE11]], i39 [[UPSCALE13]], i32 7)
+  // UNSIGNED-NEXT: [[RESIZE14:%.*]] = trunc i39 [[TMP11]] to i16
+  // CHECK-NEXT:    store i16 [[RESIZE14]], i16* %usa, align 2
+  usa = usa / ui;
+
+  // CHECK:         [[TMP12:%.*]] = load i32, i32* %lf, align 4
+  // CHECK-NEXT:    [[TMP13:%.*]] = load i32, i32* %ui, align 4
+  // CHECK-NEXT:    [[RESIZE15:%.*]] = sext i32 [[TMP12]] to i64
+  // CHECK-NEXT:    [[RESIZE16:%.*]] = zext i32 [[TMP13]] to i64
+  // CHECK-NEXT:    [[UPSCALE17:%.*]] = shl i64 [[RESIZE16]], 31
+  // CHECK-NEXT:    [[TMP14:%.*]] = call i64 @llvm.sdiv.fix.i64(i64 [[RESIZE15]], i64 [[UPSCALE17]], i32 31)
+  // CHECK-NEXT:    [[RESIZE18:%.*]] = trunc i64 [[TMP14]] to i32
+  // CHECK-NEXT:    store i32 [[RESIZE18]], i32* %lf, align 4
+  lf = lf / ui;
+
+  // CHECK:         [[TMP15:%.*]] = load i32, i32* %a, align 4
+  // CHECK-NEXT:    [[TMP16:%.*]] = load i8, i8* %b, align 1
+  // CHECK-NEXT:    [[TOBOOL:%.*]] = trunc i8 [[TMP16]] to i1
+  // CHECK-NEXT:    [[CONV:%.*]] = zext i1 [[TOBOOL]] to i32
+  // CHECK-NEXT:    [[RESIZE19:%.*]] = sext i32 [[TMP15]] to i47
+  // CHECK-NEXT:    [[RESIZE20:%.*]] = sext i32 [[CONV]] to i47
+  // CHECK-NEXT:    [[UPSCALE21:%.*]] = shl i47 [[RESIZE20]], 15
+  // CHECK-NEXT:    [[TMP17:%.*]] = call i47 @llvm.sdiv.fix.i47(i47 [[RESIZE19]], i47 [[UPSCALE21]], i32 15)
+  // CHECK-NEXT:    [[RESIZE22:%.*]] = trunc i47 [[TMP17]] to i32
+  // CHECK-NEXT:    store i32 [[RESIZE22]], i32* %a, align 4
+  a = a / b;
+
+  // CHECK:         [[TMP18:%.*]] = load i32, i32* %i, align 4
+  // CHECK-NEXT:    [[TMP19:%.*]] = load i32, i32* %a, align 4
+  // CHECK-NEXT:    [[RESIZE23:%.*]] = sext i32 [[TMP18]] to i47
+  // CHECK-NEXT:    [[UPSCALE24:%.*]] = shl i47 [[RESIZE23]], 15
+  // CHECK-NEXT:    [[RESIZE25:%.*]] = sext i32 [[TMP19]] to i47
+  // CHECK-NEXT:    [[TMP20:%.*]] = call i47 @llvm.sdiv.fix.i47(i47 [[UPSCALE24]], i47 [[RESIZE25]], i32 15)
+  // CHECK-NEXT:    [[RESIZE26:%.*]] = trunc i47 [[TMP20]] to i32
+  // CHECK-NEXT:    store i32 [[RESIZE26]], i32* %a, align 4
+  a = i / a;
+
+  // CHECK:         [[TMP21:%.*]] = load i32, i32* %ui, align 4
+  // CHECK-NEXT:    [[TMP22:%.*]] = load i16, i16* %usa, align 2
+  // SIGNED-NEXT:   [[RESIZE27:%.*]] = zext i32 [[TMP21]] to i40
+  // SIGNED-NEXT:   [[UPSCALE28:%.*]] = shl i40 [[RESIZE27]], 8
+  // SIGNED-NEXT:   [[RESIZE29:%.*]] = zext i16 [[TMP22]] to i40
+  // SIGNED-NEXT:   [[TMP23:%.*]] = call i40 @llvm.udiv.fix.i40(i40 [[UPSCALE28]], i40 [[RESIZE29]], i32 8)
+  // SIGNED-NEXT:   [[RESIZE30:%.*]] = trunc i40 [[TMP23]] to i16
+  // UNSIGNED-NEXT: [[RESIZE27:%.*]] = zext i32 [[TMP21]] to i39
+  // UNSIGNED-NEXT: [[UPSCALE28:%.*]] = shl i39 [[RESIZE27]], 7
+  // UNSIGNED-NEXT: [[RESIZE29:%.*]] = zext i16 [[TMP22]] to i39
+  // UNSIGNED-NEXT: [[TMP23:%.*]] = call i39 @llvm.udiv.fix.i39(i39 [[UPSCALE28]], i39 [[RESIZE29]], i32 7)
+  // UNSIGNED-NEXT: [[RESIZE30:%.*]] = trunc i39 [[TMP23]] to i16
+  // CHECK-NEXT:    store i16 [[RESIZE30]], i16* %usa, align 2
+  usa = ui / usa;
+
+  // CHECK:         [[TMP27:%.*]] = load i32, i32* %ui, align 4
+  // CHECK-NEXT:    [[TMP28:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:    [[RESIZE33:%.*]] = zext i32 [[TMP27]] to i40
+  // CHECK-NEXT:    [[UPSCALE34:%.*]] = shl i40 [[RESIZE33]], 7
+  // CHECK-NEXT:    [[RESIZE35:%.*]] = sext i16 [[TMP28]] to i40
+  // CHECK-NEXT:    [[TMP29:%.*]] = call i40 @llvm.sdiv.fix.i40(i40 [[UPSCALE34]], i40 [[RESIZE35]], i32 7)
+  // CHECK-NEXT:    [[RESIZE36:%.*]] = trunc i40 [[TMP29]] to i16
+  // CHECK-NEXT:    store i16 [[RESIZE36]], i16* %sa, align 2
+  sa = ui / sa;
+}
+
+void SaturatedDivision() {
+  // CHECK-LABEL: SaturatedDivision
+  short _Accum sa;
+  _Accum a;
+  long _Accum la;
+  unsigned short _Accum usa;
+  unsigned _Accum ua;
+  unsigned long _Accum ula;
+
+  _Sat short _Accum sa_sat;
+  _Sat _Accum a_sat;
+  _Sat long _Accum la_sat;
+  _Sat unsigned short _Accum usa_sat;
+  _Sat unsigned _Accum ua_sat;
+  _Sat unsigned long _Accum ula_sat;
+  _Sat unsigned _Fract uf_sat;
+
+  int i;
+  unsigned int ui;
+
+  // CHECK:      [[SA:%[0-9]+]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT: [[SA_SAT:%[0-9]+]] = load i16, i16* %sa_sat, align 2
+  // CHECK-NEXT: [[SUM:%[0-9]+]] = call i16 @llvm.sdiv.fix.sat.i16(i16 [[SA]], i16 [[SA_SAT]], i32 7)
+  // CHECK-NEXT: store i16 [[SUM]], i16* %sa_sat, align 2
+  sa_sat = sa / sa_sat;
+
+  // CHECK:      [[USA:%[0-9]+]] = load i16, i16* %usa, align 2
+  // CHECK-NEXT: [[USA_SAT:%[0-9]+]] = load i16, i16* %usa_sat, align 2
+  // SIGNED-NEXT: [[SUM:%[0-9]+]] = call i16 @llvm.udiv.fix.sat.i16(i16 [[USA]], i16 [[USA_SAT]], i32 8)
+  // SIGNED-NEXT: store i16 [[SUM]], i16* %usa_sat, align 2
+  // UNSIGNED-NEXT: [[USA_TRUNC:%[a-z0-9]+]] = trunc i16 [[USA]] to i15
+  // UNSIGNED-NEXT: [[USA_SAT_TRUNC:%[a-z0-9]+]] = trunc i16 [[USA_SAT]] to i15
+  // UNSIGNED-NEXT: [[SUM:%[0-9]+]] = call i15 @llvm.udiv.fix.sat.i15(i15 [[USA_TRUNC]], i15 [[USA_SAT_TRUNC]], i32 7)
+  // UNSIGNED-NEXT: [[SUM_EXT:%[a-z0-9]+]] = zext i15 [[SUM]] to i16
+  // UNSIGNED-NEXT: store i16 [[SUM_EXT]], i16* %usa_sat, align 2
+  usa_sat = usa / usa_sat;
+
+  // CHECK:      [[UA:%[0-9]+]] = load i32, i32* %ua, align 4
+  // CHECK-NEXT: [[USA:%[0-9]+]] = load i16, i16* %usa_sat, align 2
+  // SIGNED-NEXT: [[USA_EXT:%[a-z0-9]+]] = zext i16 [[USA]] to i32
+  // SIGNED-NEXT: [[USA:%[a-z0-9]+]] = shl i32 [[USA_EXT]], 8
+  // SIGNED-NEXT: [[SUM:%[0-9]+]] = call i32 @llvm.udiv.fix.sat.i32(i32 [[UA]], i32 [[USA]], i32 16)
+  // SIGNED-NEXT: store i32 [[SUM]], i32* %ua_sat, align 4
+  // UNSIGNED-NEXT: [[UA_TRUNC:%[a-z0-9]+]] = trunc i32 [[UA]] to i31
+  // UNSIGNED-NEXT: [[USA_EXT:%[a-z0-9]+]] = zext i16 [[USA]] to i31
+  // UNSIGNED-NEXT: [[USA:%[a-z0-9]+]] = shl i31 [[USA_EXT]], 8
+  // UNSIGNED-NEXT: [[SUM:%[0-9]+]] = call i31 @llvm.udiv.fix.sat.i31(i31 [[UA_TRUNC]], i31 [[USA]], i32 15)
+  // UNSIGNED-NEXT: [[SUM_EXT:%[a-z0-9]+]] = zext i31 [[SUM]] to i32
+  // UNSIGNED-NEXT: store i32 [[SUM_EXT]], i32* %ua_sat, align 4
+  ua_sat = ua / usa_sat;
+
+  // CHECK:      [[SA_SAT:%[0-9]+]] = load i16, i16* %sa_sat, align 2
+  // CHECK-NEXT: [[I:%[0-9]+]] = load i32, i32* %i, align 4
+  // CHECK-NEXT: [[SA_SAT_EXT:%[a-z0-9]+]] = sext i16 [[SA_SAT]] to i39
+  // CHECK-NEXT: [[I_EXT:%[a-z0-9]+]] = sext i32 [[I]] to i39
+  // CHECK-NEXT: [[I:%[a-z0-9]+]] = shl i39 [[I_EXT]], 7
+  // CHECK-NEXT: [[SUM:%[0-9]+]] = call i39 @llvm.sdiv.fix.sat.i39(i39 [[SA_SAT_EXT]], i39 [[I]], i32 7)
+  // CHECK-NEXT: [[USE_MAX:%[0-9]+]] = icmp sgt i39 [[SUM]], 32767
+  // CHECK-NEXT: [[RES:%[a-z0-9]+]] = select i1 [[USE_MAX]], i39 32767, i39 [[SUM]]
+  // CHECK-NEXT: [[USE_MIN:%[0-9]+]] = icmp slt i39 [[RES]], -32768
+  // CHECK-NEXT: [[RES2:%[a-z0-9]+]] = select i1 [[USE_MIN]], i39 -32768, i39 [[RES]]
+  // CHECK-NEXT: [[RES3:%[a-z0-9]+]] = trunc i39 [[RES2]] to i16
+  // CHECK-NEXT: store i16 [[RES3]], i16* %sa_sat, align 2
+  sa_sat = sa_sat / i;
+
+  // CHECK:      [[SA_SAT:%[0-9]+]] = load i16, i16* %sa_sat, align 2
+  // CHECK-NEXT: [[I:%[0-9]+]] = load i32, i32* %ui, align 4
+  // CHECK-NEXT: [[SA_SAT_EXT:%[a-z0-9]+]] = sext i16 [[SA_SAT]] to i40
+  // CHECK-NEXT: [[I_EXT:%[a-z0-9]+]] = zext i32 [[I]] to i40
+  // CHECK-NEXT: [[I:%[a-z0-9]+]] = shl i40 [[I_EXT]], 7
+  // CHECK-NEXT: [[SUM:%[0-9]+]] = call i40 @llvm.sdiv.fix.sat.i40(i40 [[SA_SAT_EXT]], i40 [[I]], i32 7)
+  // CHECK-NEXT: [[USE_MAX:%[0-9]+]] = icmp sgt i40 [[SUM]], 32767
+  // CHECK-NEXT: [[RES:%[a-z0-9]+]] = select i1 [[USE_MAX]], i40 32767, i40 [[SUM]]
+  // CHECK-NEXT: [[USE_MIN:%[0-9]+]] = icmp slt i40 [[RES]], -32768
+  // CHECK-NEXT: [[RES2:%[a-z0-9]+]] = select i1 [[USE_MIN]], i40 -32768, i40 [[RES]]
+  // CHECK-NEXT: [[RES3:%[a-z0-9]+]] = trunc i40 [[RES2]] to i16
+  // CHECK-NEXT: store i16 [[RES3]], i16* %sa_sat, align 2
+  sa_sat = sa_sat / ui;
+
+  // CHECK:      [[UF_SAT:%[0-9]+]] = load i16, i16* %uf_sat, align 2
+  // CHECK-NEXT: [[UF_SAT2:%[0-9]+]] = load i16, i16* %uf_sat, align 2
+  // SIGNED-NEXT: [[SUM:%[0-9]+]] = call i16 @llvm.udiv.fix.sat.i16(i16 [[UF_SAT]], i16 [[UF_SAT2]], i32 16)
+  // SIGNED-NEXT: store i16 [[SUM]], i16* %uf_sat, align 2
+  // UNSIGNED-NEXT: [[UF_SAT_TRUNC:%[a-z0-9]+]] = trunc i16 [[UF_SAT]] to i15
+  // UNSIGNED-NEXT: [[UF_SAT_TRUNC2:%[a-z0-9]+]] = trunc i16 [[UF_SAT2]] to i15
+  // UNSIGNED-NEXT: [[SUM:%[0-9]+]] = call i15 @llvm.udiv.fix.sat.i15(i15 [[UF_SAT_TRUNC]], i15 [[UF_SAT_TRUNC2]], i32 15)
+  // UNSIGNED-NEXT: [[SUM_EXT:%[a-z0-9]+]] = zext i15 [[SUM]] to i16
+  // UNSIGNED-NEXT: store i16 [[SUM_EXT]], i16* %uf_sat, align 2
+  uf_sat = uf_sat / uf_sat;
+
+  // CHECK:      [[USA_SAT:%[0-9]+]] = load i16, i16* %usa_sat, align 2
+  // CHECK-NEXT: [[I:%[0-9]+]] = load i32, i32* %i, align 4
+  // SIGNED-NEXT: [[USA_SAT_RESIZE:%[a-z0-9]+]] = zext i16 [[USA_SAT]] to i40
+  // SIGNED-NEXT: [[I_RESIZE:%[a-z0-9]+]] = sext i32 [[I]] to i40
+  // SIGNED-NEXT: [[I_UPSCALE:%[a-z0-9]+]] = shl i40 [[I_RESIZE]], 8
+  // SIGNED-NEXT: [[SUM:%[0-9]+]] = call i40 @llvm.udiv.fix.sat.i40(i40 [[USA_SAT_RESIZE]], i40 [[I_UPSCALE]], i32 8)
+  // SIGNED-NEXT: [[USE_MAX:%[0-9]+]] = icmp sgt i40 [[SUM]], 65535
+  // SIGNED-NEXT: [[RESULT:%[a-z0-9]+]] = select i1 [[USE_MAX]], i40 65535, i40 [[SUM]]
+  // SIGNED-NEXT: [[USE_MIN:%[0-9]+]] = icmp slt i40 [[RESULT]], 0
+  // SIGNED-NEXT: [[RESULT2:%[a-z0-9]+]] = select i1 [[USE_MIN]], i40 0, i40 [[RESULT]]
+  // SIGNED-NEXT: [[RESULT:%[a-z0-9]+]] = trunc i40 [[RESULT2]] to i16
+  // UNSIGNED-NEXT: [[USA_SAT_RESIZE:%[a-z0-9]+]] = zext i16 [[USA_SAT]] to i39
+  // UNSIGNED-NEXT: [[I_RESIZE:%[a-z0-9]+]] = sext i32 [[I]] to i39
+  // UNSIGNED-NEXT: [[I_UPSCALE:%[a-z0-9]+]] = shl i39 [[I_RESIZE]], 7
+  // UNSIGNED-NEXT: [[SUM:%[0-9]+]] = call i39 @llvm.udiv.fix.sat.i39(i39 [[USA_SAT_RESIZE]], i39 [[I_UPSCALE]], i32 7)
+  // UNSIGNED-NEXT: [[USE_MAX:%[0-9]+]] = icmp sgt i39 [[SUM]], 32767
+  // UNSIGNED-NEXT: [[RESULT:%[a-z0-9]+]] = select i1 [[USE_MAX]], i39 32767, i39 [[SUM]]
+  // UNSIGNED-NEXT: [[USE_MIN:%[0-9]+]] = icmp slt i39 [[RESULT]], 0
+  // UNSIGNED-NEXT: [[RESULT2:%[a-z0-9]+]] = select i1 [[USE_MIN]], i39 0, i39 [[RESULT]]
+  // UNSIGNED-NEXT: [[RESULT:%[a-z0-9]+]] = trunc i39 [[RESULT2]] to i16
+  // CHECK-NEXT: store i16 [[RESULT]], i16* %usa_sat, align 2
+  usa_sat = usa_sat / i;
+}

diff  --git a/clang/test/Frontend/fixed_point_mul.c b/clang/test/Frontend/fixed_point_mul.c
new file mode 100644
index 000000000000..274528fe9d9a
--- /dev/null
+++ b/clang/test/Frontend/fixed_point_mul.c
@@ -0,0 +1,431 @@
+// RUN: %clang_cc1 -ffixed-point -triple x86_64-unknown-linux-gnu -S -emit-llvm %s -o - | FileCheck %s --check-prefixes=CHECK,SIGNED
+// RUN: %clang_cc1 -ffixed-point -triple x86_64-unknown-linux-gnu -fpadding-on-unsigned-fixed-point -S -emit-llvm %s -o - | FileCheck %s --check-prefixes=CHECK,UNSIGNED
+
+void SignedMultiplication() {
+  // CHECK-LABEL: SignedMultiplication
+  short _Accum sa;
+  _Accum a, b, c, d;
+  long _Accum la;
+  unsigned short _Accum usa;
+  unsigned _Accum ua;
+  unsigned long _Accum ula;
+
+  short _Fract sf;
+  _Fract f;
+  long _Fract lf;
+  unsigned short _Fract usf;
+  unsigned _Fract uf;
+  unsigned long _Fract ulf;
+
+  // Same type
+  // CHECK:       [[TMP0:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:  [[TMP1:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:  [[TMP2:%.*]] = call i16 @llvm.smul.fix.i16(i16 [[TMP0]], i16 [[TMP1]], i32 7)
+  // CHECK-NEXT:  store i16 [[TMP2]], i16* %sa, align 2
+  sa = sa * sa;
+
+  // To larger scale and larger width
+  // CHECK:       [[TMP3:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:  [[TMP4:%.*]] = load i32, i32* %a, align 4
+  // CHECK-NEXT:  [[RESIZE:%.*]] = sext i16 [[TMP3]] to i32
+  // CHECK-NEXT:  [[UPSCALE:%.*]] = shl i32 [[RESIZE]], 8
+  // CHECK-NEXT:  [[TMP5:%.*]] = call i32 @llvm.smul.fix.i32(i32 [[UPSCALE]], i32 [[TMP4]], i32 15)
+  // CHECK-NEXT:  store i32 [[TMP5]], i32* %a, align 4
+  a = sa * a;
+
+  // To same scale and smaller width
+  // CHECK:       [[TMP6:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:  [[TMP7:%.*]] = load i8, i8* %sf, align 1
+  // CHECK-NEXT:  [[RESIZE1:%.*]] = sext i8 [[TMP7]] to i16
+  // CHECK-NEXT:  [[TMP8:%.*]] = call i16 @llvm.smul.fix.i16(i16 [[TMP6]], i16 [[RESIZE1]], i32 7)
+  // CHECK-NEXT:  store i16 [[TMP8]], i16* %sa, align 2
+  sa = sa * sf;
+
+  // To smaller scale and same width.
+  // CHECK:       [[TMP9:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:  [[TMP10:%.*]] = load i16, i16* %f, align 2
+  // CHECK-NEXT:  [[RESIZE2:%.*]] = sext i16 [[TMP9]] to i24
+  // CHECK-NEXT:  [[UPSCALE3:%.*]] = shl i24 [[RESIZE2]], 8
+  // CHECK-NEXT:  [[RESIZE4:%.*]] = sext i16 [[TMP10]] to i24
+  // CHECK-NEXT:  [[TMP11:%.*]] = call i24 @llvm.smul.fix.i24(i24 [[UPSCALE3]], i24 [[RESIZE4]], i32 15)
+  // CHECK-NEXT:  [[DOWNSCALE:%.*]] = ashr i24 [[TMP11]], 8
+  // CHECK-NEXT:  [[RESIZE5:%.*]] = trunc i24 [[DOWNSCALE]] to i16
+  // CHECK-NEXT:  store i16 [[RESIZE5]], i16* %sa, align 2
+  sa = sa * f;
+
+  // To smaller scale and smaller width
+  // CHECK:       [[TMP12:%.*]] = load i32, i32* %a, align 4
+  // CHECK-NEXT:  [[TMP13:%.*]] = load i8, i8* %sf, align 1
+  // CHECK-NEXT:  [[RESIZE6:%.*]] = sext i8 [[TMP13]] to i32
+  // CHECK-NEXT:  [[UPSCALE7:%.*]] = shl i32 [[RESIZE6]], 8
+  // CHECK-NEXT:  [[TMP14:%.*]] = call i32 @llvm.smul.fix.i32(i32 [[TMP12]], i32 [[UPSCALE7]], i32 15)
+  // CHECK-NEXT:  store i32 [[TMP14]], i32* %a, align 4
+  a = a * sf;
+
+  // To larger scale and same width
+  // CHECK:       [[TMP15:%.*]] = load i32, i32* %a, align 4
+  // CHECK-NEXT:  [[TMP16:%.*]] = load i32, i32* %lf, align 4
+  // CHECK-NEXT:  [[RESIZE8:%.*]] = sext i32 [[TMP15]] to i48
+  // CHECK-NEXT:  [[UPSCALE9:%.*]] = shl i48 [[RESIZE8]], 16
+  // CHECK-NEXT:  [[RESIZE10:%.*]] = sext i32 [[TMP16]] to i48
+  // CHECK-NEXT:  [[TMP17:%.*]] = call i48 @llvm.smul.fix.i48(i48 [[UPSCALE9]], i48 [[RESIZE10]], i32 31)
+  // CHECK-NEXT:  [[DOWNSCALE11:%.*]] = ashr i48 [[TMP17]], 16
+  // CHECK-NEXT:  [[RESIZE12:%.*]] = trunc i48 [[DOWNSCALE11]] to i32
+  // CHECK-NEXT:  store i32 [[RESIZE12]], i32* %a, align 4
+  a = a * lf;
+
+  // With corresponding unsigned type
+  // CHECK:        [[TMP18:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:   [[TMP19:%.*]] = load i16, i16* %usa, align 2
+  // SIGNED-NEXT:  [[RESIZE13:%.*]] = sext i16 [[TMP18]] to i17
+  // SIGNED-NEXT:  [[UPSCALE14:%.*]] = shl i17 [[RESIZE13]], 1
+  // SIGNED-NEXT:  [[RESIZE15:%.*]] = zext i16 [[TMP19]] to i17
+  // SIGNED-NEXT:  [[TMP20:%.*]] = call i17 @llvm.smul.fix.i17(i17 [[UPSCALE14]], i17 [[RESIZE15]], i32 8)
+  // SIGNED-NEXT:  [[DOWNSCALE16:%.*]] = ashr i17 [[TMP20]], 1
+  // SIGNED-NEXT:  [[RESIZE17:%.*]] = trunc i17 [[DOWNSCALE16]] to i16
+  // SIGNED-NEXT:  store i16 [[RESIZE17]], i16* %sa, align 2
+  // UNSIGNED-NEXT:[[TMP20:%.*]] = call i16 @llvm.smul.fix.i16(i16 [[TMP18]], i16 [[TMP19]], i32 7)
+  // UNSIGNED-NEXT:store i16 [[TMP20]], i16* %sa, align 2
+  sa = sa * usa;
+
+  // With unsigned of larger scale
+  // CHECK:        [[TMP21:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:   [[TMP22:%.*]] = load i32, i32* %ua, align 4
+  // SIGNED-NEXT:  [[RESIZE18:%.*]] = sext i16 [[TMP21]] to i33
+  // SIGNED-NEXT:  [[UPSCALE19:%.*]] = shl i33 [[RESIZE18]], 9
+  // SIGNED-NEXT:  [[RESIZE20:%.*]] = zext i32 [[TMP22]] to i33
+  // SIGNED-NEXT:  [[TMP23:%.*]] = call i33 @llvm.smul.fix.i33(i33 [[UPSCALE19]], i33 [[RESIZE20]], i32 16)
+  // SIGNED-NEXT:  [[DOWNSCALE21:%.*]] = ashr i33 [[TMP23]], 1
+  // SIGNED-NEXT:  [[RESIZE22:%.*]] = trunc i33 [[DOWNSCALE21]] to i32
+  // SIGNED-NEXT:  store i32 [[RESIZE22]], i32* %a, align 4
+  // UNSIGNED-NEXT:[[RESIZE13:%.*]] = sext i16 [[TMP21]] to i32
+  // UNSIGNED-NEXT:[[UPSCALE14:%.*]] = shl i32 [[RESIZE13]], 8
+  // UNSIGNED-NEXT:[[TMP23:%.*]] = call i32 @llvm.smul.fix.i32(i32 [[UPSCALE14]], i32 [[TMP22]], i32 15)
+  // UNSIGNED-NEXT:store i32 [[TMP23]], i32* %a, align 4
+  a = sa * ua;
+
+  // With unsigned of smaller width
+  // CHECK:        [[TMP24:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:   [[TMP25:%.*]] = load i8, i8* %usf, align 1
+  // SIGNED-NEXT:  [[RESIZE23:%.*]] = sext i16 [[TMP24]] to i17
+  // SIGNED-NEXT:  [[UPSCALE24:%.*]] = shl i17 [[RESIZE23]], 1
+  // SIGNED-NEXT:  [[RESIZE25:%.*]] = zext i8 [[TMP25]] to i17
+  // SIGNED-NEXT:  [[TMP26:%.*]] = call i17 @llvm.smul.fix.i17(i17 [[UPSCALE24]], i17 [[RESIZE25]], i32 8)
+  // SIGNED-NEXT:  [[DOWNSCALE26:%.*]] = ashr i17 [[TMP26]], 1
+  // SIGNED-NEXT:  [[RESIZE27:%.*]] = trunc i17 [[DOWNSCALE26]] to i16
+  // SIGNED-NEXT:  store i16 [[RESIZE27]], i16* %sa, align 2
+  // UNSIGNED-NEXT:[[RESIZE15:%.*]] = zext i8 [[TMP25]] to i16
+  // UNSIGNED-NEXT:[[TMP26:%.*]] = call i16 @llvm.smul.fix.i16(i16 [[TMP24]], i16 [[RESIZE15]], i32 7)
+  // UNSIGNED-NEXT:store i16 [[TMP26]], i16* %sa, align 2
+  sa = sa * usf;
+
+  // With unsigned of larger width and smaller scale
+  // CHECK:        [[TMP27:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:   [[TMP28:%.*]] = load i32, i32* %ulf, align 4
+  // SIGNED-NEXT:  [[RESIZE28:%.*]] = sext i16 [[TMP27]] to i41
+  // SIGNED-NEXT:  [[UPSCALE29:%.*]] = shl i41 [[RESIZE28]], 25
+  // SIGNED-NEXT:  [[RESIZE30:%.*]] = zext i32 [[TMP28]] to i41
+  // SIGNED-NEXT:  [[TMP29:%.*]] = call i41 @llvm.smul.fix.i41(i41 [[UPSCALE29]], i41 [[RESIZE30]], i32 32)
+  // SIGNED-NEXT:  [[DOWNSCALE31:%.*]] = ashr i41 [[TMP29]], 25
+  // SIGNED-NEXT:  [[RESIZE32:%.*]] = trunc i41 [[DOWNSCALE31]] to i16
+  // SIGNED-NEXT:  store i16 [[RESIZE32]], i16* %sa, align 2
+  // UNSIGNED-NEXT:[[RESIZE16:%.*]] = sext i16 [[TMP27]] to i40
+  // UNSIGNED-NEXT:[[UPSCALE17:%.*]] = shl i40 [[RESIZE16]], 24
+  // UNSIGNED-NEXT:[[RESIZE18:%.*]] = zext i32 [[TMP28]] to i40
+  // UNSIGNED-NEXT:[[TMP29:%.*]] = call i40 @llvm.smul.fix.i40(i40 [[UPSCALE17]], i40 [[RESIZE18]], i32 31)
+  // UNSIGNED-NEXT:[[DOWNSCALE19:%.*]] = ashr i40 [[TMP29]], 24
+  // UNSIGNED-NEXT:[[RESIZE20:%.*]] = trunc i40 [[DOWNSCALE19]] to i16
+  // UNSIGNED-NEXT:store i16 [[RESIZE20]], i16* %sa, align 2
+  sa = sa * ulf;
+
+  // Chained multiplications of the same signed type should result in the same
+  // CHECK:       [[TMP30:%.*]] = load i32, i32* %a, align 4
+  // CHECK-NEXT:  [[TMP31:%.*]] = load i32, i32* %b, align 4
+  // CHECK-NEXT:  [[TMP32:%.*]] = call i32 @llvm.smul.fix.i32(i32 [[TMP30]], i32 [[TMP31]], i32 15)
+  // CHECK-NEXT:  [[TMP33:%.*]] = load i32, i32* %c, align 4
+  // CHECK-NEXT:  [[TMP34:%.*]] = call i32 @llvm.smul.fix.i32(i32 [[TMP32]], i32 [[TMP33]], i32 15)
+  // CHECK-NEXT:  [[TMP35:%.*]] = load i32, i32* %d, align 4
+  // CHECK-NEXT:  [[TMP36:%.*]] = call i32 @llvm.smul.fix.i32(i32 [[TMP34]], i32 [[TMP35]], i32 15)
+  // CHECK-NEXT:  store i32 [[TMP36]], i32* %a, align 4
+  a = a * b * c * d;
+}
+
+
+void UnsignedMultiplication() {
+  // CHECK-LABEL: UnsignedMultiplication
+  unsigned short _Accum usa;
+  unsigned _Accum ua;
+  unsigned long _Accum ula;
+
+  unsigned short _Fract usf;
+  unsigned _Fract uf;
+  unsigned long _Fract ulf;
+
+  // CHECK:         [[TMP0:%.*]] = load i16, i16* %usa, align 2
+  // CHECK-NEXT:    [[TMP1:%.*]] = load i16, i16* %usa, align 2
+  // SIGNED-NEXT:   [[TMP2:%.*]] = call i16 @llvm.umul.fix.i16(i16 [[TMP0]], i16 [[TMP1]], i32 8)
+  // UNSIGNED-NEXT: [[TMP2:%.*]] = call i16 @llvm.umul.fix.i16(i16 [[TMP0]], i16 [[TMP1]], i32 7)
+  // CHECK-NEXT:    store i16 [[TMP2]], i16* %usa, align 2
+  usa = usa * usa;
+
+  // CHECK:         [[TMP3:%.*]] = load i16, i16* %usa, align 2
+  // CHECK-NEXT:    [[TMP4:%.*]] = load i32, i32* %ua, align 4
+  // CHECK-NEXT:    [[RESIZE:%.*]] = zext i16 [[TMP3]] to i32
+  // CHECK-NEXT:    [[UPSCALE:%.*]] = shl i32 [[RESIZE]], 8
+  // SIGNED-NEXT:   [[TMP5:%.*]] = call i32 @llvm.umul.fix.i32(i32 [[UPSCALE]], i32 [[TMP4]], i32 16)
+  // UNSIGNED-NEXT: [[TMP5:%.*]] = call i32 @llvm.umul.fix.i32(i32 [[UPSCALE]], i32 [[TMP4]], i32 15)
+  // CHECK-NEXT:    store i32 [[TMP5]], i32* %ua, align 4
+  ua = usa * ua;
+
+  // CHECK:         [[TMP6:%.*]] = load i16, i16* %usa, align 2
+  // CHECK-NEXT:    [[TMP7:%.*]] = load i8, i8* %usf, align 1
+  // CHECK-NEXT:    [[RESIZE1:%.*]] = zext i8 [[TMP7]] to i16
+  // SIGNED-NEXT:   [[TMP8:%.*]] = call i16 @llvm.umul.fix.i16(i16 [[TMP6]], i16 [[RESIZE1]], i32 8)
+  // UNSIGNED-NEXT: [[TMP8:%.*]] = call i16 @llvm.umul.fix.i16(i16 [[TMP6]], i16 [[RESIZE1]], i32 7)
+  // CHECK-NEXT:    store i16 [[TMP8]], i16* %usa, align 2
+  usa = usa * usf;
+
+  // CHECK:         [[TMP9:%.*]] = load i16, i16* %usa, align 2
+  // CHECK-NEXT:    [[TMP10:%.*]] = load i16, i16* %uf, align 2
+  // CHECK-NEXT:    [[RESIZE2:%.*]] = zext i16 [[TMP9]] to i24
+  // CHECK-NEXT:    [[UPSCALE3:%.*]] = shl i24 [[RESIZE2]], 8
+  // CHECK-NEXT:    [[RESIZE4:%.*]] = zext i16 [[TMP10]] to i24
+  // SIGNED-NEXT:   [[TMP11:%.*]] = call i24 @llvm.umul.fix.i24(i24 [[UPSCALE3]], i24 [[RESIZE4]], i32 16)
+  // UNSIGNED-NEXT: [[TMP11:%.*]] = call i24 @llvm.umul.fix.i24(i24 [[UPSCALE3]], i24 [[RESIZE4]], i32 15)
+  // CHECK-NEXT:    [[DOWNSCALE:%.*]] = lshr i24 [[TMP11]], 8
+  // CHECK-NEXT:    [[RESIZE5:%.*]] = trunc i24 [[DOWNSCALE]] to i16
+  // CHECK-NEXT:    store i16 [[RESIZE5]], i16* %usa, align 2
+  usa = usa * uf;
+}
+
+void IntMultiplication() {
+  // CHECK-LABEL: IntMultiplication
+  short _Accum sa;
+  _Accum a;
+  unsigned short _Accum usa;
+  int i;
+  unsigned int ui;
+  long _Fract lf;
+  _Bool b;
+
+  // CHECK:         [[TMP0:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:    [[TMP1:%.*]] = load i32, i32* %i, align 4
+  // CHECK-NEXT:    [[RESIZE:%.*]] = sext i16 [[TMP0]] to i39
+  // CHECK-NEXT:    [[RESIZE1:%.*]] = sext i32 [[TMP1]] to i39
+  // CHECK-NEXT:    [[UPSCALE:%.*]] = shl i39 [[RESIZE1]], 7
+  // CHECK-NEXT:    [[TMP2:%.*]] = call i39 @llvm.smul.fix.i39(i39 [[RESIZE]], i39 [[UPSCALE]], i32 7)
+  // CHECK-NEXT:    [[RESIZE2:%.*]] = trunc i39 [[TMP2]] to i16
+  // CHECK-NEXT:    store i16 [[RESIZE2]], i16* %sa, align 2
+  sa = sa * i;
+
+  // CHECK:         [[TMP3:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:    [[TMP4:%.*]] = load i32, i32* %ui, align 4
+  // CHECK-NEXT:    [[RESIZE3:%.*]] = sext i16 [[TMP3]] to i40
+  // CHECK-NEXT:    [[RESIZE4:%.*]] = zext i32 [[TMP4]] to i40
+  // CHECK-NEXT:    [[UPSCALE5:%.*]] = shl i40 [[RESIZE4]], 7
+  // CHECK-NEXT:    [[TMP5:%.*]] = call i40 @llvm.smul.fix.i40(i40 [[RESIZE3]], i40 [[UPSCALE5]], i32 7)
+  // CHECK-NEXT:    [[RESIZE6:%.*]] = trunc i40 [[TMP5]] to i16
+  // CHECK-NEXT:    store i16 [[RESIZE6]], i16* %sa, align 2
+  sa = sa * ui;
+
+  // CHECK:         [[TMP6:%.*]] = load i16, i16* %usa, align 2
+  // CHECK-NEXT:    [[TMP7:%.*]] = load i32, i32* %i, align 4
+  // SIGNED-NEXT:   [[RESIZE7:%.*]] = zext i16 [[TMP6]] to i40
+  // SIGNED-NEXT:   [[RESIZE8:%.*]] = sext i32 [[TMP7]] to i40
+  // SIGNED-NEXT:   [[UPSCALE9:%.*]] = shl i40 [[RESIZE8]], 8
+  // SIGNED-NEXT:   [[TMP8:%.*]] = call i40 @llvm.umul.fix.i40(i40 [[RESIZE7]], i40 [[UPSCALE9]], i32 8)
+  // SIGNED-NEXT:   [[RESIZE10:%.*]] = trunc i40 [[TMP8]] to i16
+  // UNSIGNED-NEXT: [[RESIZE7:%.*]] = zext i16 [[TMP6]] to i39
+  // UNSIGNED-NEXT: [[RESIZE8:%.*]] = sext i32 [[TMP7]] to i39
+  // UNSIGNED-NEXT: [[UPSCALE9:%.*]] = shl i39 [[RESIZE8]], 7
+  // UNSIGNED-NEXT: [[TMP8:%.*]] = call i39 @llvm.umul.fix.i39(i39 [[RESIZE7]], i39 [[UPSCALE9]], i32 7)
+  // UNSIGNED-NEXT: [[RESIZE10:%.*]] = trunc i39 [[TMP8]] to i16
+  // CHECK-NEXT:    store i16 [[RESIZE10]], i16* %usa, align 2
+  usa = usa * i;
+
+  // CHECK:         [[TMP9:%.*]] = load i16, i16* %usa, align 2
+  // CHECK-NEXT:    [[TMP10:%.*]] = load i32, i32* %ui, align 4
+  // SIGNED-NEXT:   [[RESIZE11:%.*]] = zext i16 [[TMP9]] to i40
+  // SIGNED-NEXT:   [[RESIZE12:%.*]] = zext i32 [[TMP10]] to i40
+  // SIGNED-NEXT:   [[UPSCALE13:%.*]] = shl i40 [[RESIZE12]], 8
+  // SIGNED-NEXT:   [[TMP11:%.*]] = call i40 @llvm.umul.fix.i40(i40 [[RESIZE11]], i40 [[UPSCALE13]], i32 8)
+  // SIGNED-NEXT:   [[RESIZE14:%.*]] = trunc i40 [[TMP11]] to i16
+  // UNSIGNED-NEXT: [[RESIZE11:%.*]] = zext i16 [[TMP9]] to i39
+  // UNSIGNED-NEXT: [[RESIZE12:%.*]] = zext i32 [[TMP10]] to i39
+  // UNSIGNED-NEXT: [[UPSCALE13:%.*]] = shl i39 [[RESIZE12]], 7
+  // UNSIGNED-NEXT: [[TMP11:%.*]] = call i39 @llvm.umul.fix.i39(i39 [[RESIZE11]], i39 [[UPSCALE13]], i32 7)
+  // UNSIGNED-NEXT: [[RESIZE14:%.*]] = trunc i39 [[TMP11]] to i16
+  // CHECK-NEXT:    store i16 [[RESIZE14]], i16* %usa, align 2
+  usa = usa * ui;
+
+  // CHECK:         [[TMP12:%.*]] = load i32, i32* %lf, align 4
+  // CHECK-NEXT:    [[TMP13:%.*]] = load i32, i32* %ui, align 4
+  // CHECK-NEXT:    [[RESIZE15:%.*]] = sext i32 [[TMP12]] to i64
+  // CHECK-NEXT:    [[RESIZE16:%.*]] = zext i32 [[TMP13]] to i64
+  // CHECK-NEXT:    [[UPSCALE17:%.*]] = shl i64 [[RESIZE16]], 31
+  // CHECK-NEXT:    [[TMP14:%.*]] = call i64 @llvm.smul.fix.i64(i64 [[RESIZE15]], i64 [[UPSCALE17]], i32 31)
+  // CHECK-NEXT:    [[RESIZE18:%.*]] = trunc i64 [[TMP14]] to i32
+  // CHECK-NEXT:    store i32 [[RESIZE18]], i32* %lf, align 4
+  lf = lf * ui;
+
+  // CHECK:         [[TMP15:%.*]] = load i32, i32* %a, align 4
+  // CHECK-NEXT:    [[TMP16:%.*]] = load i8, i8* %b, align 1
+  // CHECK-NEXT:    [[TOBOOL:%.*]] = trunc i8 [[TMP16]] to i1
+  // CHECK-NEXT:    [[CONV:%.*]] = zext i1 [[TOBOOL]] to i32
+  // CHECK-NEXT:    [[RESIZE19:%.*]] = sext i32 [[TMP15]] to i47
+  // CHECK-NEXT:    [[RESIZE20:%.*]] = sext i32 [[CONV]] to i47
+  // CHECK-NEXT:    [[UPSCALE21:%.*]] = shl i47 [[RESIZE20]], 15
+  // CHECK-NEXT:    [[TMP17:%.*]] = call i47 @llvm.smul.fix.i47(i47 [[RESIZE19]], i47 [[UPSCALE21]], i32 15)
+  // CHECK-NEXT:    [[RESIZE22:%.*]] = trunc i47 [[TMP17]] to i32
+  // CHECK-NEXT:    store i32 [[RESIZE22]], i32* %a, align 4
+  a = a * b;
+
+  // CHECK:         [[TMP18:%.*]] = load i32, i32* %i, align 4
+  // CHECK-NEXT:    [[TMP19:%.*]] = load i32, i32* %a, align 4
+  // CHECK-NEXT:    [[RESIZE23:%.*]] = sext i32 [[TMP18]] to i47
+  // CHECK-NEXT:    [[UPSCALE24:%.*]] = shl i47 [[RESIZE23]], 15
+  // CHECK-NEXT:    [[RESIZE25:%.*]] = sext i32 [[TMP19]] to i47
+  // CHECK-NEXT:    [[TMP20:%.*]] = call i47 @llvm.smul.fix.i47(i47 [[UPSCALE24]], i47 [[RESIZE25]], i32 15)
+  // CHECK-NEXT:    [[RESIZE26:%.*]] = trunc i47 [[TMP20]] to i32
+  // CHECK-NEXT:    store i32 [[RESIZE26]], i32* %a, align 4
+  a = i * a;
+
+  // CHECK:         [[TMP21:%.*]] = load i32, i32* %ui, align 4
+  // CHECK-NEXT:    [[TMP22:%.*]] = load i16, i16* %usa, align 2
+  // SIGNED-NEXT:   [[RESIZE27:%.*]] = zext i32 [[TMP21]] to i40
+  // SIGNED-NEXT:   [[UPSCALE28:%.*]] = shl i40 [[RESIZE27]], 8
+  // SIGNED-NEXT:   [[RESIZE29:%.*]] = zext i16 [[TMP22]] to i40
+  // SIGNED-NEXT:   [[TMP23:%.*]] = call i40 @llvm.umul.fix.i40(i40 [[UPSCALE28]], i40 [[RESIZE29]], i32 8)
+  // SIGNED-NEXT:   [[RESIZE30:%.*]] = trunc i40 [[TMP23]] to i16
+  // UNSIGNED-NEXT: [[RESIZE27:%.*]] = zext i32 [[TMP21]] to i39
+  // UNSIGNED-NEXT: [[UPSCALE28:%.*]] = shl i39 [[RESIZE27]], 7
+  // UNSIGNED-NEXT: [[RESIZE29:%.*]] = zext i16 [[TMP22]] to i39
+  // UNSIGNED-NEXT: [[TMP23:%.*]] = call i39 @llvm.umul.fix.i39(i39 [[UPSCALE28]], i39 [[RESIZE29]], i32 7)
+  // UNSIGNED-NEXT: [[RESIZE30:%.*]] = trunc i39 [[TMP23]] to i16
+  // CHECK-NEXT:    store i16 [[RESIZE30]], i16* %usa, align 2
+  usa = ui * usa;
+
+  // CHECK:         [[TMP27:%.*]] = load i32, i32* %ui, align 4
+  // CHECK-NEXT:    [[TMP28:%.*]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT:    [[RESIZE33:%.*]] = zext i32 [[TMP27]] to i40
+  // CHECK-NEXT:    [[UPSCALE34:%.*]] = shl i40 [[RESIZE33]], 7
+  // CHECK-NEXT:    [[RESIZE35:%.*]] = sext i16 [[TMP28]] to i40
+  // CHECK-NEXT:    [[TMP29:%.*]] = call i40 @llvm.smul.fix.i40(i40 [[UPSCALE34]], i40 [[RESIZE35]], i32 7)
+  // CHECK-NEXT:    [[RESIZE36:%.*]] = trunc i40 [[TMP29]] to i16
+  // CHECK-NEXT:    store i16 [[RESIZE36]], i16* %sa, align 2
+  sa = ui * sa;
+}
+
+void SaturatedMultiplication() {
+  // CHECK-LABEL: SaturatedMultiplication
+  short _Accum sa;
+  _Accum a;
+  long _Accum la;
+  unsigned short _Accum usa;
+  unsigned _Accum ua;
+  unsigned long _Accum ula;
+
+  _Sat short _Accum sa_sat;
+  _Sat _Accum a_sat;
+  _Sat long _Accum la_sat;
+  _Sat unsigned short _Accum usa_sat;
+  _Sat unsigned _Accum ua_sat;
+  _Sat unsigned long _Accum ula_sat;
+  _Sat unsigned _Fract uf_sat;
+
+  int i;
+  unsigned int ui;
+
+  // CHECK:      [[SA:%[0-9]+]] = load i16, i16* %sa, align 2
+  // CHECK-NEXT: [[SA_SAT:%[0-9]+]] = load i16, i16* %sa_sat, align 2
+  // CHECK-NEXT: [[SUM:%[0-9]+]] = call i16 @llvm.smul.fix.sat.i16(i16 [[SA]], i16 [[SA_SAT]], i32 7)
+  // CHECK-NEXT: store i16 [[SUM]], i16* %sa_sat, align 2
+  sa_sat = sa * sa_sat;
+
+  // CHECK:      [[USA:%[0-9]+]] = load i16, i16* %usa, align 2
+  // CHECK-NEXT: [[USA_SAT:%[0-9]+]] = load i16, i16* %usa_sat, align 2
+  // SIGNED-NEXT: [[SUM:%[0-9]+]] = call i16 @llvm.umul.fix.sat.i16(i16 [[USA]], i16 [[USA_SAT]], i32 8)
+  // SIGNED-NEXT: store i16 [[SUM]], i16* %usa_sat, align 2
+  // UNSIGNED-NEXT: [[USA_TRUNC:%[a-z0-9]+]] = trunc i16 [[USA]] to i15
+  // UNSIGNED-NEXT: [[USA_SAT_TRUNC:%[a-z0-9]+]] = trunc i16 [[USA_SAT]] to i15
+  // UNSIGNED-NEXT: [[SUM:%[0-9]+]] = call i15 @llvm.umul.fix.sat.i15(i15 [[USA_TRUNC]], i15 [[USA_SAT_TRUNC]], i32 7)
+  // UNSIGNED-NEXT: [[SUM_EXT:%[a-z0-9]+]] = zext i15 [[SUM]] to i16
+  // UNSIGNED-NEXT: store i16 [[SUM_EXT]], i16* %usa_sat, align 2
+  usa_sat = usa * usa_sat;
+
+  // CHECK:      [[UA:%[0-9]+]] = load i32, i32* %ua, align 4
+  // CHECK-NEXT: [[USA:%[0-9]+]] = load i16, i16* %usa_sat, align 2
+  // SIGNED-NEXT: [[USA_EXT:%[a-z0-9]+]] = zext i16 [[USA]] to i32
+  // SIGNED-NEXT: [[USA:%[a-z0-9]+]] = shl i32 [[USA_EXT]], 8
+  // SIGNED-NEXT: [[SUM:%[0-9]+]] = call i32 @llvm.umul.fix.sat.i32(i32 [[UA]], i32 [[USA]], i32 16)
+  // SIGNED-NEXT: store i32 [[SUM]], i32* %ua_sat, align 4
+  // UNSIGNED-NEXT: [[UA_TRUNC:%[a-z0-9]+]] = trunc i32 [[UA]] to i31
+  // UNSIGNED-NEXT: [[USA_EXT:%[a-z0-9]+]] = zext i16 [[USA]] to i31
+  // UNSIGNED-NEXT: [[USA:%[a-z0-9]+]] = shl i31 [[USA_EXT]], 8
+  // UNSIGNED-NEXT: [[SUM:%[0-9]+]] = call i31 @llvm.umul.fix.sat.i31(i31 [[UA_TRUNC]], i31 [[USA]], i32 15)
+  // UNSIGNED-NEXT: [[SUM_EXT:%[a-z0-9]+]] = zext i31 [[SUM]] to i32
+  // UNSIGNED-NEXT: store i32 [[SUM_EXT]], i32* %ua_sat, align 4
+  ua_sat = ua * usa_sat;
+
+  // CHECK:      [[SA_SAT:%[0-9]+]] = load i16, i16* %sa_sat, align 2
+  // CHECK-NEXT: [[I:%[0-9]+]] = load i32, i32* %i, align 4
+  // CHECK-NEXT: [[SA_SAT_EXT:%[a-z0-9]+]] = sext i16 [[SA_SAT]] to i39
+  // CHECK-NEXT: [[I_EXT:%[a-z0-9]+]] = sext i32 [[I]] to i39
+  // CHECK-NEXT: [[I:%[a-z0-9]+]] = shl i39 [[I_EXT]], 7
+  // CHECK-NEXT: [[SUM:%[0-9]+]] = call i39 @llvm.smul.fix.sat.i39(i39 [[SA_SAT_EXT]], i39 [[I]], i32 7)
+  // CHECK-NEXT: [[USE_MAX:%[0-9]+]] = icmp sgt i39 [[SUM]], 32767
+  // CHECK-NEXT: [[RES:%[a-z0-9]+]] = select i1 [[USE_MAX]], i39 32767, i39 [[SUM]]
+  // CHECK-NEXT: [[USE_MIN:%[0-9]+]] = icmp slt i39 [[RES]], -32768
+  // CHECK-NEXT: [[RES2:%[a-z0-9]+]] = select i1 [[USE_MIN]], i39 -32768, i39 [[RES]]
+  // CHECK-NEXT: [[RES3:%[a-z0-9]+]] = trunc i39 [[RES2]] to i16
+  // CHECK-NEXT: store i16 [[RES3]], i16* %sa_sat, align 2
+  sa_sat = sa_sat * i;
+
+  // CHECK:      [[SA_SAT:%[0-9]+]] = load i16, i16* %sa_sat, align 2
+  // CHECK-NEXT: [[I:%[0-9]+]] = load i32, i32* %ui, align 4
+  // CHECK-NEXT: [[SA_SAT_EXT:%[a-z0-9]+]] = sext i16 [[SA_SAT]] to i40
+  // CHECK-NEXT: [[I_EXT:%[a-z0-9]+]] = zext i32 [[I]] to i40
+  // CHECK-NEXT: [[I:%[a-z0-9]+]] = shl i40 [[I_EXT]], 7
+  // CHECK-NEXT: [[SUM:%[0-9]+]] = call i40 @llvm.smul.fix.sat.i40(i40 [[SA_SAT_EXT]], i40 [[I]], i32 7)
+  // CHECK-NEXT: [[USE_MAX:%[0-9]+]] = icmp sgt i40 [[SUM]], 32767
+  // CHECK-NEXT: [[RES:%[a-z0-9]+]] = select i1 [[USE_MAX]], i40 32767, i40 [[SUM]]
+  // CHECK-NEXT: [[USE_MIN:%[0-9]+]] = icmp slt i40 [[RES]], -32768
+  // CHECK-NEXT: [[RES2:%[a-z0-9]+]] = select i1 [[USE_MIN]], i40 -32768, i40 [[RES]]
+  // CHECK-NEXT: [[RES3:%[a-z0-9]+]] = trunc i40 [[RES2]] to i16
+  // CHECK-NEXT: store i16 [[RES3]], i16* %sa_sat, align 2
+  sa_sat = sa_sat * ui;
+
+  // CHECK:      [[UF_SAT:%[0-9]+]] = load i16, i16* %uf_sat, align 2
+  // CHECK-NEXT: [[UF_SAT2:%[0-9]+]] = load i16, i16* %uf_sat, align 2
+  // SIGNED-NEXT: [[SUM:%[0-9]+]] = call i16 @llvm.umul.fix.sat.i16(i16 [[UF_SAT]], i16 [[UF_SAT2]], i32 16)
+  // SIGNED-NEXT: store i16 [[SUM]], i16* %uf_sat, align 2
+  // UNSIGNED-NEXT: [[UF_SAT_TRUNC:%[a-z0-9]+]] = trunc i16 [[UF_SAT]] to i15
+  // UNSIGNED-NEXT: [[UF_SAT_TRUNC2:%[a-z0-9]+]] = trunc i16 [[UF_SAT2]] to i15
+  // UNSIGNED-NEXT: [[SUM:%[0-9]+]] = call i15 @llvm.umul.fix.sat.i15(i15 [[UF_SAT_TRUNC]], i15 [[UF_SAT_TRUNC2]], i32 15)
+  // UNSIGNED-NEXT: [[SUM_EXT:%[a-z0-9]+]] = zext i15 [[SUM]] to i16
+  // UNSIGNED-NEXT: store i16 [[SUM_EXT]], i16* %uf_sat, align 2
+  uf_sat = uf_sat * uf_sat;
+
+  // CHECK:      [[USA_SAT:%[0-9]+]] = load i16, i16* %usa_sat, align 2
+  // CHECK-NEXT: [[I:%[0-9]+]] = load i32, i32* %i, align 4
+  // SIGNED-NEXT: [[USA_SAT_RESIZE:%[a-z0-9]+]] = zext i16 [[USA_SAT]] to i40
+  // SIGNED-NEXT: [[I_RESIZE:%[a-z0-9]+]] = sext i32 [[I]] to i40
+  // SIGNED-NEXT: [[I_UPSCALE:%[a-z0-9]+]] = shl i40 [[I_RESIZE]], 8
+  // SIGNED-NEXT: [[SUM:%[0-9]+]] = call i40 @llvm.umul.fix.sat.i40(i40 [[USA_SAT_RESIZE]], i40 [[I_UPSCALE]], i32 8)
+  // SIGNED-NEXT: [[USE_MAX:%[0-9]+]] = icmp sgt i40 [[SUM]], 65535
+  // SIGNED-NEXT: [[RESULT:%[a-z0-9]+]] = select i1 [[USE_MAX]], i40 65535, i40 [[SUM]]
+  // SIGNED-NEXT: [[USE_MIN:%[0-9]+]] = icmp slt i40 [[RESULT]], 0
+  // SIGNED-NEXT: [[RESULT2:%[a-z0-9]+]] = select i1 [[USE_MIN]], i40 0, i40 [[RESULT]]
+  // SIGNED-NEXT: [[RESULT:%[a-z0-9]+]] = trunc i40 [[RESULT2]] to i16
+  // UNSIGNED-NEXT: [[USA_SAT_RESIZE:%[a-z0-9]+]] = zext i16 [[USA_SAT]] to i39
+  // UNSIGNED-NEXT: [[I_RESIZE:%[a-z0-9]+]] = sext i32 [[I]] to i39
+  // UNSIGNED-NEXT: [[I_UPSCALE:%[a-z0-9]+]] = shl i39 [[I_RESIZE]], 7
+  // UNSIGNED-NEXT: [[SUM:%[0-9]+]] = call i39 @llvm.umul.fix.sat.i39(i39 [[USA_SAT_RESIZE]], i39 [[I_UPSCALE]], i32 7)
+  // UNSIGNED-NEXT: [[USE_MAX:%[0-9]+]] = icmp sgt i39 [[SUM]], 32767
+  // UNSIGNED-NEXT: [[RESULT:%[a-z0-9]+]] = select i1 [[USE_MAX]], i39 32767, i39 [[SUM]]
+  // UNSIGNED-NEXT: [[USE_MIN:%[0-9]+]] = icmp slt i39 [[RESULT]], 0
+  // UNSIGNED-NEXT: [[RESULT2:%[a-z0-9]+]] = select i1 [[USE_MIN]], i39 0, i39 [[RESULT]]
+  // UNSIGNED-NEXT: [[RESULT:%[a-z0-9]+]] = trunc i39 [[RESULT2]] to i16
+  // CHECK-NEXT: store i16 [[RESULT]], i16* %usa_sat, align 2
+  usa_sat = usa_sat * i;
+}


        


More information about the cfe-commits mailing list