[Mlir-commits] [mlir] [mlir][Tensor] Fold destination-style ops into `tensor.unpack` operation. (PR #71468)
llvmlistbot at llvm.org
llvmlistbot at llvm.org
Mon Nov 6 16:46:03 PST 2023
https://github.com/MaheshRavishankar created https://github.com/llvm/llvm-project/pull/71468
The destination operand of the `tensor.unpack` operation is only needed to carry shape information. So if the producer of the destination operand implements the `DestinationStyleOpInterface`, then fold it into the `tensor.unpack` operation by replacing the destination operand with the destination for the source.
>From db1e68b908e76444fe2d38427d214f03a9b77bb0 Mon Sep 17 00:00:00 2001
From: MaheshRavishankar <mahesh at nod-labs.com>
Date: Mon, 6 Nov 2023 16:35:15 -0800
Subject: [PATCH] [mlir][Tensor] Fold destination-style ops into
`tensor.unpack` operation.
The destination operand of the `tensor.unpack` operation is only
needed to carry shape information. So if the producer of the
destination operand implements the `DestinationStyleOpInterface`, then
fold it into the `tensor.unpack` operation by replacing the
destination operand with the destination for the source.
---
mlir/lib/Dialect/Tensor/IR/TensorOps.cpp | 28 +++++++++++++++-------
mlir/test/Dialect/Tensor/canonicalize.mlir | 16 +++++++++++++
2 files changed, 35 insertions(+), 9 deletions(-)
diff --git a/mlir/lib/Dialect/Tensor/IR/TensorOps.cpp b/mlir/lib/Dialect/Tensor/IR/TensorOps.cpp
index f719cfed6b6dd30..79ea99a192d0e31 100644
--- a/mlir/lib/Dialect/Tensor/IR/TensorOps.cpp
+++ b/mlir/lib/Dialect/Tensor/IR/TensorOps.cpp
@@ -3925,15 +3925,25 @@ UnPackOp UnPackOp::createTransposedClone(OpBuilder &b, Location loc,
/// pack(unpack(x)) -> x
LogicalResult UnPackOp::canonicalize(UnPackOp unPackOp,
PatternRewriter &rewriter) {
- PackOp packOp = unPackOp.getSource().getDefiningOp<tensor::PackOp>();
- if (!packOp || packOp.getDestType() != unPackOp.getSourceType())
- return failure();
- if (packOp.getPaddingValue() ||
- !hasSameInnerOuterAttribute(packOp, unPackOp) ||
- !haveSameTiles(packOp, unPackOp))
- return failure();
- rewriter.replaceOp(unPackOp, packOp.getSource());
- return success();
+ if (PackOp packOp = unPackOp.getSource().getDefiningOp<tensor::PackOp>()) {
+ if (packOp.getDestType() != unPackOp.getSourceType())
+ return failure();
+ if (packOp.getPaddingValue() ||
+ !hasSameInnerOuterAttribute(packOp, unPackOp) ||
+ !haveSameTiles(packOp, unPackOp))
+ return failure();
+ rewriter.replaceOp(unPackOp, packOp.getSource());
+ return success();
+ }
+ if (DestinationStyleOpInterface dstStyleOp =
+ unPackOp.getDest().getDefiningOp<DestinationStyleOpInterface>()) {
+ OpResult destValue = unPackOp.getDest().cast<OpResult>();
+ Value newDest = dstStyleOp.getDpsInits()[destValue.getResultNumber()];
+ rewriter.updateRootInPlace(
+ unPackOp, [&]() { unPackOp.setDpsInitOperand(0, newDest); });
+ return success();
+ }
+ return failure();
}
bool UnPackOp::isLikeUnPad() {
diff --git a/mlir/test/Dialect/Tensor/canonicalize.mlir b/mlir/test/Dialect/Tensor/canonicalize.mlir
index c40c9efeb152ac6..b7b34a63640dbef 100644
--- a/mlir/test/Dialect/Tensor/canonicalize.mlir
+++ b/mlir/test/Dialect/Tensor/canonicalize.mlir
@@ -1861,3 +1861,19 @@ func.func @invalid_empty_negative_size() -> (tensor<4x5x?xf32>) {
%1 = tensor.empty(%0) : tensor<4x5x?xf32>
return %1 : tensor<4x5x?xf32>
}
+
+// -----
+
+// Fold DstStyleOp -> tensor.unpack operations.
+func.func @fold_dst_style_ops_into_unpack(%arg0 : tensor<?x?x16x64xf32>, %init : tensor<?x?xf32>) -> tensor<?x?xf32> {
+ %cst = arith.constant 0.0 : f32
+ %fill = linalg.fill ins(%cst : f32) outs(%init : tensor<?x?xf32>) -> tensor<?x?xf32>
+ %unpack = tensor.unpack %arg0 inner_dims_pos = [0, 1] inner_tiles = [16, 64] into %fill : tensor<?x?x16x64xf32> -> tensor<?x?xf32>
+ return %unpack : tensor<?x?xf32>
+}
+// CHECK-LABEL: func @fold_dst_style_ops_into_unpack
+// CHECK-SAME: %[[ARG0:.+]]: tensor<?x?x16x64xf32>
+// CHECK-SAME: %[[ARG1:.+]]: tensor<?x?xf32>
+// CHECK: %[[UNPACK:.+]] = tensor.unpack %[[ARG0]]
+// CHECK-SAME: into %[[ARG1]]
+// CHECK: return %[[UNPACK]]
More information about the Mlir-commits
mailing list