[Mlir-commits] [mlir] [mlir][Linalg] Add speculation for LinalgStructuredOps (PR #108032)
Kunwar Grover
llvmlistbot at llvm.org
Tue Sep 10 09:45:28 PDT 2024
https://github.com/Groverkss updated https://github.com/llvm/llvm-project/pull/108032
>From 658f42a4a9f6937baf2dc8f49e068162c5478f53 Mon Sep 17 00:00:00 2001
From: Kunwar Grover <groverkss at gmail.com>
Date: Tue, 10 Sep 2024 13:51:07 +0100
Subject: [PATCH 1/5] Add speculation for linalg structured ops
---
.../Dialect/Linalg/IR/LinalgStructuredOps.td | 1 +
mlir/lib/Dialect/Linalg/IR/LinalgOps.cpp | 34 +++++++++++++++++++
.../mlir-linalg-ods-yaml-gen.cpp | 5 ++-
3 files changed, 39 insertions(+), 1 deletion(-)
diff --git a/mlir/include/mlir/Dialect/Linalg/IR/LinalgStructuredOps.td b/mlir/include/mlir/Dialect/Linalg/IR/LinalgStructuredOps.td
index ac61117c3d6e36..31f29139247267 100644
--- a/mlir/include/mlir/Dialect/Linalg/IR/LinalgStructuredOps.td
+++ b/mlir/include/mlir/Dialect/Linalg/IR/LinalgStructuredOps.td
@@ -29,6 +29,7 @@ class LinalgStructuredBase_Op<string mnemonic, list<Trait> props>
: Op<Linalg_Dialect, mnemonic, !listconcat([
SingleBlockImplicitTerminator<"YieldOp">,
DeclareOpInterfaceMethods<MemoryEffectsOpInterface>,
+ DeclareOpInterfaceMethods<ConditionallySpeculatable>,
DestinationStyleOpInterface,
LinalgStructuredInterface,
ReifyRankedShapedTypeOpInterface], props)> {
diff --git a/mlir/lib/Dialect/Linalg/IR/LinalgOps.cpp b/mlir/lib/Dialect/Linalg/IR/LinalgOps.cpp
index 76df3ecf2d2bd4..790e2ac50d9e39 100644
--- a/mlir/lib/Dialect/Linalg/IR/LinalgOps.cpp
+++ b/mlir/lib/Dialect/Linalg/IR/LinalgOps.cpp
@@ -34,6 +34,7 @@
#include "mlir/IR/OperationSupport.h"
#include "mlir/IR/PatternMatch.h"
#include "mlir/Interfaces/InferTypeOpInterface.h"
+#include "mlir/Interfaces/SideEffectInterfaces.h"
#include "llvm/ADT/DenseMap.h"
#include "llvm/ADT/SmallSet.h"
@@ -1202,6 +1203,23 @@ void GenericOp::getEffects(
getGenericEffectsImpl(effects, cast<LinalgOp>(getOperation()));
}
+static Speculation::Speculatability
+getGenericSpeculatabilityImpl(LinalgOp linalgOp) {
+ // Operands with value semantics are speculatable, while operands with memory
+ // semantics are not.
+ for (Value operand : linalgOp->getOperands()) {
+ if (isa<MemRefType>(operand.getType())) {
+ return Speculation::NotSpeculatable;
+ }
+ }
+ // The body of the op can still have speculation in it's region.
+ return Speculation::RecursivelySpeculatable;
+}
+
+Speculation::Speculatability GenericOp::getSpeculatability() {
+ return getGenericSpeculatabilityImpl(cast<LinalgOp>(getOperation()));
+}
+
LogicalResult GenericOp::verify() { return success(); }
namespace {
@@ -1553,6 +1571,10 @@ void MapOp::getEffects(
getGenericEffectsImpl(effects, cast<LinalgOp>(getOperation()));
}
+Speculation::Speculatability MapOp::getSpeculatability() {
+ return getGenericSpeculatabilityImpl(cast<LinalgOp>(getOperation()));
+}
+
//===----------------------------------------------------------------------===//
// ReduceOp
//===----------------------------------------------------------------------===//
@@ -1621,6 +1643,10 @@ void ReduceOp::getEffects(
getGenericEffectsImpl(effects, cast<LinalgOp>(getOperation()));
}
+Speculation::Speculatability ReduceOp::getSpeculatability() {
+ return getGenericSpeculatabilityImpl(cast<LinalgOp>(getOperation()));
+}
+
static ParseResult parseDenseI64ArrayAttr(OpAsmParser &parser,
NamedAttrList &attributes,
StringRef attributeName) {
@@ -1906,6 +1932,10 @@ void TransposeOp::getEffects(
getGenericEffectsImpl(effects, cast<LinalgOp>(getOperation()));
}
+Speculation::Speculatability TransposeOp::getSpeculatability() {
+ return getGenericSpeculatabilityImpl(cast<LinalgOp>(getOperation()));
+}
+
LogicalResult TransposeOp::fold(FoldAdaptor adaptor,
SmallVectorImpl<OpFoldResult> &result) {
// Only the tensor type is supported.
@@ -2134,6 +2164,10 @@ void BroadcastOp::getEffects(
getGenericEffectsImpl(effects, cast<LinalgOp>(getOperation()));
}
+Speculation::Speculatability BroadcastOp::getSpeculatability() {
+ return getGenericSpeculatabilityImpl(cast<LinalgOp>(getOperation()));
+}
+
void BroadcastOp::getCanonicalizationPatterns(RewritePatternSet &results,
MLIRContext *context) {
results.add<EraseIdentityLinalgOp<BroadcastOp>>(context);
diff --git a/mlir/tools/mlir-linalg-ods-gen/mlir-linalg-ods-yaml-gen.cpp b/mlir/tools/mlir-linalg-ods-gen/mlir-linalg-ods-yaml-gen.cpp
index 7311cdd39d0755..e5ce845b94b39b 100644
--- a/mlir/tools/mlir-linalg-ods-gen/mlir-linalg-ods-yaml-gen.cpp
+++ b/mlir/tools/mlir-linalg-ods-gen/mlir-linalg-ods-yaml-gen.cpp
@@ -656,7 +656,7 @@ ArrayAttr {0}::getIndexingMaps() {{
}
)FMT";
-// Implementations of fold and getEffects.
+// Implementations of fold, getEffects and getSpeculatability.
// Parameters:
// {0}: Class name
const char structuredOpFoldersFormat[] = R"FMT(
@@ -669,6 +669,9 @@ void {0}::getEffects(SmallVectorImpl<
if (hasPureTensorSemantics()) return;
getGenericEffectsImpl(effects, cast<LinalgOp>(getOperation()));
}
+Speculation::Speculatability {0}::getSpeculatability() {{
+ return getGenericSpeculatabilityImpl(cast<LinalgOp>(getOperation()));
+}
)FMT";
// Implementation of parse/print.
>From 63bc3dfd0a1993da2581d1af6f267321b1f9dcc9 Mon Sep 17 00:00:00 2001
From: Kunwar Grover <groverkss at gmail.com>
Date: Tue, 10 Sep 2024 15:14:05 +0100
Subject: [PATCH 2/5] Use hasPureTensorSemantics
---
mlir/lib/Dialect/Linalg/IR/LinalgOps.cpp | 7 ++-----
1 file changed, 2 insertions(+), 5 deletions(-)
diff --git a/mlir/lib/Dialect/Linalg/IR/LinalgOps.cpp b/mlir/lib/Dialect/Linalg/IR/LinalgOps.cpp
index 790e2ac50d9e39..e2af2e1caad811 100644
--- a/mlir/lib/Dialect/Linalg/IR/LinalgOps.cpp
+++ b/mlir/lib/Dialect/Linalg/IR/LinalgOps.cpp
@@ -1207,11 +1207,8 @@ static Speculation::Speculatability
getGenericSpeculatabilityImpl(LinalgOp linalgOp) {
// Operands with value semantics are speculatable, while operands with memory
// semantics are not.
- for (Value operand : linalgOp->getOperands()) {
- if (isa<MemRefType>(operand.getType())) {
- return Speculation::NotSpeculatable;
- }
- }
+ if (!linalgOp.hasPureTensorSemantics())
+ return Speculation::NotSpeculatable;
// The body of the op can still have speculation in it's region.
return Speculation::RecursivelySpeculatable;
}
>From a0b97b000cd380b48ab21dfa78c664f4020192bb Mon Sep 17 00:00:00 2001
From: Kunwar Grover <groverkss at gmail.com>
Date: Tue, 10 Sep 2024 15:15:11 +0100
Subject: [PATCH 3/5] Add test
---
.../loop-invariant-code-motion.mlir | 45 +++++++++++++++++++
1 file changed, 45 insertions(+)
diff --git a/mlir/test/Transforms/loop-invariant-code-motion.mlir b/mlir/test/Transforms/loop-invariant-code-motion.mlir
index 47a49465e8a7cd..dbcac818a728b5 100644
--- a/mlir/test/Transforms/loop-invariant-code-motion.mlir
+++ b/mlir/test/Transforms/loop-invariant-code-motion.mlir
@@ -1118,3 +1118,48 @@ func.func @hoist_from_scf_while(%arg0: i32, %arg1: i32) -> i32 {
}
return %0 : i32
}
+
+// -----
+
+#trait = {
+ indexing_maps = [
+ affine_map<(m, n, k) -> (m, k)>,
+ affine_map<(m, n, k) -> (k, n)>,
+ affine_map<(m, n, k) -> (m, n)>
+ ],
+ iterator_types = ["parallel", "parallel", "reduction"]
+}
+
+// CHECK-LABEL: func @hoist_linalg_ops
+// CHECK: linalg.generic
+// CHECK: scf.for
+// CHECK-NOT: linalg.generic
+// CHECK: tensor.insert_slice
+// CHECK: scf.yield
+func.func @hoist_linalg_ops(%a : tensor<128x128xf32>,
+ %b : tensor<128x128xf32>,
+ %c: tensor<128x128xf32>,
+ %lb : index,
+ %ub : index,
+ %step : index,
+ %output : tensor<?x128xf32>) -> tensor<?x128xf32> {
+ %final =
+ scf.for %i = %lb to %ub step %step iter_args(%acc = %output)
+ -> tensor<?x128xf32> {
+ %compute = linalg.generic #trait
+ ins(%a, %b : tensor<128x128xf32>, tensor<128x128xf32>)
+ outs(%c : tensor<128x128xf32>) {
+ ^bb0(%in : f32, %in2 : f32, %in3 : f32):
+ %mul = arith.mulf %in, %in2 : f32
+ %add = arith.addf %mul, %in3 : f32
+ linalg.yield %in3 : f32
+ } -> tensor<128x128xf32>
+
+ %newacc = tensor.insert_slice %compute into
+ %output[%i, 0][128, 128][1, 1]
+ : tensor<128x128xf32> into tensor<?x128xf32>
+ scf.yield %newacc : tensor<?x128xf32>
+ }
+
+ func.return %final : tensor<?x128xf32>
+}
>From aabaa72036c6888abf45ea3bc731d8eb104187e9 Mon Sep 17 00:00:00 2001
From: Kunwar Grover <groverkss at gmail.com>
Date: Tue, 10 Sep 2024 17:44:08 +0100
Subject: [PATCH 4/5] add test for non speculatable
---
.../loop-invariant-code-motion.mlir | 91 +++++++++++++++++++
1 file changed, 91 insertions(+)
diff --git a/mlir/test/Transforms/loop-invariant-code-motion.mlir b/mlir/test/Transforms/loop-invariant-code-motion.mlir
index dbcac818a728b5..3a86d3077c6369 100644
--- a/mlir/test/Transforms/loop-invariant-code-motion.mlir
+++ b/mlir/test/Transforms/loop-invariant-code-motion.mlir
@@ -1163,3 +1163,94 @@ func.func @hoist_linalg_ops(%a : tensor<128x128xf32>,
func.return %final : tensor<?x128xf32>
}
+
+// -----
+
+#trait = {
+ indexing_maps = [
+ affine_map<(m, n, k) -> (m, k)>,
+ affine_map<(m, n, k) -> (k, n)>,
+ affine_map<(m, n, k) -> (m, n)>
+ ],
+ iterator_types = ["parallel", "parallel", "reduction"]
+}
+
+// CHECK-LABEL: func @hoist_linalg_ops
+// CHECK: linalg.generic
+// CHECK: scf.for
+// CHECK-NOT: linalg.generic
+// CHECK: tensor.insert_slice
+// CHECK: scf.yield
+func.func @hoist_linalg_ops_nonspeculatable(%a : tensor<128x128xf32>,
+ %b : tensor<128x128xf32>,
+ %c: tensor<128x128xf32>,
+ %lb : index,
+ %ub : index,
+ %step : index,
+ %output : tensor<?x128xf32>) -> tensor<?x128xf32> {
+ %final =
+ scf.for %i = %lb to %ub step %step iter_args(%acc = %output)
+ -> tensor<?x128xf32> {
+ %compute = linalg.generic #trait
+ ins(%a, %b : tensor<128x128xf32>, tensor<128x128xf32>)
+ outs(%c : tensor<128x128xf32>) {
+ ^bb0(%in : f32, %in2 : f32, %in3 : f32):
+ %mul = arith.mulf %in, %in2 : f32
+ %add = arith.divf %mul, %in3 : f32
+ linalg.yield %in3 : f32
+ } -> tensor<128x128xf32>
+
+ %newacc = tensor.insert_slice %compute into
+ %output[%i, 0][128, 128][1, 1]
+ : tensor<128x128xf32> into tensor<?x128xf32>
+ scf.yield %newacc : tensor<?x128xf32>
+ }
+
+ func.return %final : tensor<?x128xf32>
+}
+
+// -----
+
+#trait = {
+ indexing_maps = [
+ affine_map<(m, n, k) -> (m, k)>,
+ affine_map<(m, n, k) -> (k, n)>,
+ affine_map<(m, n, k) -> (m, n)>
+ ],
+ iterator_types = ["parallel", "parallel", "reduction"]
+}
+
+// CHECK-LABEL: func @hoist_linalg_ops_div_by_zero
+// CHECK-NOT: linalg.generic
+// CHECK: scf.for
+// CHECK: linalg.generic
+// CHECK: tensor.insert_slice
+// CHECK: scf.yield
+func.func @hoist_linalg_ops_div_by_zero(%a : tensor<128x128xi32>,
+ %b : tensor<128x128xi32>,
+ %c: tensor<128x128xi32>,
+ %lb : index,
+ %ub : index,
+ %step : index,
+ %output : tensor<?x128xi32>) -> tensor<?x128xi32> {
+ %cst0 = arith.constant 0 : i32
+ %final =
+ scf.for %i = %lb to %ub step %step iter_args(%acc = %output)
+ -> tensor<?x128xi32> {
+ %compute = linalg.generic #trait
+ ins(%a, %b : tensor<128x128xi32>, tensor<128x128xi32>)
+ outs(%c : tensor<128x128xi32>) {
+ ^bb0(%in : i32, %in2 : i32, %in3 : i32):
+ %div = arith.divui %in, %in2 : i32
+ %add = arith.addi %div, %in3 : i32
+ linalg.yield %in3 : i32
+ } -> tensor<128x128xi32>
+
+ %newacc = tensor.insert_slice %compute into
+ %output[%i, 0][128, 128][1, 1]
+ : tensor<128x128xi32> into tensor<?x128xi32>
+ scf.yield %newacc : tensor<?x128xi32>
+ }
+
+ func.return %final : tensor<?x128xi32>
+}
>From 421a8775f1200fc2bd41e88fdcfd30d2d22aea35 Mon Sep 17 00:00:00 2001
From: Kunwar Grover <groverkss at gmail.com>
Date: Tue, 10 Sep 2024 17:45:09 +0100
Subject: [PATCH 5/5] Address comments
---
mlir/lib/Dialect/Linalg/IR/LinalgOps.cpp | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/mlir/lib/Dialect/Linalg/IR/LinalgOps.cpp b/mlir/lib/Dialect/Linalg/IR/LinalgOps.cpp
index e2af2e1caad811..630985d76a0ebf 100644
--- a/mlir/lib/Dialect/Linalg/IR/LinalgOps.cpp
+++ b/mlir/lib/Dialect/Linalg/IR/LinalgOps.cpp
@@ -1209,7 +1209,7 @@ getGenericSpeculatabilityImpl(LinalgOp linalgOp) {
// semantics are not.
if (!linalgOp.hasPureTensorSemantics())
return Speculation::NotSpeculatable;
- // The body of the op can still have speculation in it's region.
+ // The body of the op can still have speculation in its region.
return Speculation::RecursivelySpeculatable;
}
More information about the Mlir-commits
mailing list