[Mlir-commits] [mlir] b7fe0d3 - Lower math.tan to __nv_tan[f] / __ocml_tan_f{32|64}
Goran Flegar
llvmlistbot at llvm.org
Thu Jan 12 06:27:01 PST 2023
Author: Goran Flegar
Date: 2023-01-12T15:26:11+01:00
New Revision: b7fe0d346b3049df6e3950159a5de74438d2737a
URL: https://github.com/llvm/llvm-project/commit/b7fe0d346b3049df6e3950159a5de74438d2737a
DIFF: https://github.com/llvm/llvm-project/commit/b7fe0d346b3049df6e3950159a5de74438d2737a.diff
LOG: Lower math.tan to __nv_tan[f] / __ocml_tan_f{32|64}
At present math.tan fails to lower for NVVM and ROCDL.
Differential Revision: https://reviews.llvm.org/D141505
Added:
Modified:
mlir/lib/Conversion/GPUToNVVM/LowerGpuOpsToNVVMOps.cpp
mlir/lib/Conversion/GPUToROCDL/LowerGpuOpsToROCDLOps.cpp
mlir/test/Conversion/GPUToNVVM/gpu-to-nvvm.mlir
mlir/test/Conversion/GPUToROCDL/gpu-to-rocdl.mlir
Removed:
################################################################################
diff --git a/mlir/lib/Conversion/GPUToNVVM/LowerGpuOpsToNVVMOps.cpp b/mlir/lib/Conversion/GPUToNVVM/LowerGpuOpsToNVVMOps.cpp
index 5050f6b6bf6bb..fe578f7560acc 100644
--- a/mlir/lib/Conversion/GPUToNVVM/LowerGpuOpsToNVVMOps.cpp
+++ b/mlir/lib/Conversion/GPUToNVVM/LowerGpuOpsToNVVMOps.cpp
@@ -294,6 +294,7 @@ void mlir::populateGpuToNVVMConversionPatterns(LLVMTypeConverter &converter,
"__nv_sqrt");
populateOpPatterns<math::TanhOp>(converter, patterns, "__nv_tanhf",
"__nv_tanh");
+ populateOpPatterns<math::TanOp>(converter, patterns, "__nv_tanf", "__nv_tan");
}
std::unique_ptr<OperationPass<gpu::GPUModuleOp>>
diff --git a/mlir/lib/Conversion/GPUToROCDL/LowerGpuOpsToROCDLOps.cpp b/mlir/lib/Conversion/GPUToROCDL/LowerGpuOpsToROCDLOps.cpp
index fead2891ef9dc..018288646a14f 100644
--- a/mlir/lib/Conversion/GPUToROCDL/LowerGpuOpsToROCDLOps.cpp
+++ b/mlir/lib/Conversion/GPUToROCDL/LowerGpuOpsToROCDLOps.cpp
@@ -256,6 +256,8 @@ void mlir::populateGpuToROCDLConversionPatterns(
"__ocml_sqrt_f64");
populateOpPatterns<math::TanhOp>(converter, patterns, "__ocml_tanh_f32",
"__ocml_tanh_f64");
+ populateOpPatterns<math::TanOp>(converter, patterns, "__ocml_tan_f32",
+ "__ocml_tan_f64");
}
std::unique_ptr<OperationPass<gpu::GPUModuleOp>>
diff --git a/mlir/test/Conversion/GPUToNVVM/gpu-to-nvvm.mlir b/mlir/test/Conversion/GPUToNVVM/gpu-to-nvvm.mlir
index 40035fa29efa3..4a7e6f5c6631b 100644
--- a/mlir/test/Conversion/GPUToNVVM/gpu-to-nvvm.mlir
+++ b/mlir/test/Conversion/GPUToNVVM/gpu-to-nvvm.mlir
@@ -343,6 +343,25 @@ gpu.module @test_module {
// -----
+gpu.module @test_module {
+ // CHECK: llvm.func @__nv_tanf(f32) -> f32
+ // CHECK: llvm.func @__nv_tan(f64) -> f64
+ // CHECK-LABEL: func @gpu_tan
+ func.func @gpu_tan(%arg_f16 : f16, %arg_f32 : f32, %arg_f64 : f64) -> (f16, f32, f64) {
+ %result16 = math.tan %arg_f16 : f16
+ // CHECK: llvm.fpext %{{.*}} : f16 to f32
+ // CHECK-NEXT: llvm.call @__nv_tanf(%{{.*}}) : (f32) -> f32
+ // CHECK-NEXT: llvm.fptrunc %{{.*}} : f32 to f16
+ %result32 = math.tan %arg_f32 : f32
+ // CHECK: llvm.call @__nv_tanf(%{{.*}}) : (f32) -> f32
+ %result64 = math.tan %arg_f64 : f64
+ // CHECK: llvm.call @__nv_tan(%{{.*}}) : (f64) -> f64
+ func.return %result16, %result32, %result64 : f16, f32, f64
+ }
+}
+
+// -----
+
gpu.module @test_module {
// CHECK: llvm.func @__nv_tanhf(f32) -> f32
// CHECK: llvm.func @__nv_tanh(f64) -> f64
diff --git a/mlir/test/Conversion/GPUToROCDL/gpu-to-rocdl.mlir b/mlir/test/Conversion/GPUToROCDL/gpu-to-rocdl.mlir
index 0403cdb2615d5..ccf8e47f8da0b 100644
--- a/mlir/test/Conversion/GPUToROCDL/gpu-to-rocdl.mlir
+++ b/mlir/test/Conversion/GPUToROCDL/gpu-to-rocdl.mlir
@@ -362,6 +362,21 @@ gpu.module @test_module {
// -----
+gpu.module @test_module {
+ // CHECK: llvm.func @__ocml_tan_f32(f32) -> f32
+ // CHECK: llvm.func @__ocml_tan_f64(f64) -> f64
+ // CHECK-LABEL: func @gpu_tan
+ func.func @gpu_tan(%arg_f32 : f32, %arg_f64 : f64) -> (f32, f64) {
+ %result32 = math.tan %arg_f32 : f32
+ // CHECK: llvm.call @__ocml_tan_f32(%{{.*}}) : (f32) -> f32
+ %result64 = math.tan %arg_f64 : f64
+ // CHECK: llvm.call @__ocml_tan_f64(%{{.*}}) : (f64) -> f64
+ func.return %result32, %result64 : f32, f64
+ }
+}
+
+// -----
+
gpu.module @test_module {
// CHECK: llvm.func @__ocml_tanh_f32(f32) -> f32
// CHECK: llvm.func @__ocml_tanh_f64(f64) -> f64
More information about the Mlir-commits
mailing list