[llvm-branch-commits] [llvm] 5f18e2f - Move createScalarizeMaskedMemIntrinPass to Scalar.h

Benjamin Kramer via llvm-branch-commits llvm-branch-commits at lists.llvm.org
Tue Dec 8 10:12:44 PST 2020


Author: Benjamin Kramer
Date: 2020-12-08T19:08:09+01:00
New Revision: 5f18e2f31ecbda5ace100cd2925ef7a0cafe3c3b

URL: https://github.com/llvm/llvm-project/commit/5f18e2f31ecbda5ace100cd2925ef7a0cafe3c3b
DIFF: https://github.com/llvm/llvm-project/commit/5f18e2f31ecbda5ace100cd2925ef7a0cafe3c3b.diff

LOG: Move createScalarizeMaskedMemIntrinPass to Scalar.h

Added: 
    

Modified: 
    llvm/include/llvm/CodeGen/Passes.h
    llvm/include/llvm/Transforms/Scalar.h
    llvm/lib/Transforms/Scalar/ScalarizeMaskedMemIntrin.cpp

Removed: 
    


################################################################################
diff  --git a/llvm/include/llvm/CodeGen/Passes.h b/llvm/include/llvm/CodeGen/Passes.h
index e3aa32bffb71..a74334e6200c 100644
--- a/llvm/include/llvm/CodeGen/Passes.h
+++ b/llvm/include/llvm/CodeGen/Passes.h
@@ -76,10 +76,6 @@ namespace llvm {
   /// matching during instruction selection.
   FunctionPass *createCodeGenPreparePass();
 
-  /// createScalarizeMaskedMemIntrinPass - Replace masked load, store, gather
-  /// and scatter intrinsics with scalar code when target doesn't support them.
-  FunctionPass *createScalarizeMaskedMemIntrinPass();
-
   /// AtomicExpandID -- Lowers atomic operations in terms of either cmpxchg
   /// load-linked/store-conditional loops.
   extern char &AtomicExpandID;

diff  --git a/llvm/include/llvm/Transforms/Scalar.h b/llvm/include/llvm/Transforms/Scalar.h
index 68f678432471..2cebd89fdddb 100644
--- a/llvm/include/llvm/Transforms/Scalar.h
+++ b/llvm/include/llvm/Transforms/Scalar.h
@@ -545,6 +545,14 @@ Pass *createWarnMissedTransformationsPass();
 // instruction in a function.
 //
 FunctionPass *createInstSimplifyLegacyPass();
+
+
+//===----------------------------------------------------------------------===//
+//
+// createScalarizeMaskedMemIntrinPass - Replace masked load, store, gather
+// and scatter intrinsics with scalar code when target doesn't support them.
+//
+FunctionPass *createScalarizeMaskedMemIntrinPass();
 } // End llvm namespace
 
 #endif

diff  --git a/llvm/lib/Transforms/Scalar/ScalarizeMaskedMemIntrin.cpp b/llvm/lib/Transforms/Scalar/ScalarizeMaskedMemIntrin.cpp
index fb5ad3fb1c88..98d85f4f5844 100644
--- a/llvm/lib/Transforms/Scalar/ScalarizeMaskedMemIntrin.cpp
+++ b/llvm/lib/Transforms/Scalar/ScalarizeMaskedMemIntrin.cpp
@@ -31,6 +31,7 @@
 #include "llvm/InitializePasses.h"
 #include "llvm/Pass.h"
 #include "llvm/Support/Casting.h"
+#include "llvm/Transforms/Scalar.h"
 #include <algorithm>
 #include <cassert>
 


        


More information about the llvm-branch-commits mailing list