[llvm] 5fa9d41 - [Support/Parallel] Add a special case for 0/1 items to llvm::parallel_for_each.
Chris Lattner via llvm-commits
llvm-commits at lists.llvm.org
Mon May 3 10:08:09 PDT 2021
Author: Chris Lattner
Date: 2021-05-03T10:08:00-07:00
New Revision: 5fa9d4163421304255d12498fd672e79893422a4
URL: https://github.com/llvm/llvm-project/commit/5fa9d4163421304255d12498fd672e79893422a4
DIFF: https://github.com/llvm/llvm-project/commit/5fa9d4163421304255d12498fd672e79893422a4.diff
LOG: [Support/Parallel] Add a special case for 0/1 items to llvm::parallel_for_each.
This avoids the non-trivial overhead of creating a TaskGroup in these degenerate
cases, but also exposes parallelism. It turns out that the default executor
underlying TaskGroup prevents recursive parallelism - so an instance of a task
group being alive will make nested ones become serial.
This is a big issue in MLIR in some dialects, if they have a single instance of
an outer op (e.g. a firrtl.circuit) that has many parallel ops within it (e.g.
a firrtl.module). This patch side-steps the problem by avoiding creating the
TaskGroup in the unneeded case. See this issue for more details:
https://github.com/llvm/circt/issues/993
Note that this isn't a really great solution for the general case of nested
parallelism. A redesign of the TaskGroup stuff would be better, but would be
a much more invasive change.
Differential Revision: https://reviews.llvm.org/D101699
Added:
Modified:
llvm/include/llvm/Support/Parallel.h
Removed:
################################################################################
diff --git a/llvm/include/llvm/Support/Parallel.h b/llvm/include/llvm/Support/Parallel.h
index d2f006773836..28d171d45256 100644
--- a/llvm/include/llvm/Support/Parallel.h
+++ b/llvm/include/llvm/Support/Parallel.h
@@ -129,9 +129,20 @@ enum { MaxTasksPerGroup = 1024 };
template <class IterTy, class FuncTy>
void parallel_for_each(IterTy Begin, IterTy End, FuncTy Fn) {
+ // If we have zero or one items, then do not incur the overhead of spinning up
+ // a task group. They are surprisingly expensive, and because they do not
+ // support nested parallelism, a single entry task group can block parallel
+ // execution underneath them.
+ auto NumItems = std::distance(Begin, End);
+ if (NumItems <= 1) {
+ if (NumItems)
+ Fn(*Begin);
+ return;
+ }
+
// Limit the number of tasks to MaxTasksPerGroup to limit job scheduling
// overhead on large inputs.
- ptr
diff _t TaskSize = std::distance(Begin, End) / MaxTasksPerGroup;
+ ptr
diff _t TaskSize = NumItems / MaxTasksPerGroup;
if (TaskSize == 0)
TaskSize = 1;
@@ -145,9 +156,20 @@ void parallel_for_each(IterTy Begin, IterTy End, FuncTy Fn) {
template <class IndexTy, class FuncTy>
void parallel_for_each_n(IndexTy Begin, IndexTy End, FuncTy Fn) {
+ // If we have zero or one items, then do not incur the overhead of spinning up
+ // a task group. They are surprisingly expensive, and because they do not
+ // support nested parallelism, a single entry task group can block parallel
+ // execution underneath them.
+ auto NumItems = End - Begin;
+ if (NumItems <= 1) {
+ if (NumItems)
+ Fn(Begin);
+ return;
+ }
+
// Limit the number of tasks to MaxTasksPerGroup to limit job scheduling
// overhead on large inputs.
- ptr
diff _t TaskSize = (End - Begin) / MaxTasksPerGroup;
+ ptr
diff _t TaskSize = NumItems / MaxTasksPerGroup;
if (TaskSize == 0)
TaskSize = 1;
More information about the llvm-commits
mailing list