[llvm] r255444 - Add a C++11 ThreadPool implementation in LLVM
Vedant Kumar via llvm-commits
llvm-commits at lists.llvm.org
Fri Dec 18 07:38:30 PST 2015
Hi Mehdi,
I have some inline comments below --
> On Dec 12, 2015, at 2:55 PM, Mehdi AMINI via llvm-commits <llvm-commits at lists.llvm.org> wrote:
>
> Author: mehdi_amini
> Date: Sat Dec 12 16:55:25 2015
> New Revision: 255444
>
> URL: http://llvm.org/viewvc/llvm-project?rev=255444&view=rev
> Log:
> Add a C++11 ThreadPool implementation in LLVM
>
> This is a very simple implementation of a thread pool using C++11
> thread. It accepts any std::function<void()> for asynchronous
> execution. Individual task can be synchronize using the returned
> future, or the client can block on the full queue completion.
>
> In case LLVM is configured with Threading disabled, it falls back
> to sequential execution using std::async with launch:deferred.
>
> This is intended to support parallelism for ThinLTO processing in
> linker plugin, but is generic enough for any other uses.
>
> Differential Revision: http://reviews.llvm.org/D15464
>
> From: Mehdi Amini <mehdi.amini at apple.com>
>
> Added:
> llvm/trunk/include/llvm/Support/ThreadPool.h
> llvm/trunk/lib/Support/ThreadPool.cpp
> llvm/trunk/unittests/Support/ThreadPool.cpp
> Modified:
> llvm/trunk/include/llvm/Support/thread.h
> llvm/trunk/lib/Support/CMakeLists.txt
> llvm/trunk/unittests/Support/CMakeLists.txt
>
> Added: llvm/trunk/include/llvm/Support/ThreadPool.h
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/include/llvm/Support/ThreadPool.h?rev=255444&view=auto
> ==============================================================================
> --- llvm/trunk/include/llvm/Support/ThreadPool.h (added)
> +++ llvm/trunk/include/llvm/Support/ThreadPool.h Sat Dec 12 16:55:25 2015
> @@ -0,0 +1,95 @@
> +//===-- llvm/Support/ThreadPool.h - A ThreadPool implementation -*- C++ -*-===//
> +//
> +// The LLVM Compiler Infrastructure
> +//
> +// This file is distributed under the University of Illinois Open Source
> +// License. See LICENSE.TXT for details.
> +//
> +//===----------------------------------------------------------------------===//
> +//
> +// This file defines a crude C++11 based thread pool.
> +//
> +//===----------------------------------------------------------------------===//
> +
> +#ifndef LLVM_SUPPORT_THREAD_POOL_H
> +#define LLVM_SUPPORT_THREAD_POOL_H
> +
> +#include "llvm/Support/thread.h"
> +
> +#include <condition_variable>
> +#include <functional>
> +#include <future>
> +#include <memory>
> +#include <mutex>
> +#include <queue>
> +#include <utility>
> +
> +namespace llvm {
> +
> +/// A ThreadPool for asynchronous parallel execution on a defined number of
> +/// threads.
> +///
> +/// The pool keeps a vector of threads alive, waiting on a condition variable
> +/// for some work to become available.
> +class ThreadPool {
> +public:
> + using TaskTy = std::function<void()>;
> +
> + /// Construct a pool with the number of core available on the system (or
> + /// whatever the value returned by std::thread::hardware_concurrency() is).
> + ThreadPool();
> +
> + /// Construct a pool of \p ThreadCount threads
> + ThreadPool(unsigned ThreadCount);
> +
> + /// Blocking destructor: the pool will wait for all the threads to complete.
> + ~ThreadPool();
> +
> + /// Asynchronous submission of a task to the pool. The returned future can be
> + /// used to wait for the task to finish and is *non-blocking* on destruction.
> + template <typename Function, typename... Args>
> + inline std::shared_future<void> async(Function &&F, Args &&... ArgList) {
> + auto Task =
> + std::bind(std::forward<Function>(F), std::forward<Args...>(ArgList...));
> + return asyncImpl(Task);
> + }
> +
> + /// Asynchronous submission of a task to the pool. The returned future can be
> + /// used to wait for the task to finish and is *non-blocking* on destruction.
> + template <typename Function>
> + inline std::shared_future<void> async(Function &&F) {
> + return asyncImpl(F);
> + }
> +
> + /// Blocking wait for all the threads to complete and the queue to be empty.
> + /// It is an error to try to add new tasks while blocking on this call.
> + void wait();
> +
> +private:
> + /// Asynchronous submission of a task to the pool. The returned future can be
> + /// used to wait for the task to finish and is *non-blocking* on destruction.
> + std::shared_future<void> asyncImpl(TaskTy f);
> +
> + /// Threads in flight
> + std::vector<llvm::thread> Threads;
> +
> + /// Tasks waiting for execution in the pool.
> + std::queue<std::packaged_task<void()>> Tasks;
> +
> + /// Locking and signaling for accessing the Tasks queue.
> + std::mutex QueueLock;
> + std::condition_variable QueueCondition;
> +
> + /// Locking and signaling for job completion
> + std::mutex CompletionLock;
> + std::condition_variable CompletionCondition;
> +
> + /// Keep track of the number of thread actually busy
> + std::atomic<unsigned> ActiveThreads;
> +
> + /// Signal for the destruction of the pool, asking thread to exit.
> + bool EnableFlag;
> +};
> +}
> +
> +#endif // LLVM_SUPPORT_THREAD_POOL_H
>
> Modified: llvm/trunk/include/llvm/Support/thread.h
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/include/llvm/Support/thread.h?rev=255444&r1=255443&r2=255444&view=diff
> ==============================================================================
> --- llvm/trunk/include/llvm/Support/thread.h (original)
> +++ llvm/trunk/include/llvm/Support/thread.h Sat Dec 12 16:55:25 2015
> @@ -43,6 +43,8 @@ typedef std::thread thread;
>
> #else // !LLVM_ENABLE_THREADS
>
> +#include <utility>
> +
> namespace llvm {
>
> struct thread {
>
> Modified: llvm/trunk/lib/Support/CMakeLists.txt
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/Support/CMakeLists.txt?rev=255444&r1=255443&r2=255444&view=diff
> ==============================================================================
> --- llvm/trunk/lib/Support/CMakeLists.txt (original)
> +++ llvm/trunk/lib/Support/CMakeLists.txt Sat Dec 12 16:55:25 2015
> @@ -89,6 +89,7 @@ add_llvm_library(LLVMSupport
> StringRef.cpp
> SystemUtils.cpp
> TargetParser.cpp
> + ThreadPool.cpp
> Timer.cpp
> ToolOutputFile.cpp
> Triple.cpp
>
> Added: llvm/trunk/lib/Support/ThreadPool.cpp
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/Support/ThreadPool.cpp?rev=255444&view=auto
> ==============================================================================
> --- llvm/trunk/lib/Support/ThreadPool.cpp (added)
> +++ llvm/trunk/lib/Support/ThreadPool.cpp Sat Dec 12 16:55:25 2015
> @@ -0,0 +1,143 @@
> +//==-- llvm/Support/ThreadPool.cpp - A ThreadPool implementation -*- C++ -*-==//
> +//
> +// The LLVM Compiler Infrastructure
> +//
> +// This file is distributed under the University of Illinois Open Source
> +// License. See LICENSE.TXT for details.
> +//
> +//===----------------------------------------------------------------------===//
> +//
> +// This file implements a crude C++11 based thread pool.
> +//
> +//===----------------------------------------------------------------------===//
> +
> +#include "llvm/Support/ThreadPool.h"
> +
> +#include "llvm/Config/llvm-config.h"
> +#include "llvm/Support/raw_ostream.h"
> +
> +using namespace llvm;
> +
> +#if LLVM_ENABLE_THREADS
> +
> +// Default to std::thread::hardware_concurrency
> +ThreadPool::ThreadPool() : ThreadPool(std::thread::hardware_concurrency()) {}
> +
> +ThreadPool::ThreadPool(unsigned ThreadCount)
> + : ActiveThreads(0), EnableFlag(true) {
> + // Create ThreadCount threads that will loop forever, wait on QueueCondition
> + // for tasks to be queued or the Pool to be destroyed.
> + Threads.reserve(ThreadCount);
> + for (unsigned ThreadID = 0; ThreadID < ThreadCount; ++ThreadID) {
> + Threads.emplace_back([&] {
> + while (true) {
> + std::packaged_task<void()> Task;
> + {
> + std::unique_lock<std::mutex> LockGuard(QueueLock);
> + // Wait for tasks to be pushed in the queue
> + QueueCondition.wait(LockGuard,
> + [&] { return !EnableFlag || !Tasks.empty(); });
> + // Exit condition
> + if (!EnableFlag && Tasks.empty())
> + return;
> + // Yeah, we have a task, grab it and release the lock on the queue
> +
> + // We first need to signal that we are active before popping the queue
> + // in order for wait() to properly detect that even if the queue is
> + // empty, there is still a task in flight.
> + {
> + ++ActiveThreads;
> + std::unique_lock<std::mutex> LockGuard(CompletionLock);
> + }
> + Task = std::move(Tasks.front());
> + Tasks.pop();
> + }
> + // Run the task we just grabbed
> + Task();
> +
> + {
> + // Adjust `ActiveThreads`, in case someone waits on ThreadPool::wait()
> + std::unique_lock<std::mutex> LockGuard(CompletionLock);
> + --ActiveThreads;
> + }
> +
> + // Notify task completion, in case someone waits on ThreadPool::wait()
> + CompletionCondition.notify_all();
> + }
> + });
> + }
> +}
> +
> +void ThreadPool::wait() {
> + // Wait for all threads to complete and the queue to be empty
> + std::unique_lock<std::mutex> LockGuard(CompletionLock);
> + CompletionCondition.wait(LockGuard,
> + [&] { return Tasks.empty() && !ActiveThreads; });
> +}
> +
> +std::shared_future<void> ThreadPool::asyncImpl(TaskTy Task) {
> + /// Wrap the Task in a packaged_task to return a future object.
> + std::packaged_task<void()> PackagedTask(std::move(Task));
> + std::future<void> Future = PackagedTask.get_future();
> + {
> + // Lock the queue and push the new task
> + std::unique_lock<std::mutex> LockGuard(QueueLock);
> +
> + // Don't allow enqueueing after disabling the pool
> + assert(EnableFlag && "Queuing a thread during ThreadPool destruction");
> +
> + Tasks.push(std::move(PackagedTask));
> + }
> + QueueCondition.notify_one();
> + return Future.share();
> +}
> +
> +// The destructor joins all threads, waiting for completion.
> +ThreadPool::~ThreadPool() {
> + {
> + std::unique_lock<std::mutex> LockGuard(QueueLock);
> + EnableFlag = false;
> + }
> + QueueCondition.notify_all();
> + for (auto &Worker : Threads)
> + Worker.join();
> +}
> +
> +#else // LLVM_ENABLE_THREADS Disabled
> +
> +ThreadPool::ThreadPool() : ThreadPool(0) {}
> +
> +// No threads are launched, issue a warning if ThreadCount is not 0
> +ThreadPool::ThreadPool(unsigned ThreadCount)
> + : ActiveThreads(0), EnableFlag(true) {
> + if (ThreadCount) {
> + errs() << "Warning: request a ThreadPool with " << ThreadCount
> + << " threads, but LLVM_ENABLE_THREADS has been turned off\n";
> + }
> +}
> +
> +void ThreadPool::wait() {
> + // Sequential implementation running the tasks
> + while (!Tasks.empty()) {
> + auto Task = std::move(Tasks.front());
> + Tasks.pop();
> + Task();
> + }
> +}
> +
> +std::shared_future<void> ThreadPool::asyncImpl(TaskTy Task) {
> + // Get a Future with launch::deferred execution using std::async
> + auto Future = std::async(std::launch::deferred, std::move(Task)).share();
> + // Wrap the future so that both ThreadPool::wait() can operate and the
> + // returned future can be sync'ed on.
> + std::packaged_task<void()> PackagedTask([Future]() { Future.get(); });
> + Tasks.push(std::move(PackagedTask));
> + return Future;
> +}
> +
> +ThreadPool::~ThreadPool() {
> + EnableFlag = false;
> + wait();
> +}
> +
> +#endif
>
> Modified: llvm/trunk/unittests/Support/CMakeLists.txt
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/unittests/Support/CMakeLists.txt?rev=255444&r1=255443&r2=255444&view=diff
> ==============================================================================
> --- llvm/trunk/unittests/Support/CMakeLists.txt (original)
> +++ llvm/trunk/unittests/Support/CMakeLists.txt Sat Dec 12 16:55:25 2015
> @@ -41,6 +41,7 @@ add_llvm_unittest(SupportTests
> SwapByteOrderTest.cpp
> TargetRegistry.cpp
> ThreadLocalTest.cpp
> + ThreadPool.cpp
> TimeValueTest.cpp
> TrailingObjectsTest.cpp
> UnicodeTest.cpp
>
> Added: llvm/trunk/unittests/Support/ThreadPool.cpp
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/unittests/Support/ThreadPool.cpp?rev=255444&view=auto
> ==============================================================================
> --- llvm/trunk/unittests/Support/ThreadPool.cpp (added)
> +++ llvm/trunk/unittests/Support/ThreadPool.cpp Sat Dec 12 16:55:25 2015
> @@ -0,0 +1,91 @@
> +//========- unittests/Support/ThreadPools.cpp - ThreadPools.h tests --========//
> +//
> +// The LLVM Compiler Infrastructure
> +//
> +// This file is distributed under the University of Illinois Open Source
> +// License. See LICENSE.TXT for details.
> +//
> +//===----------------------------------------------------------------------===//
> +
> +#include "llvm/Support/ThreadPool.h"
> +
> +#include "llvm/ADT/STLExtras.h"
> +
> +#include "gtest/gtest.h"
> +
> +using namespace llvm;
> +using namespace std::chrono;
> +
> +/// Try best to make this thread not progress faster than the main thread
> +static void yield() {
> +#ifdef LLVM_ENABLE_THREADS
> + std::this_thread::yield();
> +#endif
> + std::this_thread::sleep_for(milliseconds(200));
> +#ifdef LLVM_ENABLE_THREADS
> + std::this_thread::yield();
> +#endif
> +}
> +
> +TEST(ThreadPoolTest, AsyncBarrier) {
> + // test that async & barrier work together properly.
> +
> + std::atomic_int checked_in{0};
> +
> + ThreadPool Pool;
> + for (size_t i = 0; i < 5; ++i) {
> + Pool.async([&checked_in, i] {
> + yield();
> + ++checked_in;
> + });
> + }
> + ASSERT_EQ(0, checked_in);
> + Pool.wait();
> + ASSERT_EQ(5, checked_in);
> +}
> +
> +TEST(ThreadPoolTest, Async) {
> + ThreadPool Pool;
> + std::atomic_int i{0};
> + // sleep here just to ensure that the not-equal is correct.
I'm not sure that the usage of yield() ensures this... please see my comment at the end of the email.
This is related to https://llvm.org/bugs/show_bug.cgi?id=25829.
> + Pool.async([&i] {
> + yield();
> + ++i;
> + });
> + Pool.async([&i] { ++i; });
> + ASSERT_NE(2, i.load());
> + Pool.wait();
> + ASSERT_EQ(2, i.load());
> +}
> +
> +TEST(ThreadPoolTest, GetFuture) {
> + ThreadPool Pool;
> + std::atomic_int i{0};
> + // sleep here just to ensure that the not-equal is correct.
> + Pool.async([&i] {
> + yield();
> + ++i;
> + });
> + // Force the future using get()
> + Pool.async([&i] { ++i; }).get();
> + ASSERT_NE(2, i.load());
> + Pool.wait();
> + ASSERT_EQ(2, i.load());
> +}
> +
> +TEST(ThreadPoolTest, PoolDestruction) {
> + // Test that we are waiting on destruction
> + std::atomic_int checked_in{0};
> +
> + {
> + ThreadPool Pool;
> + for (size_t i = 0; i < 5; ++i) {
> + Pool.async([&checked_in, i] {
> + yield();
Mehdi, I don't think the usage of yield() here guarantees that no updates to `checked_in` happen before we hit `ASSERT_EQ(0, checked_in)`. Yaron referenced a failure in this spot on the -dev list earlier -- could you take a look?
> + ++checked_in;
> + });
> + }
> + ASSERT_EQ(0, checked_in);
> + }
> + ASSERT_EQ(5, checked_in);
> +}
vedant
More information about the llvm-commits
mailing list