| Index: src/utils/SkTaskGroup.cpp
|
| diff --git a/src/utils/SkTaskGroup.cpp b/src/utils/SkTaskGroup.cpp
|
| deleted file mode 100644
|
| index f1ec7f42788a198057aa0a2a4bcd916de9bb0007..0000000000000000000000000000000000000000
|
| --- a/src/utils/SkTaskGroup.cpp
|
| +++ /dev/null
|
| @@ -1,160 +0,0 @@
|
| -#include "SkTaskGroup.h"
|
| -
|
| -#include "SkCondVar.h"
|
| -#include "SkTDArray.h"
|
| -#include "SkThread.h"
|
| -#include "SkThreadUtils.h"
|
| -
|
| -#if defined(SK_BUILD_FOR_WIN32)
|
| - static inline int num_cores() {
|
| - SYSTEM_INFO sysinfo;
|
| - GetSystemInfo(&sysinfo);
|
| - return sysinfo.dwNumberOfProcessors;
|
| - }
|
| -#else
|
| - #include <unistd.h>
|
| - static inline int num_cores() {
|
| - return (int) sysconf(_SC_NPROCESSORS_ONLN);
|
| - }
|
| -#endif
|
| -
|
| -namespace {
|
| -
|
| -class ThreadPool : SkNoncopyable {
|
| -public:
|
| - static void Add(SkRunnable* task, int32_t* pending) {
|
| - if (!gGlobal) { // If we have no threads, run synchronously.
|
| - return task->run();
|
| - }
|
| - gGlobal->add(&CallRunnable, task, pending);
|
| - }
|
| -
|
| - static void Add(void (*fn)(void*), void* arg, int32_t* pending) {
|
| - if (!gGlobal) {
|
| - return fn(arg);
|
| - }
|
| - gGlobal->add(fn, arg, pending);
|
| - }
|
| -
|
| - static void Wait(int32_t* pending) {
|
| - if (!gGlobal) { // If we have no threads, the work must already be done.
|
| - SkASSERT(*pending == 0);
|
| - return;
|
| - }
|
| - while (sk_acquire_load(pending) > 0) { // Pairs with sk_atomic_dec here or in Loop.
|
| - // Lend a hand until our SkTaskGroup of interest is done.
|
| - Work work;
|
| - {
|
| - AutoLock lock(&gGlobal->fReady);
|
| - if (gGlobal->fWork.isEmpty()) {
|
| - // Someone has picked up all the work (including ours). How nice of them!
|
| - // (They may still be working on it, so we can't assert *pending == 0 here.)
|
| - continue;
|
| - }
|
| - gGlobal->fWork.pop(&work);
|
| - }
|
| - // This Work isn't necessarily part of our SkTaskGroup of interest, but that's fine.
|
| - // We threads gotta stick together. We're always making forward progress.
|
| - work.fn(work.arg);
|
| - sk_atomic_dec(work.pending); // Release pairs with the sk_acquire_load() just above.
|
| - }
|
| - }
|
| -
|
| -private:
|
| - struct AutoLock {
|
| - AutoLock(SkCondVar* c) : fC(c) { fC->lock(); }
|
| - ~AutoLock() { fC->unlock(); }
|
| - private:
|
| - SkCondVar* fC;
|
| - };
|
| -
|
| - static void CallRunnable(void* arg) { static_cast<SkRunnable*>(arg)->run(); }
|
| -
|
| - struct Work {
|
| - void (*fn)(void*); // A function to call,
|
| - void* arg; // its argument,
|
| - int32_t* pending; // then sk_atomic_dec(pending) afterwards.
|
| - };
|
| -
|
| - explicit ThreadPool(int threads) : fDraining(false) {
|
| - if (threads == -1) {
|
| - threads = num_cores();
|
| - }
|
| - for (int i = 0; i < threads; i++) {
|
| - fThreads.push(SkNEW_ARGS(SkThread, (&ThreadPool::Loop, this)));
|
| - fThreads.top()->start();
|
| - }
|
| - }
|
| -
|
| - ~ThreadPool() {
|
| - SkASSERT(fWork.isEmpty()); // All SkTaskGroups should be destroyed by now.
|
| - {
|
| - AutoLock lock(&fReady);
|
| - fDraining = true;
|
| - fReady.broadcast();
|
| - }
|
| - for (int i = 0; i < fThreads.count(); i++) {
|
| - fThreads[i]->join();
|
| - }
|
| - SkASSERT(fWork.isEmpty()); // Can't hurt to double check.
|
| - fThreads.deleteAll();
|
| - }
|
| -
|
| - void add(void (*fn)(void*), void* arg, int32_t* pending) {
|
| - Work work = { fn, arg, pending };
|
| - sk_atomic_inc(pending); // No barrier needed.
|
| - {
|
| - AutoLock lock(&fReady);
|
| - fWork.push(work);
|
| - fReady.signal();
|
| - }
|
| - }
|
| -
|
| - static void Loop(void* arg) {
|
| - ThreadPool* pool = (ThreadPool*)arg;
|
| - Work work;
|
| - while (true) {
|
| - {
|
| - AutoLock lock(&pool->fReady);
|
| - while (pool->fWork.isEmpty()) {
|
| - if (pool->fDraining) {
|
| - return;
|
| - }
|
| - pool->fReady.wait();
|
| - }
|
| - pool->fWork.pop(&work);
|
| - }
|
| - work.fn(work.arg);
|
| - sk_atomic_dec(work.pending); // Release pairs with sk_acquire_load() in Wait().
|
| - }
|
| - }
|
| -
|
| - SkTDArray<Work> fWork;
|
| - SkTDArray<SkThread*> fThreads;
|
| - SkCondVar fReady;
|
| - bool fDraining;
|
| -
|
| - static ThreadPool* gGlobal;
|
| - friend struct SkTaskGroup::Enabler;
|
| -};
|
| -ThreadPool* ThreadPool::gGlobal = NULL;
|
| -
|
| -} // namespace
|
| -
|
| -SkTaskGroup::Enabler::Enabler(int threads) {
|
| - SkASSERT(ThreadPool::gGlobal == NULL);
|
| - if (threads != 0) {
|
| - ThreadPool::gGlobal = SkNEW_ARGS(ThreadPool, (threads));
|
| - }
|
| -}
|
| -
|
| -SkTaskGroup::Enabler::~Enabler() {
|
| - SkDELETE(ThreadPool::gGlobal);
|
| -}
|
| -
|
| -SkTaskGroup::SkTaskGroup() : fPending(0) {}
|
| -
|
| -void SkTaskGroup::add(SkRunnable* task) { ThreadPool::Add(task, &fPending); }
|
| -void SkTaskGroup::add(void (*fn)(void*), void* arg) { ThreadPool::Add(fn, arg, &fPending); }
|
| -void SkTaskGroup::wait() { ThreadPool::Wait(&fPending); }
|
| -
|
|
|