office-gobmx/comphelper/source/misc/threadpool.cxx
Luboš Luňák 3a6a7f71ec fix possible >= 0 warning
Change-Id: I7b5595265d00a33e933a1768d8c80a7ffd25c8a6
Reviewed-on: https://gerrit.libreoffice.org/c/core/+/103997
Tested-by: Jenkins
Reviewed-by: Miklos Vajna <vmiklos@collabora.com>
2020-10-06 12:00:54 +02:00

380 lines
9.6 KiB
C++

/* -*- Mode: C++; tab-width: 4; indent-tabs-mode: nil; c-basic-offset: 4 -*- */
/*
* This file is part of the LibreOffice project.
*
* This Source Code Form is subject to the terms of the Mozilla Public
* License, v. 2.0. If a copy of the MPL was not distributed with this
* file, You can obtain one at http://mozilla.org/MPL/2.0/.
*/
#include <comphelper/threadpool.hxx>
#include <com/sun/star/uno/Exception.hpp>
#include <config_options.h>
#include <sal/config.h>
#include <sal/log.hxx>
#include <rtl/instance.hxx>
#include <salhelper/thread.hxx>
#include <algorithm>
#include <memory>
#include <thread>
#include <chrono>
#include <comphelper/debuggerinfo.hxx>
#if defined HAVE_VALGRIND_HEADERS
#include <valgrind/memcheck.h>
#endif
#if defined(_WIN32)
#define WIN32_LEAN_AND_MEAN
#include <windows.h>
#elif defined UNX
#include <unistd.h>
#include <fcntl.h>
#endif
namespace comphelper {
/** prevent waiting for a task from inside a task */
#if defined DBG_UTIL && (defined LINUX || defined _WIN32)
static thread_local bool gbIsWorkerThread;
#endif
// used to group thread-tasks for waiting in waitTillDone()
class ThreadTaskTag
{
std::mutex maMutex;
sal_Int32 mnTasksWorking;
std::condition_variable maTasksComplete;
public:
ThreadTaskTag();
bool isDone();
void waitUntilDone();
void onTaskWorkerDone();
void onTaskPushed();
};
class ThreadPool::ThreadWorker : public salhelper::Thread
{
ThreadPool *mpPool;
public:
explicit ThreadWorker( ThreadPool *pPool ) :
salhelper::Thread("thread-pool"),
mpPool( pPool )
{
}
virtual void execute() override
{
#if defined DBG_UTIL && (defined LINUX || defined _WIN32)
gbIsWorkerThread = true;
#endif
std::unique_lock< std::mutex > aGuard( mpPool->maMutex );
while( !mpPool->mbTerminate )
{
std::unique_ptr<ThreadTask> pTask = mpPool->popWorkLocked( aGuard, true );
if( pTask )
{
mpPool->incBusyWorker();
aGuard.unlock();
pTask->exec();
pTask.reset();
aGuard.lock();
mpPool->decBusyWorker();
}
}
}
};
ThreadPool::ThreadPool(sal_Int32 nWorkers)
: mbTerminate(true)
, mnMaxWorkers(nWorkers)
, mnBusyWorkers(0)
{
}
ThreadPool::~ThreadPool()
{
// note: calling shutdown from global variable dtor blocks forever on Win7
// note2: there isn't enough MSVCRT left on exit to call assert() properly
// so these asserts just print something to stderr but exit status is
// still 0, but hopefully they will be more helpful on non-WNT platforms
assert(mbTerminate);
assert(maTasks.empty());
assert(mnBusyWorkers == 0);
}
namespace {
struct ThreadPoolStatic : public rtl::StaticWithInit< std::shared_ptr< ThreadPool >,
ThreadPoolStatic >
{
std::shared_ptr< ThreadPool > operator () () {
const sal_Int32 nThreads = ThreadPool::getPreferredConcurrency();
return std::make_shared< ThreadPool >( nThreads );
};
};
}
ThreadPool& ThreadPool::getSharedOptimalPool()
{
return *ThreadPoolStatic::get();
}
sal_Int32 ThreadPool::getPreferredConcurrency()
{
static sal_Int32 ThreadCount = [&]()
{
const sal_Int32 nHardThreads = std::max(std::thread::hardware_concurrency(), 1U);
sal_Int32 nThreads = nHardThreads;
const char *pEnv = getenv("MAX_CONCURRENCY");
if (pEnv != nullptr)
{
// Override with user/admin preference.
nThreads = rtl_str_toInt32(pEnv, 10);
}
nThreads = std::min(nHardThreads, nThreads);
return std::max<sal_Int32>(nThreads, 1);
}();
return ThreadCount;
}
// Used to order shutdown, and to ensure there are no lingering
// threads after LibreOfficeKit pre-init.
void ThreadPool::shutdown()
{
// if (mbTerminate)
// return;
std::unique_lock< std::mutex > aGuard( maMutex );
shutdownLocked(aGuard);
}
void ThreadPool::shutdownLocked(std::unique_lock<std::mutex>& aGuard)
{
if( maWorkers.empty() )
{ // no threads at all -> execute the work in-line
std::unique_ptr<ThreadTask> pTask;
while ( ( pTask = popWorkLocked(aGuard, false) ) )
pTask->exec();
}
else
{
while( !maTasks.empty() )
maTasksChanged.wait( aGuard );
}
assert( maTasks.empty() );
// coverity[missing_lock] - on purpose
mbTerminate = true;
maTasksChanged.notify_all();
decltype(maWorkers) aWorkers;
std::swap(maWorkers, aWorkers);
aGuard.unlock();
while (!aWorkers.empty())
{
rtl::Reference<ThreadWorker> xWorker = aWorkers.back();
aWorkers.pop_back();
assert(std::find(aWorkers.begin(), aWorkers.end(), xWorker)
== aWorkers.end());
{
xWorker->join();
xWorker.clear();
}
}
}
void ThreadPool::pushTask( std::unique_ptr<ThreadTask> pTask )
{
std::scoped_lock< std::mutex > aGuard( maMutex );
mbTerminate = false;
// Worked on tasks are already removed from maTasks, so include the count of busy workers.
if (maWorkers.size() < mnMaxWorkers && maWorkers.size() <= maTasks.size() + mnBusyWorkers)
{
maWorkers.push_back( new ThreadWorker( this ) );
maWorkers.back()->launch();
}
pTask->mpTag->onTaskPushed();
maTasks.insert( maTasks.begin(), std::move(pTask) );
maTasksChanged.notify_one();
}
std::unique_ptr<ThreadTask> ThreadPool::popWorkLocked( std::unique_lock< std::mutex > & rGuard, bool bWait )
{
do
{
if( !maTasks.empty() )
{
std::unique_ptr<ThreadTask> pTask = std::move(maTasks.back());
maTasks.pop_back();
return pTask;
}
else if (!bWait || mbTerminate)
return nullptr;
maTasksChanged.wait( rGuard );
} while (!mbTerminate);
return nullptr;
}
void ThreadPool::incBusyWorker()
{
++mnBusyWorkers;
}
void ThreadPool::decBusyWorker()
{
assert(mnBusyWorkers >= 1);
--mnBusyWorkers;
}
void ThreadPool::waitUntilDone(const std::shared_ptr<ThreadTaskTag>& rTag, bool bJoinAll)
{
#if defined DBG_UTIL && (defined LINUX || defined _WIN32)
assert(!gbIsWorkerThread && "cannot wait for tasks from inside a task");
#endif
{
std::unique_lock< std::mutex > aGuard( maMutex );
if( maWorkers.empty() )
{ // no threads at all -> execute the work in-line
while (!rTag->isDone())
{
std::unique_ptr<ThreadTask> pTask = popWorkLocked(aGuard, false);
if (!pTask)
break;
pTask->exec();
}
}
}
rTag->waitUntilDone();
if (bJoinAll)
joinAll();
}
void ThreadPool::joinAll()
{
std::unique_lock< std::mutex > aGuard( maMutex );
if (maTasks.empty()) // check if there are still tasks from another tag
{
shutdownLocked(aGuard);
}
}
std::shared_ptr<ThreadTaskTag> ThreadPool::createThreadTaskTag()
{
return std::make_shared<ThreadTaskTag>();
}
bool ThreadPool::isTaskTagDone(const std::shared_ptr<ThreadTaskTag>& pTag)
{
return pTag->isDone();
}
ThreadTask::ThreadTask(const std::shared_ptr<ThreadTaskTag>& pTag)
: mpTag(pTag)
{
}
void ThreadTask::exec()
{
std::shared_ptr<ThreadTaskTag> pTag(mpTag);
try {
doWork();
}
catch (const std::exception &e)
{
SAL_WARN("comphelper", "exception in thread worker while calling doWork(): " << e.what());
}
catch (const css::uno::Exception &e)
{
SAL_WARN("comphelper", "exception in thread worker while calling doWork(): " << e);
}
catch (...)
{
SAL_WARN("comphelper", "unknown exception in thread worker while calling doWork()");
}
pTag->onTaskWorkerDone();
}
ThreadTaskTag::ThreadTaskTag() : mnTasksWorking(0)
{
}
void ThreadTaskTag::onTaskPushed()
{
std::scoped_lock< std::mutex > aGuard( maMutex );
mnTasksWorking++;
assert( mnTasksWorking < 65536 ); // sanity checking
}
void ThreadTaskTag::onTaskWorkerDone()
{
std::scoped_lock< std::mutex > aGuard( maMutex );
mnTasksWorking--;
assert(mnTasksWorking >= 0);
if (mnTasksWorking == 0)
maTasksComplete.notify_all();
}
bool ThreadTaskTag::isDone()
{
std::scoped_lock< std::mutex > aGuard( maMutex );
return mnTasksWorking == 0;
}
void ThreadTaskTag::waitUntilDone()
{
std::unique_lock< std::mutex > aGuard( maMutex );
while( mnTasksWorking > 0 )
{
#if defined DBG_UTIL && !defined NDEBUG
// 10 minute timeout in debug mode, unless the code is built with
// sanitizers or debugged in valgrind or gdb, in which case the threads
// should not time out in the middle of a debugging session
int maxTimeout = 10 * 60;
#if !ENABLE_RUNTIME_OPTIMIZATIONS
maxTimeout = 30 * 60;
#endif
#if defined HAVE_VALGRIND_HEADERS
if( RUNNING_ON_VALGRIND )
maxTimeout = 30 * 60;
#endif
if( isDebuggerAttached())
maxTimeout = 300 * 60;
std::cv_status result = maTasksComplete.wait_for(
aGuard, std::chrono::seconds( maxTimeout ));
assert(result != std::cv_status::timeout);
#else
// 10 minute timeout in production so the app eventually throws some kind of error
if (maTasksComplete.wait_for(
aGuard, std::chrono::seconds( 10 * 60 )) == std::cv_status::timeout)
throw std::runtime_error("timeout waiting for threadpool tasks");
#endif
}
}
} // namespace comphelper
/* vim:set shiftwidth=4 softtabstop=4 expandtab: */