mirror of
https://github.com/mariadb-corporation/mariadb-columnstore-engine.git
synced 2025-04-18 21:44:02 +03:00
* fix(threadpool): MCOL-5565 queries stuck in FairThreadScheduler. (#3100) Meta Primitive Jobs, .e.g ADD_JOINER, LAST_JOINER stuck in Fair scheduler without out-of-band scheduler. Add OOB scheduler back to remedy the issue. * fix(messageqcpp): MCOL-5636 same node communication crashes transmiting PP errors to EM b/c error messaging leveraged socket that was a nullptr. (#3106) * fix(threadpool): MCOL-5645 errenous threadpool Job ctor implictly sets socket shared_ptr to nullptr causing sigabrt when threadpool returns an error (#3125) --------- Co-authored-by: drrtuy <roman.nozdrin@mariadb.com>
337 lines
8.6 KiB
C++
337 lines
8.6 KiB
C++
/* Copyright (C) 2014 InfiniDB, Inc.
|
|
|
|
This program is free software; you can redistribute it and/or
|
|
modify it under the terms of the GNU General Public License
|
|
as published by the Free Software Foundation; version 2 of
|
|
the License.
|
|
|
|
This program is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
GNU General Public License for more details.
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
along with this program; if not, write to the Free Software
|
|
Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston,
|
|
MA 02110-1301, USA. */
|
|
|
|
/***********************************************************************
|
|
* $Id: threadpool.cpp 553 2008-02-27 17:51:16Z rdempsey $
|
|
*
|
|
*
|
|
***********************************************************************/
|
|
|
|
#include <unistd.h>
|
|
#include <exception>
|
|
using namespace std;
|
|
|
|
#include "messageobj.h"
|
|
#include "messagelog.h"
|
|
#include "threadnaming.h"
|
|
using namespace logging;
|
|
|
|
#include "prioritythreadpool.h"
|
|
using namespace boost;
|
|
|
|
#include "dbcon/joblist/primitivemsg.h"
|
|
|
|
namespace error_handling
|
|
{
|
|
messageqcpp::SBS makePrimitiveErrorMsg(const uint16_t status, const uint32_t id, const uint32_t step)
|
|
{
|
|
ISMPacketHeader ism;
|
|
ism.Status = status;
|
|
|
|
PrimitiveHeader ph = {0, 0, 0, step, id, 0};
|
|
|
|
messageqcpp::SBS errorMsg(new messageqcpp::ByteStream(sizeof(ISMPacketHeader) + sizeof(PrimitiveHeader)));
|
|
errorMsg->append((uint8_t*)&ism, sizeof(ism));
|
|
errorMsg->append((uint8_t*)&ph, sizeof(ph));
|
|
|
|
return errorMsg;
|
|
}
|
|
|
|
void sendErrorMsg(const uint16_t status, const uint32_t id, const uint32_t step,
|
|
primitiveprocessor::SP_UM_IOSOCK sock)
|
|
{
|
|
auto errorMsg = error_handling::makePrimitiveErrorMsg(status, id, step);
|
|
|
|
sock->write(errorMsg);
|
|
}
|
|
|
|
} // namespace error_handling
|
|
|
|
namespace threadpool
|
|
{
|
|
PriorityThreadPool::PriorityThreadPool(uint targetWeightPerRun, uint highThreads, uint midThreads,
|
|
uint lowThreads, uint ID)
|
|
: _stop(false), weightPerRun(targetWeightPerRun), id(ID), blockedThreads(0), extraThreads(0), stopExtra(true)
|
|
{
|
|
boost::thread* newThread;
|
|
for (uint32_t i = 0; i < highThreads; i++)
|
|
{
|
|
newThread = threads.create_thread(ThreadHelper(this, HIGH));
|
|
newThread->detach();
|
|
}
|
|
for (uint32_t i = 0; i < midThreads; i++)
|
|
{
|
|
newThread = threads.create_thread(ThreadHelper(this, MEDIUM));
|
|
newThread->detach();
|
|
}
|
|
for (uint32_t i = 0; i < lowThreads; i++)
|
|
{
|
|
newThread = threads.create_thread(ThreadHelper(this, LOW));
|
|
newThread->detach();
|
|
}
|
|
cout << "started " << highThreads << " high, " << midThreads << " med, " << lowThreads << " low.\n";
|
|
defaultThreadCounts[HIGH] = threadCounts[HIGH] = highThreads;
|
|
defaultThreadCounts[MEDIUM] = threadCounts[MEDIUM] = midThreads;
|
|
defaultThreadCounts[LOW] = threadCounts[LOW] = lowThreads;
|
|
}
|
|
|
|
PriorityThreadPool::~PriorityThreadPool()
|
|
{
|
|
stop();
|
|
}
|
|
|
|
void PriorityThreadPool::addJob(const Job& job, bool useLock)
|
|
{
|
|
boost::thread* newThread;
|
|
boost::mutex::scoped_lock lk(mutex, boost::defer_lock_t());
|
|
|
|
if (useLock)
|
|
lk.lock();
|
|
|
|
// Create any missing threads
|
|
if (defaultThreadCounts[HIGH] != threadCounts[HIGH])
|
|
{
|
|
newThread = threads.create_thread(ThreadHelper(this, HIGH));
|
|
newThread->detach();
|
|
threadCounts[HIGH]++;
|
|
}
|
|
|
|
if (defaultThreadCounts[MEDIUM] != threadCounts[MEDIUM])
|
|
{
|
|
newThread = threads.create_thread(ThreadHelper(this, MEDIUM));
|
|
newThread->detach();
|
|
threadCounts[MEDIUM]++;
|
|
}
|
|
|
|
if (defaultThreadCounts[LOW] != threadCounts[LOW])
|
|
{
|
|
newThread = threads.create_thread(ThreadHelper(this, LOW));
|
|
newThread->detach();
|
|
threadCounts[LOW]++;
|
|
}
|
|
|
|
// If some threads have blocked (because of output queue full)
|
|
// Temporarily add some extra worker threads to make up for the blocked threads.
|
|
if (blockedThreads > extraThreads)
|
|
{
|
|
stopExtra = false;
|
|
newThread = threads.create_thread(ThreadHelper(this, EXTRA));
|
|
newThread->detach();
|
|
extraThreads++;
|
|
}
|
|
else if (blockedThreads == 0)
|
|
{
|
|
// Release the temporary threads -- some threads have become unblocked.
|
|
stopExtra = true;
|
|
}
|
|
|
|
if (job.priority > 66)
|
|
jobQueues[HIGH].push_back(job);
|
|
else if (job.priority > 33)
|
|
jobQueues[MEDIUM].push_back(job);
|
|
else
|
|
jobQueues[LOW].push_back(job);
|
|
|
|
if (useLock)
|
|
newJob.notify_one();
|
|
}
|
|
|
|
void PriorityThreadPool::removeJobs(uint32_t id)
|
|
{
|
|
list<Job>::iterator it;
|
|
|
|
boost::mutex::scoped_lock lk(mutex);
|
|
|
|
for (uint32_t i = 0; i < _COUNT; i++)
|
|
for (it = jobQueues[i].begin(); it != jobQueues[i].end();)
|
|
if (it->id == id)
|
|
it = jobQueues[i].erase(it);
|
|
else
|
|
++it;
|
|
}
|
|
|
|
PriorityThreadPool::Priority PriorityThreadPool::pickAQueue(Priority preference)
|
|
{
|
|
if (preference != EXTRA && !jobQueues[preference].empty())
|
|
return preference;
|
|
else if (!jobQueues[HIGH].empty())
|
|
return HIGH;
|
|
else if (!jobQueues[MEDIUM].empty())
|
|
return MEDIUM;
|
|
else
|
|
return LOW;
|
|
}
|
|
|
|
void PriorityThreadPool::threadFcn(const Priority preferredQueue) throw()
|
|
{
|
|
if (preferredQueue == EXTRA)
|
|
utils::setThreadName("Extra");
|
|
else
|
|
utils::setThreadName("Idle");
|
|
Priority queue = LOW;
|
|
uint32_t weight, i = 0;
|
|
vector<Job> runList;
|
|
vector<bool> reschedule;
|
|
uint32_t rescheduleCount;
|
|
uint32_t queueSize;
|
|
bool running = false;
|
|
|
|
try
|
|
{
|
|
while (!_stop)
|
|
{
|
|
boost::mutex::scoped_lock lk(mutex);
|
|
|
|
queue = pickAQueue(preferredQueue);
|
|
|
|
if (jobQueues[queue].empty())
|
|
{
|
|
// If this is an EXTRA thread due toother threads blocking, and all blockers are unblocked,
|
|
// we don't want this one any more.
|
|
if (preferredQueue == EXTRA && stopExtra)
|
|
{
|
|
extraThreads--;
|
|
return;
|
|
}
|
|
|
|
newJob.wait(lk);
|
|
continue;
|
|
}
|
|
|
|
queueSize = jobQueues[queue].size();
|
|
weight = 0;
|
|
// 3 conditions stop this thread from grabbing all jobs in the queue
|
|
//
|
|
// 1: The weight limit has been exceeded
|
|
// 2: The queue is empty
|
|
// 3: It has grabbed more than half of the jobs available &
|
|
// should leave some to the other threads
|
|
|
|
while ((weight < weightPerRun) && (!jobQueues[queue].empty()) && (runList.size() <= queueSize / 2))
|
|
{
|
|
runList.push_back(jobQueues[queue].front());
|
|
jobQueues[queue].pop_front();
|
|
weight += runList.back().weight;
|
|
}
|
|
|
|
lk.unlock();
|
|
|
|
reschedule.resize(runList.size());
|
|
rescheduleCount = 0;
|
|
|
|
for (i = 0; i < runList.size() && !_stop; i++)
|
|
{
|
|
reschedule[i] = false;
|
|
running = true;
|
|
reschedule[i] = (*(runList[i].functor))();
|
|
running = false;
|
|
|
|
if (reschedule[i])
|
|
rescheduleCount++;
|
|
}
|
|
if (preferredQueue == EXTRA)
|
|
utils::setThreadName("Extra (used)");
|
|
else
|
|
utils::setThreadName("Idle");
|
|
|
|
// no real work was done, prevent intensive busy waiting
|
|
if (rescheduleCount == runList.size())
|
|
usleep(1000);
|
|
|
|
if (rescheduleCount > 0)
|
|
{
|
|
lk.lock();
|
|
|
|
for (i = 0; i < runList.size(); i++)
|
|
if (reschedule[i])
|
|
addJob(runList[i], false);
|
|
|
|
if (rescheduleCount > 1)
|
|
newJob.notify_all();
|
|
else
|
|
newJob.notify_one();
|
|
|
|
lk.unlock();
|
|
}
|
|
|
|
runList.clear();
|
|
}
|
|
}
|
|
catch (std::exception& ex)
|
|
{
|
|
// Log the exception and exit this thread
|
|
try
|
|
{
|
|
threadCounts[queue]--;
|
|
#ifndef NOLOGGING
|
|
logging::Message::Args args;
|
|
logging::Message message(5);
|
|
args.add("threadFcn: Caught exception: ");
|
|
args.add(ex.what());
|
|
|
|
message.format(args);
|
|
|
|
logging::LoggingID lid(22);
|
|
logging::MessageLog ml(lid);
|
|
|
|
ml.logErrorMessage(message);
|
|
#endif
|
|
|
|
if (running)
|
|
error_handling::sendErrorMsg(logging::primitiveServerErr, runList[i].uniqueID, runList[i].stepID,
|
|
runList[i].sock);
|
|
}
|
|
catch (...)
|
|
{
|
|
}
|
|
}
|
|
catch (...)
|
|
{
|
|
// Log the exception and exit this thread
|
|
try
|
|
{
|
|
threadCounts[queue]--;
|
|
#ifndef NOLOGGING
|
|
logging::Message::Args args;
|
|
logging::Message message(6);
|
|
args.add("threadFcn: Caught unknown exception!");
|
|
|
|
message.format(args);
|
|
|
|
logging::LoggingID lid(22);
|
|
logging::MessageLog ml(lid);
|
|
|
|
ml.logErrorMessage(message);
|
|
#endif
|
|
|
|
if (running)
|
|
error_handling::sendErrorMsg(logging::primitiveServerErr, runList[i].uniqueID, runList[i].stepID,
|
|
runList[i].sock);
|
|
}
|
|
catch (...)
|
|
{
|
|
}
|
|
}
|
|
}
|
|
|
|
void PriorityThreadPool::stop()
|
|
{
|
|
_stop = true;
|
|
}
|
|
|
|
} // namespace threadpool
|