LUMIERA.clone/src/backend/threadpool.c

158 lines
5.3 KiB
C
Raw Normal View History

2009-11-23 01:55:08 +01:00
/*
threadpool.c - Manage pools of threads
Copyright (C) Lumiera.org
2009, Michael Ploujnikov <ploujj@gmail.com>
This program is free software; you can redistribute it and/or
modify it under the terms of the GNU General Public License as
published by the Free Software Foundation; either version 2 of the
License, or (at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program; if not, write to the Free Software
Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
*/
//TODO: Support library includes//
#include "include/logging.h"
#include "lib/safeclib.h"
//TODO: Lumiera header includes//
#include "backend/threadpool.h"
//TODO: internal/static forward declarations//
static lumiera_threadpool threadpool;
2009-11-23 01:55:08 +01:00
//TODO: System includes//
#include <pthread.h>
/**
* @file
*
*/
NOBUG_DEFINE_FLAG_PARENT (threadpool, threads_dbg); /*TODO insert a suitable/better parent flag here */
2009-11-23 01:55:08 +01:00
//code goes here//
void* pool_thread_loop(void * arg)
{
(void) arg;
while (1)
{
;
}
return arg;
}
void
lumiera_threadpool_init(unsigned limit)
2009-11-23 01:55:08 +01:00
{
for (int i = 0; i < LUMIERA_THREADCLASS_COUNT; ++i)
{
llist_init(&threadpool.pool[i].list);
threadpool.pool[i].max_threads = limit;
threadpool.pool[i].working_thread_count = 0;
threadpool.pool[i].idle_thread_count = 0;
lumiera_mutex_init(&threadpool.pool[i].lock,"pool of threads", &NOBUG_FLAG(threadpool));
}
}
2009-11-23 01:55:08 +01:00
void
lumiera_threadpool_destroy(void)
{
ECHO ("destroying threadpool");
for (int i = 0; i < LUMIERA_THREADCLASS_COUNT; ++i)
{
ECHO ("destroying individual pool #%d", i);
// no locking is done at this point
ECHO ("number of threads in the pool=%d", llist_count(&threadpool.pool[i].list));
LLIST_WHILE_HEAD(&threadpool.pool[i].list, thread)
lumiera_thread_delete((LumieraThread)thread);
ECHO ("destroying the pool mutex");
lumiera_mutex_destroy (&threadpool.pool[i].lock, &NOBUG_FLAG (threadpool));
ECHO ("pool mutex destroyed");
}
}
2009-11-23 01:55:08 +01:00
LumieraThread
lumiera_threadpool_acquire_thread(enum lumiera_thread_class kind,
const char* purpose,
struct nobug_flag* flag)
{
LumieraThread ret;
REQUIRE (kind < LUMIERA_THREADCLASS_COUNT, "unknown pool kind specified: %d", kind);
if (llist_is_empty (&threadpool.pool[kind].list))
{
// TODO: fill in the reccondition argument, currently NULL
FIXME ("this max thread logic needs to be deeply thought about and made more efficient as well as rebust");
if (threadpool.pool[kind].working_thread_count
+ threadpool.pool[kind].idle_thread_count
< threadpool.pool[kind].max_threads) {
ret = lumiera_thread_new (kind, NULL, purpose, flag);
threadpool.pool[kind].working_thread_count++;
ENSURE (ret, "did not create a valid thread");
}
else
{
//ERROR (threadpool, "did not create a new thread because per-pool limit was reached: %d", threadpool.pool[kind].max_threads);
LUMIERA_DIE(ERRNO);
}
}
else
{
// use an existing thread, pick the first one
// remove it from the pool's list
LUMIERA_MUTEX_SECTION (threadpool, &threadpool.pool[kind].lock)
{
ret = (LumieraThread)(llist_unlink(llist_head (&threadpool.pool[kind].list)));
threadpool.pool[kind].working_thread_count++;
threadpool.pool[kind].idle_thread_count--; // cheaper than using llist_count
ENSURE (threadpool.pool[kind].idle_thread_count ==
llist_count(&threadpool.pool[kind].list),
"idle thread count %d is wrong, should be %d",
threadpool.pool[kind].idle_thread_count,
llist_count(&threadpool.pool[kind].list));
}
ENSURE (ret, "did not find a valid thread");
}
return ret;
2009-11-23 01:55:08 +01:00
}
void
lumiera_threadpool_release_thread(LumieraThread thread)
{
REQUIRE (thread, "invalid thread given");
REQUIRE (thread->kind < LUMIERA_THREADCLASS_COUNT, "thread belongs to an unknown pool kind: %d", thread->kind);
LUMIERA_MUTEX_SECTION (threadpool, &threadpool.pool[thread->kind].lock)
{
REQUIRE (llist_is_single(&thread->node), "thread already belongs to some list");
llist_insert_head(&threadpool.pool[thread->kind].list, &thread->node);
threadpool.pool[thread->kind].working_thread_count--;
threadpool.pool[thread->kind].idle_thread_count++; // cheaper than using llist_count
ENSURE (threadpool.pool[thread->kind].idle_thread_count ==
llist_count(&threadpool.pool[thread->kind].list),
"idle thread count %d is wrong, should be %d",
threadpool.pool[thread->kind].idle_thread_count,
llist_count(&threadpool.pool[thread->kind].list));
// REQUIRE (!llist_is_empty (&threadpool.pool[thread->kind].list), "thread pool is still empty after insertion");
}
}
2009-11-23 01:55:08 +01:00
/*
// Local Variables:
// mode: C
// c-file-style: "gnu"
// indent-tabs-mode: nil
// End:
*/