cOMS/thread/ThreadPool.h

191 lines
5.3 KiB
C

/**
* Jingga
*
* @copyright Jingga
* @license OMS License 2.0
* @version 1.0.0
* @link https://jingga.app
*/
#ifndef TOS_THREADS_THREAD_POOL_H
#define TOS_THREADS_THREAD_POOL_H
#include <stdio.h>
#include <stdlib.h>
#include "../stdlib/Types.h"
#include "../memory/Queue.h"
#include "../memory/BufferMemory.h"
#ifdef _WIN32
#include "../platform/win32/threading/Thread.h"
#include "../platform/win32/threading/Atomic.h"
#elif __linux__
#include "../platform/linux/threading/Thread.h"
#include "../platform/linux/threading/Atomic.h"
#endif
#include "ThreadJob.h"
struct ThreadPool {
// This is not a threaded queue since we want to handle the mutex in here, not in the queue for finer control
Queue work_queue;
pthread_mutex_t work_mutex;
pthread_cond_t work_cond;
pthread_cond_t working_cond;
int32 working_cnt;
int32 thread_cnt;
int32 size;
int32 state;
uint32 id_counter;
};
static THREAD_RETURN thread_pool_worker(void* arg)
{
ThreadPool* pool = (ThreadPool *) arg;
PoolWorker* work;
while (true) {
pthread_mutex_lock(&pool->work_mutex);
while (queue_is_empty(&pool->work_queue) && !pool->state) {
pthread_cond_wait(&pool->work_cond, &pool->work_mutex);
}
if (pool->state == 1) {
pthread_mutex_unlock(&pool->work_mutex);
break;
}
work = (PoolWorker *) queue_dequeue_keep(&pool->work_queue, sizeof(PoolWorker), 64);
pthread_mutex_unlock(&pool->work_mutex);
if (!work) {
continue;
}
atomic_increment(&pool->working_cnt);
atomic_set(&work->state, 2);
work->func(work);
atomic_set(&work->state, 1);
// Job gets marked after completion -> can be overwritten now
if (atomic_get(&work->id) == -1) {
atomic_set(&work->id, 0);
}
atomic_decrement(&pool->working_cnt);
if (atomic_get(&pool->state) == 0 && atomic_get(&pool->working_cnt) == 0) {
pthread_cond_signal(&pool->working_cond);
}
}
pthread_cond_signal(&pool->working_cond);
atomic_decrement(&pool->thread_cnt);
return NULL;
}
void thread_pool_create(ThreadPool* pool, BufferMemory* buf, int32 thread_count)
{
queue_init(&pool->work_queue, buf, 64, sizeof(PoolWorker), 64);
pool->thread_cnt = thread_count;
// @todo switch from pool mutex and pool cond to threadjob mutex/cond
// thread_pool_wait etc. should just iterate over all mutexes
pthread_mutex_init(&pool->work_mutex, NULL);
pthread_cond_init(&pool->work_cond, NULL);
pthread_cond_init(&pool->working_cond, NULL);
pthread_t thread;
for (pool->size = 0; pool->size < thread_count; ++pool->size) {
pthread_create(&thread, NULL, thread_pool_worker, pool);
pthread_detach(thread);
}
}
void thread_pool_wait(ThreadPool* pool)
{
pthread_mutex_lock(&pool->work_mutex);
while ((!pool->state && pool->working_cnt != 0) || (pool->state && pool->thread_cnt != 0)) {
pthread_cond_wait(&pool->working_cond, &pool->work_mutex);
}
pthread_mutex_unlock(&pool->work_mutex);
}
void thread_pool_destroy(ThreadPool* pool)
{
// This sets the queue to empty
atomic_set((void **) &pool->work_queue.tail, (void **) &pool->work_queue.head);
// This sets the state to "shutdown"
atomic_set(&pool->state, 1);
pthread_cond_broadcast(&pool->work_cond);
thread_pool_wait(pool);
pthread_mutex_destroy(&pool->work_mutex);
pthread_cond_destroy(&pool->work_cond);
pthread_cond_destroy(&pool->working_cond);
}
PoolWorker* thread_pool_add_work(ThreadPool* pool, const PoolWorker* job)
{
pthread_mutex_lock(&pool->work_mutex);
PoolWorker* temp_job = (PoolWorker *) ring_get_memory_nomove(&pool->work_queue, sizeof(PoolWorker), 64);
if (atomic_get(&temp_job->id) > 0) {
pthread_mutex_unlock(&pool->work_mutex);
ASSERT_SIMPLE(temp_job->id == 0);
return NULL;
}
memcpy(temp_job, job, sizeof(PoolWorker));
ring_move_pointer(&pool->work_queue, &pool->work_queue.head, sizeof(PoolWorker), 64);
if (temp_job->id == 0) {
temp_job->id = atomic_add_fetch(&pool->id_counter, 1);
}
pthread_cond_broadcast(&pool->work_cond);
pthread_mutex_unlock(&pool->work_mutex);
return temp_job;
}
// This is basically the same as thread_pool_add_work but allows us to directly write into the memory in the caller
// This makes it faster, since we can avoid a memcpy
PoolWorker* thread_pool_add_work_start(ThreadPool* pool)
{
pthread_mutex_lock(&pool->work_mutex);
PoolWorker* temp_job = (PoolWorker *) queue_enqueue_start(&pool->work_queue, sizeof(PoolWorker), 64);
if (atomic_get(&temp_job->id) > 0) {
pthread_mutex_unlock(&pool->work_mutex);
ASSERT_SIMPLE(temp_job->id == 0);
return NULL;
}
if (temp_job->id == 0) {
// +1 because otherwise the very first job would be id = 0 which is not a valid id
temp_job->id = atomic_add_fetch(&pool->id_counter, 1) + 1;
}
return temp_job;
}
void thread_pool_add_work_end(ThreadPool* pool)
{
queue_enqueue_end(&pool->work_queue, sizeof(PoolWorker), 64);
pthread_cond_broadcast(&pool->work_cond);
pthread_mutex_unlock(&pool->work_mutex);
}
#endif