aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorgingerBill <bill@gingerbill.org>2023-01-02 21:35:40 +0000
committergingerBill <bill@gingerbill.org>2023-01-02 21:35:40 +0000
commit7ffffeecccc6a1fa1b26238f8ed4608d93ec9bb0 (patch)
treeb24281a3c9210bd4633d319119c5614feb5799d7
parentf16d8e77b3f9b3ddf82e97672ca38dc19264eb5e (diff)
Comment out many mutex guards in `type_(size|align)_of_internal`
-rw-r--r--src/thread_pool.cpp21
-rw-r--r--src/types.cpp22
2 files changed, 21 insertions, 22 deletions
diff --git a/src/thread_pool.cpp b/src/thread_pool.cpp
index a429e47ff..b89e00454 100644
--- a/src/thread_pool.cpp
+++ b/src/thread_pool.cpp
@@ -30,7 +30,7 @@ gb_internal void thread_pool_init(ThreadPool *pool, gbAllocator const &a, isize
slice_init(&pool->threads, a, thread_count + 1);
// NOTE: this needs to be initialized before any thread starts
- pool->running.store(true);
+ pool->running.store(true, std::memory_order_seq_cst);
// setup the main thread
thread_init(pool, &pool->threads[0], 0);
@@ -43,7 +43,7 @@ gb_internal void thread_pool_init(ThreadPool *pool, gbAllocator const &a, isize
}
gb_internal void thread_pool_destroy(ThreadPool *pool) {
- pool->running.store(false);
+ pool->running.store(false, std::memory_order_seq_cst);
for_array_off(i, 1, pool->threads) {
Thread *t = &pool->threads[i];
@@ -114,7 +114,7 @@ gb_internal bool thread_pool_add_task(ThreadPool *pool, WorkerTaskProc *proc, vo
gb_internal void thread_pool_wait(ThreadPool *pool) {
WorkerTask task;
- while (pool->tasks_left.load()) {
+ while (pool->tasks_left.load(std::memory_order_acquire)) {
// if we've got tasks on our queue, run them
while (thread_pool_queue_pop(current_thread, &task)) {
task.do_work(task.data);
@@ -126,7 +126,7 @@ gb_internal void thread_pool_wait(ThreadPool *pool) {
// This *must* be executed in this order, so the futex wakes immediately
// if rem_tasks has changed since we checked last, otherwise the program
// will permanently sleep
- Footex rem_tasks = pool->tasks_left.load();
+ Footex rem_tasks = pool->tasks_left.load(std::memory_order_acquire);
if (rem_tasks == 0) {
return;
}
@@ -141,7 +141,7 @@ gb_internal THREAD_PROC(thread_pool_thread_proc) {
ThreadPool *pool = current_thread->pool;
// debugf("worker id: %td\n", current_thread->idx);
- while (pool->running.load()) {
+ while (pool->running.load(std::memory_order_seq_cst)) {
// If we've got tasks to process, work through them
usize finished_tasks = 0;
i32 state;
@@ -152,30 +152,29 @@ gb_internal THREAD_PROC(thread_pool_thread_proc) {
finished_tasks += 1;
}
- if (finished_tasks > 0 && pool->tasks_left.load() == 0) {
+ if (finished_tasks > 0 && pool->tasks_left.load(std::memory_order_acquire) == 0) {
futex_signal(&pool->tasks_left);
}
// If there's still work somewhere and we don't have it, steal it
- if (pool->tasks_left.load()) {
+ if (pool->tasks_left.load(std::memory_order_acquire)) {
usize idx = cast(usize)current_thread->idx;
for_array(i, pool->threads) {
- if (pool->tasks_left.load() == 0) {
+ if (pool->tasks_left.load(std::memory_order_acquire) == 0) {
break;
}
idx = (idx + 1) % cast(usize)pool->threads.count;
Thread *thread = &pool->threads.data[idx];
- WorkerTask task;
+ WorkerTask task, another_task;
if (!thread_pool_queue_pop(thread, &task)) {
continue;
}
-
task.do_work(task.data);
pool->tasks_left.fetch_sub(1, std::memory_order_release);
- if (pool->tasks_left.load() == 0) {
+ if (pool->tasks_left.load(std::memory_order_acquire) == 0) {
futex_signal(&pool->tasks_left);
}
diff --git a/src/types.cpp b/src/types.cpp
index c49f43f7c..1e2d85ac6 100644
--- a/src/types.cpp
+++ b/src/types.cpp
@@ -3406,7 +3406,7 @@ gb_internal i64 type_align_of_internal(Type *t, TypePath *path) {
} break;
case Type_Array: {
- MUTEX_GUARD(&g_type_mutex);
+ // MUTEX_GUARD(&g_type_mutex);
Type *elem = t->Array.elem;
bool pop = type_path_push(path, elem);
@@ -3419,7 +3419,7 @@ gb_internal i64 type_align_of_internal(Type *t, TypePath *path) {
}
case Type_EnumeratedArray: {
- MUTEX_GUARD(&g_type_mutex);
+ // MUTEX_GUARD(&g_type_mutex);
Type *elem = t->EnumeratedArray.elem;
bool pop = type_path_push(path, elem);
@@ -3440,7 +3440,7 @@ gb_internal i64 type_align_of_internal(Type *t, TypePath *path) {
case Type_Tuple: {
- MUTEX_GUARD(&g_type_mutex);
+ // MUTEX_GUARD(&g_type_mutex);
i64 max = 1;
for_array(i, t->Tuple.variables) {
@@ -3465,7 +3465,7 @@ gb_internal i64 type_align_of_internal(Type *t, TypePath *path) {
return gb_max(t->Union.custom_align, 1);
}
- MUTEX_GUARD(&g_type_mutex);
+ // MUTEX_GUARD(&g_type_mutex);
i64 max = 1;
for_array(i, t->Union.variants) {
@@ -3492,7 +3492,7 @@ gb_internal i64 type_align_of_internal(Type *t, TypePath *path) {
return 1;
}
- MUTEX_GUARD(&g_type_mutex);
+ // MUTEX_GUARD(&g_type_mutex);
i64 max = 1;
for_array(i, t->Struct.fields) {
@@ -3605,7 +3605,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) {
switch (t->kind) {
case Type_Named: {
- MUTEX_GUARD(&g_type_mutex);
+ // MUTEX_GUARD(&g_type_mutex);
bool pop = type_path_push(path, t);
if (path->failure) {
@@ -3644,7 +3644,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) {
return build_context.word_size*2;
case Type_Array: {
- MUTEX_GUARD(&g_type_mutex);
+ // MUTEX_GUARD(&g_type_mutex);
i64 count, align, size, alignment;
count = t->Array.count;
@@ -3661,7 +3661,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) {
} break;
case Type_EnumeratedArray: {
- MUTEX_GUARD(&g_type_mutex);
+ // MUTEX_GUARD(&g_type_mutex);
i64 count, align, size, alignment;
count = t->EnumeratedArray.count;
@@ -3695,7 +3695,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) {
return (1 + 1 + 2)*build_context.word_size;
case Type_Tuple: {
- MUTEX_GUARD(&g_type_mutex);
+ // MUTEX_GUARD(&g_type_mutex);
i64 count, align, size;
count = t->Tuple.variables.count;
@@ -3715,7 +3715,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) {
if (t->Union.variants.count == 0) {
return 0;
}
- MUTEX_GUARD(&g_type_mutex);
+ // MUTEX_GUARD(&g_type_mutex);
i64 align = type_align_of_internal(t, path);
if (path->failure) {
@@ -3754,7 +3754,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) {
case Type_Struct: {
- MUTEX_GUARD(&g_type_mutex);
+ // MUTEX_GUARD(&g_type_mutex);
if (t->Struct.is_raw_union) {
i64 count = t->Struct.fields.count;