summaryrefslogtreecommitdiff
path: root/kern/thread.c
diff options
context:
space:
mode:
authorRichard Braun <rbraun@sceen.net>2019-08-16 20:02:37 +0200
committerRichard Braun <rbraun@sceen.net>2019-08-16 20:30:48 +0200
commit84d47c358b1d07b9155474d7075e082c52147ebc (patch)
tree0a97fbcfe7f56637e18afc9e1ba4cefa3cea1aa5 /kern/thread.c
parent27dbd3b2fdafd6e12bdef5f653532b2ca752a6f9 (diff)
Replace some barrier()s with local atomic fences
Diffstat (limited to 'kern/thread.c')
-rw-r--r--kern/thread.c4
1 files changed, 3 insertions, 1 deletions
diff --git a/kern/thread.c b/kern/thread.c
index 71943c04..0bad0752 100644
--- a/kern/thread.c
+++ b/kern/thread.c
@@ -96,6 +96,7 @@
#include <kern/cpumap.h>
#include <kern/init.h>
#include <kern/kmem.h>
+#include <kern/latomic.h>
#include <kern/list.h>
#include <kern/macros.h>
#include <kern/panic.h>
@@ -674,7 +675,9 @@ thread_runq_schedule(struct thread_runq *runq)
* and locking the run queue again is equivalent to a full memory
* barrier.
*/
+ latomic_fence(LATOMIC_ACQ_REL);
tcb_switch(&prev->tcb, &next->tcb);
+ latomic_fence(LATOMIC_ACQ_REL);
/*
* The thread is dispatched on a processor once again.
@@ -685,7 +688,6 @@ thread_runq_schedule(struct thread_runq *runq)
* referenced any more.
* - The current thread may have been migrated to another processor.
*/
- barrier();
thread_runq_schedule_load(prev);
next = NULL;