summaryrefslogtreecommitdiffstats
path: root/lib/libpthread/arch/powerpc/_atomic_lock.c
diff options
context:
space:
mode:
authord <d@openbsd.org>1998-12-21 07:22:26 +0000
committerd <d@openbsd.org>1998-12-21 07:22:26 +0000
commitba8540c203c0a63ab3dd47f90358c2cea0a9832d (patch)
tree9acab3253dacb0daa59844f33a9ce921d9cea8ab /lib/libpthread/arch/powerpc/_atomic_lock.c
parentclean (diff)
downloadwireguard-openbsd-ba8540c203c0a63ab3dd47f90358c2cea0a9832d.tar.xz
wireguard-openbsd-ba8540c203c0a63ab3dd47f90358c2cea0a9832d.zip
unfinished powerpc md stuff.. rahnds?
Diffstat (limited to 'lib/libpthread/arch/powerpc/_atomic_lock.c')
-rw-r--r--lib/libpthread/arch/powerpc/_atomic_lock.c43
1 files changed, 43 insertions, 0 deletions
diff --git a/lib/libpthread/arch/powerpc/_atomic_lock.c b/lib/libpthread/arch/powerpc/_atomic_lock.c
new file mode 100644
index 00000000000..ab1a5321763
--- /dev/null
+++ b/lib/libpthread/arch/powerpc/_atomic_lock.c
@@ -0,0 +1,43 @@
+/* $OpenBSD: _atomic_lock.c,v 1.1 1998/12/21 07:22:26 d Exp $ */
+/*
+ * Atomic lock for powerpc
+ */
+
+#include "spinlock.h"
+
+int
+_atomic_lock(volatile _spinlock_lock_t *lock)
+{
+ _spinlock_lock_t old;
+
+ __asm__("1: lwarx %0,0,%1 \n"
+ " stwcx. %2,0,%1 \n"
+ " bne- 1b \n"
+ : "=r" (old), "=r" (lock)
+ : "r" (_SPINLOCK_LOCKED), "1" (lock)
+ );
+
+ return (old != _SPINLOCK_UNLOCKED);
+
+ /*
+ * Dale <rahnds@openbsd.org> sez:
+ * Side note. to prevent two processes from accessing
+ * the same address with the lwarx in one instrution
+ * and the stwcx in another process, the current powerpc
+ * kernel uses a lwarx instruction without the corresponding
+ * stwcx which effectively causes any reservation of a
+ * process to be removed. if a context switch occurs
+ * between the two accesses the store will not occur
+ * and the condition code will cause it to loop. If on
+ * a dual processor machine, the reserve will cause
+ * appropriate bus cycle accesses to notify other
+ * processors.
+ */
+}
+
+int
+_atomic_is_locked(volatile _spinlock_lock_t *lock)
+{
+
+ return (*lock != _SPINLOCK_UNLOCKED);
+}