*/
#include <urcu/compiler.h>
+#include <urcu/system.h>
+
+#ifdef __cplusplus
+extern "C" {
+#endif
#ifndef __SIZEOF_LONG__
#ifdef __powerpc64__
#endif
#endif
+#ifdef __NO_LWSYNC__
+#define LWSYNC_OPCODE "sync\n"
+#else
+#define LWSYNC_OPCODE "lwsync\n"
+#endif
+
#ifndef BITS_PER_LONG
#define BITS_PER_LONG (__SIZEOF_LONG__ * 8)
#endif
#define ILLEGAL_INSTR ".long 0xd00d00"
-#define uatomic_set(addr, v) \
-do { \
- ACCESS_ONCE(*(addr)) = (v); \
-} while (0)
-
-#define uatomic_read(addr) ACCESS_ONCE(*(addr))
+#define uatomic_set(addr, v) STORE_SHARED(*(addr), (v))
+#define uatomic_read(addr) LOAD_SHARED(*(addr))
/*
* Using a isync as second barrier for exchange to provide acquire semantic.
unsigned int result;
__asm__ __volatile__(
- "lwsync\n"
+ LWSYNC_OPCODE
"1:\t" "lwarx %0,0,%1\n" /* load and reserve */
"stwcx. %2,0,%1\n" /* else store conditional */
"bne- 1b\n" /* retry if lost reservation */
unsigned long result;
__asm__ __volatile__(
- "lwsync\n"
+ LWSYNC_OPCODE
"1:\t" "ldarx %0,0,%1\n" /* load and reserve */
"stdcx. %2,0,%1\n" /* else store conditional */
"bne- 1b\n" /* retry if lost reservation */
unsigned int old_val;
__asm__ __volatile__(
- "lwsync\n"
+ LWSYNC_OPCODE
"1:\t" "lwarx %0,0,%1\n" /* load and reserve */
"cmpd %0,%3\n" /* if load is not equal to */
"bne 2f\n" /* old, fail */
unsigned long old_val;
__asm__ __volatile__(
- "lwsync\n"
+ LWSYNC_OPCODE
"1:\t" "ldarx %0,0,%1\n" /* load and reserve */
"cmpd %0,%3\n" /* if load is not equal to */
"bne 2f\n" /* old, fail */
unsigned int result;
__asm__ __volatile__(
- "lwsync\n"
+ LWSYNC_OPCODE
"1:\t" "lwarx %0,0,%1\n" /* load and reserve */
"add %0,%2,%0\n" /* add val to value loaded */
"stwcx. %0,0,%1\n" /* store conditional */
unsigned long result;
__asm__ __volatile__(
- "lwsync\n"
+ LWSYNC_OPCODE
"1:\t" "ldarx %0,0,%1\n" /* load and reserve */
"add %0,%2,%0\n" /* add val to value loaded */
"stdcx. %0,0,%1\n" /* store conditional */
#define uatomic_inc(addr) uatomic_add((addr), 1)
#define uatomic_dec(addr) uatomic_add((addr), -1)
-#define URCU_CAS_AVAIL() 1
-#define compat_uatomic_cmpxchg(ptr, old, _new) uatomic_cmpxchg(ptr, old, _new)
+#ifdef __cplusplus
+}
+#endif
#endif /* _URCU_ARCH_UATOMIC_PPC_H */