#endif
#ifndef uatomic_set
-#define uatomic_set(addr, v) CMM_STORE_SHARED(*(addr), (v))
+#define uatomic_set(addr, v) ((void) CMM_STORE_SHARED(*(addr), (v)))
#endif
#ifndef uatomic_read
void _uatomic_link_error()
{
#ifdef ILLEGAL_INSTR
- /* generate an illegal instruction. Cannot catch this with linker tricks
- * when optimizations are disabled. */
+ /*
+ * generate an illegal instruction. Cannot catch this with
+ * linker tricks when optimizations are disabled.
+ */
__asm__ __volatile__(ILLEGAL_INSTR);
#else
__builtin_trap ();
(_uatomic_and((addr), \
caa_cast_long_keep_sign(v), \
sizeof(*(addr))))
+#define cmm_smp_mb__before_uatomic_and() cmm_barrier()
+#define cmm_smp_mb__after_uatomic_and() cmm_barrier()
+
#endif
/* uatomic_or */
(_uatomic_or((addr), \
caa_cast_long_keep_sign(v), \
sizeof(*(addr))))
+#define cmm_smp_mb__before_uatomic_or() cmm_barrier()
+#define cmm_smp_mb__after_uatomic_or() cmm_barrier()
+
#endif
+
/* uatomic_add_return */
#ifndef uatomic_add_return
(_uatomic_and((addr), \
caa_cast_long_keep_sign(v), \
sizeof(*(addr))))
+#define cmm_smp_mb__before_uatomic_and() cmm_barrier()
+#define cmm_smp_mb__after_uatomic_and() cmm_barrier()
+
#endif /* #ifndef uatomic_and */
#ifndef uatomic_or
(_uatomic_or((addr), \
caa_cast_long_keep_sign(v), \
sizeof(*(addr))))
+#define cmm_smp_mb__before_uatomic_or() cmm_barrier()
+#define cmm_smp_mb__after_uatomic_or() cmm_barrier()
+
#endif /* #ifndef uatomic_or */
#ifndef uatomic_add_return
#ifndef uatomic_add
#define uatomic_add(addr, v) (void)uatomic_add_return((addr), (v))
+#define cmm_smp_mb__before_uatomic_add() cmm_barrier()
+#define cmm_smp_mb__after_uatomic_add() cmm_barrier()
#endif
#define uatomic_sub_return(addr, v) \
uatomic_add_return((addr), -(caa_cast_long_keep_sign(v)))
#define uatomic_sub(addr, v) \
uatomic_add((addr), -(caa_cast_long_keep_sign(v)))
+#define cmm_smp_mb__before_uatomic_sub() cmm_smp_mb__before_uatomic_add()
+#define cmm_smp_mb__after_uatomic_sub() cmm_smp_mb__after_uatomic_add()
#ifndef uatomic_inc
#define uatomic_inc(addr) uatomic_add((addr), 1)
+#define cmm_smp_mb__before_uatomic_inc() cmm_smp_mb__before_uatomic_add()
+#define cmm_smp_mb__after_uatomic_inc() cmm_smp_mb__after_uatomic_add()
#endif
#ifndef uatomic_dec
#define uatomic_dec(addr) uatomic_add((addr), -1)
+#define cmm_smp_mb__before_uatomic_dec() cmm_smp_mb__before_uatomic_add()
+#define cmm_smp_mb__after_uatomic_dec() cmm_smp_mb__after_uatomic_add()
#endif
#ifdef __cplusplus