#endif
#ifndef uatomic_set
-#define uatomic_set(addr, v) STORE_SHARED(*(addr), (v))
+#define uatomic_set(addr, v) CMM_STORE_SHARED(*(addr), (v))
#endif
#ifndef uatomic_read
-#define uatomic_read(addr) LOAD_SHARED(*(addr))
+#define uatomic_read(addr) CMM_LOAD_SHARED(*(addr))
#endif
#if !defined __OPTIMIZE__ || defined UATOMIC_NO_LINK_ERROR
#endif
case 4:
return __sync_val_compare_and_swap_4(addr, old, _new);
-#if (BITS_PER_LONG == 64)
+#if (CAA_BITS_PER_LONG == 64)
case 8:
return __sync_val_compare_and_swap_8(addr, old, _new);
#endif
#endif
case 4:
return __sync_add_and_fetch_4(addr, val);
-#if (BITS_PER_LONG == 64)
+#if (CAA_BITS_PER_LONG == 64)
case 8:
return __sync_add_and_fetch_8(addr, val);
#endif
return old;
}
-#if (BITS_PER_LONG == 64)
+#if (CAA_BITS_PER_LONG == 64)
case 8:
{
unsigned long old;
return old + val;
}
-#if (BITS_PER_LONG == 64)
+#if (CAA_BITS_PER_LONG == 64)
case 8:
{
unsigned long old, oldt;
return old;
}
-#if (BITS_PER_LONG == 64)
+#if (CAA_BITS_PER_LONG == 64)
case 8:
{
unsigned long old, oldt;