Cleanup: remove trailing whitespaces at EOL
[urcu.git] / urcu / uatomic / x86.h
index e064b9ebd6e1bc798c85a0d464d42940cacf2d3e..129a2f5670385880870912235a14a63a6d0ea99e 100644 (file)
@@ -1,7 +1,7 @@
 #ifndef _URCU_ARCH_UATOMIC_X86_H
 #define _URCU_ARCH_UATOMIC_X86_H
 
-/* 
+/*
  * Copyright (c) 1991-1994 by Xerox Corporation.  All rights reserved.
  * Copyright (c) 1996-1999 by Silicon Graphics.  All rights reserved.
  * Copyright (c) 1999-2004 Hewlett-Packard Development Company, L.P.
@@ -28,7 +28,7 @@
 
 #ifdef __cplusplus
 extern "C" {
-#endif 
+#endif
 
 /*
  * Derived from AO_compare_and_swap() and AO_test_and_set_full().
@@ -39,7 +39,7 @@ struct __uatomic_dummy {
 };
 #define __hp(x)        ((struct __uatomic_dummy *)(x))
 
-#define _uatomic_set(addr, v)  CMM_STORE_SHARED(*(addr), (v))
+#define _uatomic_set(addr, v)  ((void) CMM_STORE_SHARED(*(addr), (v)))
 
 /* cmpxchg */
 
@@ -95,15 +95,18 @@ unsigned long __uatomic_cmpxchg(void *addr, unsigned long old,
        }
 #endif
        }
-       /* generate an illegal instruction. Cannot catch this with linker tricks
-        * when optimizations are disabled. */
+       /*
+        * generate an illegal instruction. Cannot catch this with
+        * linker tricks when optimizations are disabled.
+        */
        __asm__ __volatile__("ud2");
        return 0;
 }
 
 #define _uatomic_cmpxchg(addr, old, _new)                                    \
-       ((__typeof__(*(addr))) __uatomic_cmpxchg((addr), (unsigned long)(old),\
-                                               (unsigned long)(_new),        \
+       ((__typeof__(*(addr))) __uatomic_cmpxchg((addr),                      \
+                                               caa_cast_long_keep_sign(old), \
+                                               caa_cast_long_keep_sign(_new),\
                                                sizeof(*(addr))))
 
 /* xchg */
@@ -156,14 +159,17 @@ unsigned long __uatomic_exchange(void *addr, unsigned long val, int len)
        }
 #endif
        }
-       /* generate an illegal instruction. Cannot catch this with linker tricks
-        * when optimizations are disabled. */
+       /*
+        * generate an illegal instruction. Cannot catch this with
+        * linker tricks when optimizations are disabled.
+        */
        __asm__ __volatile__("ud2");
        return 0;
 }
 
 #define _uatomic_xchg(addr, v)                                               \
-       ((__typeof__(*(addr))) __uatomic_exchange((addr), (unsigned long)(v), \
+       ((__typeof__(*(addr))) __uatomic_exchange((addr),                     \
+                                               caa_cast_long_keep_sign(v),   \
                                                sizeof(*(addr))))
 
 /* uatomic_add_return */
@@ -220,16 +226,18 @@ unsigned long __uatomic_add_return(void *addr, unsigned long val,
        }
 #endif
        }
-       /* generate an illegal instruction. Cannot catch this with linker tricks
-        * when optimizations are disabled. */
+       /*
+        * generate an illegal instruction. Cannot catch this with
+        * linker tricks when optimizations are disabled.
+        */
        __asm__ __volatile__("ud2");
        return 0;
 }
 
-#define _uatomic_add_return(addr, v)                                   \
-       ((__typeof__(*(addr))) __uatomic_add_return((addr),             \
-                                                 (unsigned long)(v),   \
-                                                 sizeof(*(addr))))
+#define _uatomic_add_return(addr, v)                                       \
+       ((__typeof__(*(addr))) __uatomic_add_return((addr),                 \
+                                               caa_cast_long_keep_sign(v), \
+                                               sizeof(*(addr))))
 
 /* uatomic_and */
 
@@ -276,14 +284,16 @@ void __uatomic_and(void *addr, unsigned long val, int len)
        }
 #endif
        }
-       /* generate an illegal instruction. Cannot catch this with linker tricks
-        * when optimizations are disabled. */
+       /*
+        * generate an illegal instruction. Cannot catch this with
+        * linker tricks when optimizations are disabled.
+        */
        __asm__ __volatile__("ud2");
        return;
 }
 
 #define _uatomic_and(addr, v)                                             \
-       (__uatomic_and((addr), (unsigned long)(v), sizeof(*(addr))))
+       (__uatomic_and((addr), caa_cast_long_keep_sign(v), sizeof(*(addr))))
 
 /* uatomic_or */
 
@@ -330,14 +340,16 @@ void __uatomic_or(void *addr, unsigned long val, int len)
        }
 #endif
        }
-       /* generate an illegal instruction. Cannot catch this with linker tricks
-        * when optimizations are disabled. */
+       /*
+        * generate an illegal instruction. Cannot catch this with
+        * linker tricks when optimizations are disabled.
+        */
        __asm__ __volatile__("ud2");
        return;
 }
 
 #define _uatomic_or(addr, v)                                              \
-       (__uatomic_or((addr), (unsigned long)(v), sizeof(*(addr))))
+       (__uatomic_or((addr), caa_cast_long_keep_sign(v), sizeof(*(addr))))
 
 /* uatomic_add */
 
@@ -384,14 +396,16 @@ void __uatomic_add(void *addr, unsigned long val, int len)
        }
 #endif
        }
-       /* generate an illegal instruction. Cannot catch this with linker tricks
-        * when optimizations are disabled. */
+       /*
+        * generate an illegal instruction. Cannot catch this with
+        * linker tricks when optimizations are disabled.
+        */
        __asm__ __volatile__("ud2");
        return;
 }
 
 #define _uatomic_add(addr, v)                                             \
-       (__uatomic_add((addr), (unsigned long)(v), sizeof(*(addr))))
+       (__uatomic_add((addr), caa_cast_long_keep_sign(v), sizeof(*(addr))))
 
 
 /* uatomic_inc */
@@ -492,8 +506,10 @@ void __uatomic_dec(void *addr, int len)
        }
 #endif
        }
-       /* generate an illegal instruction. Cannot catch this with linker tricks
-        * when optimizations are disabled. */
+       /*
+        * generate an illegal instruction. Cannot catch this with
+        * linker tricks when optimizations are disabled.
+        */
        __asm__ __volatile__("ud2");
        return;
 }
@@ -505,55 +521,59 @@ extern int __rcu_cas_avail;
 extern int __rcu_cas_init(void);
 
 #define UATOMIC_COMPAT(insn)                                                   \
-       ((likely(__rcu_cas_avail > 0))                                          \
+       ((caa_likely(__rcu_cas_avail > 0))                                              \
        ? (_uatomic_##insn)                                                     \
-               : ((unlikely(__rcu_cas_avail < 0)                               \
+               : ((caa_unlikely(__rcu_cas_avail < 0)                           \
                        ? ((__rcu_cas_init() > 0)                               \
                                ? (_uatomic_##insn)                             \
                                : (compat_uatomic_##insn))                      \
                        : (compat_uatomic_##insn))))
 
+/*
+ * We leave the return value so we don't break the ABI, but remove the
+ * return value from the API.
+ */
 extern unsigned long _compat_uatomic_set(void *addr,
                                         unsigned long _new, int len);
 #define compat_uatomic_set(addr, _new)                                        \
-       ((__typeof__(*(addr))) _compat_uatomic_set((addr),                     \
-                                               (unsigned long)(_new),         \
-                                               sizeof(*(addr))))
+       ((void) _compat_uatomic_set((addr),                                    \
+                               caa_cast_long_keep_sign(_new),                 \
+                               sizeof(*(addr))))
 
 
 extern unsigned long _compat_uatomic_xchg(void *addr,
                                          unsigned long _new, int len);
 #define compat_uatomic_xchg(addr, _new)                                               \
        ((__typeof__(*(addr))) _compat_uatomic_xchg((addr),                    \
-                                               (unsigned long)(_new),         \
+                                               caa_cast_long_keep_sign(_new), \
                                                sizeof(*(addr))))
 
 extern unsigned long _compat_uatomic_cmpxchg(void *addr, unsigned long old,
                                             unsigned long _new, int len);
 #define compat_uatomic_cmpxchg(addr, old, _new)                                       \
        ((__typeof__(*(addr))) _compat_uatomic_cmpxchg((addr),                 \
-                                               (unsigned long)(old),          \
-                                               (unsigned long)(_new),         \
+                                               caa_cast_long_keep_sign(old),  \
+                                               caa_cast_long_keep_sign(_new), \
                                                sizeof(*(addr))))
 
 extern void _compat_uatomic_and(void *addr, unsigned long _new, int len);
 #define compat_uatomic_and(addr, v)                                   \
        (_compat_uatomic_and((addr),                                   \
-                       (unsigned long)(v),                            \
+                       caa_cast_long_keep_sign(v),                    \
                        sizeof(*(addr))))
 
 extern void _compat_uatomic_or(void *addr, unsigned long _new, int len);
 #define compat_uatomic_or(addr, v)                                    \
        (_compat_uatomic_or((addr),                                    \
-                         (unsigned long)(v),                          \
+                         caa_cast_long_keep_sign(v),                  \
                          sizeof(*(addr))))
 
 extern unsigned long _compat_uatomic_add_return(void *addr,
                                                unsigned long _new, int len);
-#define compat_uatomic_add_return(addr, v)                            \
-       ((__typeof__(*(addr))) _compat_uatomic_add_return((addr),      \
-                                                  (unsigned long)(v), \
-                                                  sizeof(*(addr))))
+#define compat_uatomic_add_return(addr, v)                                 \
+       ((__typeof__(*(addr))) _compat_uatomic_add_return((addr),           \
+                                               caa_cast_long_keep_sign(v), \
+                                               sizeof(*(addr))))
 
 #define compat_uatomic_add(addr, v)                                           \
                ((void)compat_uatomic_add_return((addr), (v)))
@@ -574,18 +594,33 @@ extern unsigned long _compat_uatomic_add_return(void *addr,
                UATOMIC_COMPAT(cmpxchg(addr, old, _new))
 #define uatomic_xchg(addr, v)                  \
                UATOMIC_COMPAT(xchg(addr, v))
+
 #define uatomic_and(addr, v)           \
                UATOMIC_COMPAT(and(addr, v))
+#define cmm_smp_mb__before_uatomic_and()       cmm_barrier()
+#define cmm_smp_mb__after_uatomic_and()                cmm_barrier()
+
 #define uatomic_or(addr, v)            \
                UATOMIC_COMPAT(or(addr, v))
+#define cmm_smp_mb__before_uatomic_or()                cmm_barrier()
+#define cmm_smp_mb__after_uatomic_or()         cmm_barrier()
+
 #define uatomic_add_return(addr, v)            \
                UATOMIC_COMPAT(add_return(addr, v))
 
 #define uatomic_add(addr, v)   UATOMIC_COMPAT(add(addr, v))
+#define cmm_smp_mb__before_uatomic_add()       cmm_barrier()
+#define cmm_smp_mb__after_uatomic_add()                cmm_barrier()
+
 #define uatomic_inc(addr)      UATOMIC_COMPAT(inc(addr))
+#define cmm_smp_mb__before_uatomic_inc()       cmm_barrier()
+#define cmm_smp_mb__after_uatomic_inc()                cmm_barrier()
+
 #define uatomic_dec(addr)      UATOMIC_COMPAT(dec(addr))
+#define cmm_smp_mb__before_uatomic_dec()       cmm_barrier()
+#define cmm_smp_mb__after_uatomic_dec()                cmm_barrier()
 
-#ifdef __cplusplus 
+#ifdef __cplusplus
 }
 #endif
 
This page took 0.026772 seconds and 4 git commands to generate.