Cleanup: remove trailing whitespaces at EOL
[urcu.git] / urcu / uatomic / x86.h
index 99b0e6c6b405191538ff6ab50de35320c47652ea..129a2f5670385880870912235a14a63a6d0ea99e 100644 (file)
@@ -1,7 +1,7 @@
 #ifndef _URCU_ARCH_UATOMIC_X86_H
 #define _URCU_ARCH_UATOMIC_X86_H
 
-/* 
+/*
  * Copyright (c) 1991-1994 by Xerox Corporation.  All rights reserved.
  * Copyright (c) 1996-1999 by Silicon Graphics.  All rights reserved.
  * Copyright (c) 1999-2004 Hewlett-Packard Development Company, L.P.
@@ -28,7 +28,7 @@
 
 #ifdef __cplusplus
 extern "C" {
-#endif 
+#endif
 
 /*
  * Derived from AO_compare_and_swap() and AO_test_and_set_full().
@@ -39,7 +39,7 @@ struct __uatomic_dummy {
 };
 #define __hp(x)        ((struct __uatomic_dummy *)(x))
 
-#define _uatomic_set(addr, v)  CMM_STORE_SHARED(*(addr), (v))
+#define _uatomic_set(addr, v)  ((void) CMM_STORE_SHARED(*(addr), (v)))
 
 /* cmpxchg */
 
@@ -95,8 +95,10 @@ unsigned long __uatomic_cmpxchg(void *addr, unsigned long old,
        }
 #endif
        }
-       /* generate an illegal instruction. Cannot catch this with linker tricks
-        * when optimizations are disabled. */
+       /*
+        * generate an illegal instruction. Cannot catch this with
+        * linker tricks when optimizations are disabled.
+        */
        __asm__ __volatile__("ud2");
        return 0;
 }
@@ -157,8 +159,10 @@ unsigned long __uatomic_exchange(void *addr, unsigned long val, int len)
        }
 #endif
        }
-       /* generate an illegal instruction. Cannot catch this with linker tricks
-        * when optimizations are disabled. */
+       /*
+        * generate an illegal instruction. Cannot catch this with
+        * linker tricks when optimizations are disabled.
+        */
        __asm__ __volatile__("ud2");
        return 0;
 }
@@ -222,8 +226,10 @@ unsigned long __uatomic_add_return(void *addr, unsigned long val,
        }
 #endif
        }
-       /* generate an illegal instruction. Cannot catch this with linker tricks
-        * when optimizations are disabled. */
+       /*
+        * generate an illegal instruction. Cannot catch this with
+        * linker tricks when optimizations are disabled.
+        */
        __asm__ __volatile__("ud2");
        return 0;
 }
@@ -278,8 +284,10 @@ void __uatomic_and(void *addr, unsigned long val, int len)
        }
 #endif
        }
-       /* generate an illegal instruction. Cannot catch this with linker tricks
-        * when optimizations are disabled. */
+       /*
+        * generate an illegal instruction. Cannot catch this with
+        * linker tricks when optimizations are disabled.
+        */
        __asm__ __volatile__("ud2");
        return;
 }
@@ -332,8 +340,10 @@ void __uatomic_or(void *addr, unsigned long val, int len)
        }
 #endif
        }
-       /* generate an illegal instruction. Cannot catch this with linker tricks
-        * when optimizations are disabled. */
+       /*
+        * generate an illegal instruction. Cannot catch this with
+        * linker tricks when optimizations are disabled.
+        */
        __asm__ __volatile__("ud2");
        return;
 }
@@ -386,8 +396,10 @@ void __uatomic_add(void *addr, unsigned long val, int len)
        }
 #endif
        }
-       /* generate an illegal instruction. Cannot catch this with linker tricks
-        * when optimizations are disabled. */
+       /*
+        * generate an illegal instruction. Cannot catch this with
+        * linker tricks when optimizations are disabled.
+        */
        __asm__ __volatile__("ud2");
        return;
 }
@@ -494,8 +506,10 @@ void __uatomic_dec(void *addr, int len)
        }
 #endif
        }
-       /* generate an illegal instruction. Cannot catch this with linker tricks
-        * when optimizations are disabled. */
+       /*
+        * generate an illegal instruction. Cannot catch this with
+        * linker tricks when optimizations are disabled.
+        */
        __asm__ __volatile__("ud2");
        return;
 }
@@ -515,12 +529,16 @@ extern int __rcu_cas_init(void);
                                : (compat_uatomic_##insn))                      \
                        : (compat_uatomic_##insn))))
 
+/*
+ * We leave the return value so we don't break the ABI, but remove the
+ * return value from the API.
+ */
 extern unsigned long _compat_uatomic_set(void *addr,
                                         unsigned long _new, int len);
 #define compat_uatomic_set(addr, _new)                                        \
-       ((__typeof__(*(addr))) _compat_uatomic_set((addr),                     \
-                                               caa_cast_long_keep_sign(_new), \
-                                               sizeof(*(addr))))
+       ((void) _compat_uatomic_set((addr),                                    \
+                               caa_cast_long_keep_sign(_new),                 \
+                               sizeof(*(addr))))
 
 
 extern unsigned long _compat_uatomic_xchg(void *addr,
@@ -579,30 +597,30 @@ extern unsigned long _compat_uatomic_add_return(void *addr,
 
 #define uatomic_and(addr, v)           \
                UATOMIC_COMPAT(and(addr, v))
-#define cmm_smp_mb__before_and()       cmm_barrier()
-#define cmm_smp_mb__after_and()                cmm_barrier()
+#define cmm_smp_mb__before_uatomic_and()       cmm_barrier()
+#define cmm_smp_mb__after_uatomic_and()                cmm_barrier()
 
 #define uatomic_or(addr, v)            \
                UATOMIC_COMPAT(or(addr, v))
-#define cmm_smp_mb__before_or()                cmm_barrier()
-#define cmm_smp_mb__after_or()         cmm_barrier()
+#define cmm_smp_mb__before_uatomic_or()                cmm_barrier()
+#define cmm_smp_mb__after_uatomic_or()         cmm_barrier()
 
 #define uatomic_add_return(addr, v)            \
                UATOMIC_COMPAT(add_return(addr, v))
 
 #define uatomic_add(addr, v)   UATOMIC_COMPAT(add(addr, v))
-#define cmm_smp_mb__before_add()       cmm_barrier()
-#define cmm_smp_mb__after_add()                cmm_barrier()
+#define cmm_smp_mb__before_uatomic_add()       cmm_barrier()
+#define cmm_smp_mb__after_uatomic_add()                cmm_barrier()
 
 #define uatomic_inc(addr)      UATOMIC_COMPAT(inc(addr))
-#define cmm_smp_mb__before_inc()       cmm_barrier()
-#define cmm_smp_mb__after_inc()                cmm_barrier()
+#define cmm_smp_mb__before_uatomic_inc()       cmm_barrier()
+#define cmm_smp_mb__after_uatomic_inc()                cmm_barrier()
 
 #define uatomic_dec(addr)      UATOMIC_COMPAT(dec(addr))
-#define cmm_smp_mb__before_dec()       cmm_barrier()
-#define cmm_smp_mb__after_dec()                cmm_barrier()
+#define cmm_smp_mb__before_uatomic_dec()       cmm_barrier()
+#define cmm_smp_mb__after_uatomic_dec()                cmm_barrier()
 
-#ifdef __cplusplus 
+#ifdef __cplusplus
 }
 #endif
 
This page took 0.025771 seconds and 4 git commands to generate.