X-Git-Url: https://git.liburcu.org/?a=blobdiff_plain;f=include%2Furcu%2Fuatomic%2Fx86.h;h=e9f2f78275f472618d2ad302205eda8bb37629c8;hb=835b9ab3ca3777fe42e37e92096226ebd19ca75b;hp=2a4ea1c032a6019f0257ca967200e28fee27edaa;hpb=375db287452ed4ec9e062e6d670d4cc198a8253d;p=urcu.git diff --git a/include/urcu/uatomic/x86.h b/include/urcu/uatomic/x86.h index 2a4ea1c..e9f2f78 100644 --- a/include/urcu/uatomic/x86.h +++ b/include/urcu/uatomic/x86.h @@ -20,6 +20,7 @@ * Boehm-Demers-Weiser conservative garbage collector. */ +#include #include #include #include @@ -35,10 +36,14 @@ extern "C" { * Derived from AO_compare_and_swap() and AO_test_and_set_full(). */ -struct __uatomic_dummy { - unsigned long v[10]; -}; -#define __hp(x) ((struct __uatomic_dummy *)(x)) +/* + * The __hp() macro casts the void pointer "x" to a pointer to a structure + * containing an array of char of the specified size. This allows passing the + * @addr arguments of the following inline functions as "m" and "+m" operands + * to the assembly. + */ + +#define __hp(size, x) ((struct { char v[size]; } *)(x)) #define _uatomic_set(addr, v) ((void) CMM_STORE_SHARED(*(addr), (v))) @@ -55,7 +60,7 @@ unsigned long __uatomic_cmpxchg(void *addr, unsigned long old, __asm__ __volatile__( "lock; cmpxchgb %2, %1" - : "+a"(result), "+m"(*__hp(addr)) + : "+a"(result), "+m"(*__hp(len, addr)) : "q"((unsigned char)_new) : "memory"); return result; @@ -66,7 +71,7 @@ unsigned long __uatomic_cmpxchg(void *addr, unsigned long old, __asm__ __volatile__( "lock; cmpxchgw %2, %1" - : "+a"(result), "+m"(*__hp(addr)) + : "+a"(result), "+m"(*__hp(len, addr)) : "r"((unsigned short)_new) : "memory"); return result; @@ -77,7 +82,7 @@ unsigned long __uatomic_cmpxchg(void *addr, unsigned long old, __asm__ __volatile__( "lock; cmpxchgl %2, %1" - : "+a"(result), "+m"(*__hp(addr)) + : "+a"(result), "+m"(*__hp(len, addr)) : "r"((unsigned int)_new) : "memory"); return result; @@ -89,7 +94,7 @@ unsigned long __uatomic_cmpxchg(void *addr, unsigned long old, __asm__ __volatile__( "lock; cmpxchgq %2, %1" - : "+a"(result), "+m"(*__hp(addr)) + : "+a"(result), "+m"(*__hp(len, addr)) : "r"((unsigned long)_new) : "memory"); return result; @@ -122,7 +127,7 @@ unsigned long __uatomic_exchange(void *addr, unsigned long val, int len) unsigned char result; __asm__ __volatile__( "xchgb %0, %1" - : "=q"(result), "+m"(*__hp(addr)) + : "=q"(result), "+m"(*__hp(len, addr)) : "0" ((unsigned char)val) : "memory"); return result; @@ -132,7 +137,7 @@ unsigned long __uatomic_exchange(void *addr, unsigned long val, int len) unsigned short result; __asm__ __volatile__( "xchgw %0, %1" - : "=r"(result), "+m"(*__hp(addr)) + : "=r"(result), "+m"(*__hp(len, addr)) : "0" ((unsigned short)val) : "memory"); return result; @@ -142,7 +147,7 @@ unsigned long __uatomic_exchange(void *addr, unsigned long val, int len) unsigned int result; __asm__ __volatile__( "xchgl %0, %1" - : "=r"(result), "+m"(*__hp(addr)) + : "=r"(result), "+m"(*__hp(len, addr)) : "0" ((unsigned int)val) : "memory"); return result; @@ -153,7 +158,7 @@ unsigned long __uatomic_exchange(void *addr, unsigned long val, int len) unsigned long result; __asm__ __volatile__( "xchgq %0, %1" - : "=r"(result), "+m"(*__hp(addr)) + : "=r"(result), "+m"(*__hp(len, addr)) : "0" ((unsigned long)val) : "memory"); return result; @@ -186,7 +191,7 @@ unsigned long __uatomic_add_return(void *addr, unsigned long val, __asm__ __volatile__( "lock; xaddb %1, %0" - : "+m"(*__hp(addr)), "+q" (result) + : "+m"(*__hp(len, addr)), "+q" (result) : : "memory"); return result + (unsigned char)val; @@ -197,7 +202,7 @@ unsigned long __uatomic_add_return(void *addr, unsigned long val, __asm__ __volatile__( "lock; xaddw %1, %0" - : "+m"(*__hp(addr)), "+r" (result) + : "+m"(*__hp(len, addr)), "+r" (result) : : "memory"); return result + (unsigned short)val; @@ -208,7 +213,7 @@ unsigned long __uatomic_add_return(void *addr, unsigned long val, __asm__ __volatile__( "lock; xaddl %1, %0" - : "+m"(*__hp(addr)), "+r" (result) + : "+m"(*__hp(len, addr)), "+r" (result) : : "memory"); return result + (unsigned int)val; @@ -220,7 +225,7 @@ unsigned long __uatomic_add_return(void *addr, unsigned long val, __asm__ __volatile__( "lock; xaddq %1, %0" - : "+m"(*__hp(addr)), "+r" (result) + : "+m"(*__hp(len, addr)), "+r" (result) : : "memory"); return result + (unsigned long)val; @@ -250,7 +255,7 @@ void __uatomic_and(void *addr, unsigned long val, int len) { __asm__ __volatile__( "lock; andb %1, %0" - : "=m"(*__hp(addr)) + : "=m"(*__hp(len, addr)) : "iq" ((unsigned char)val) : "memory"); return; @@ -259,7 +264,7 @@ void __uatomic_and(void *addr, unsigned long val, int len) { __asm__ __volatile__( "lock; andw %1, %0" - : "=m"(*__hp(addr)) + : "=m"(*__hp(len, addr)) : "ir" ((unsigned short)val) : "memory"); return; @@ -268,7 +273,7 @@ void __uatomic_and(void *addr, unsigned long val, int len) { __asm__ __volatile__( "lock; andl %1, %0" - : "=m"(*__hp(addr)) + : "=m"(*__hp(len, addr)) : "ir" ((unsigned int)val) : "memory"); return; @@ -278,7 +283,7 @@ void __uatomic_and(void *addr, unsigned long val, int len) { __asm__ __volatile__( "lock; andq %1, %0" - : "=m"(*__hp(addr)) + : "=m"(*__hp(len, addr)) : "er" ((unsigned long)val) : "memory"); return; @@ -306,7 +311,7 @@ void __uatomic_or(void *addr, unsigned long val, int len) { __asm__ __volatile__( "lock; orb %1, %0" - : "=m"(*__hp(addr)) + : "=m"(*__hp(len, addr)) : "iq" ((unsigned char)val) : "memory"); return; @@ -315,7 +320,7 @@ void __uatomic_or(void *addr, unsigned long val, int len) { __asm__ __volatile__( "lock; orw %1, %0" - : "=m"(*__hp(addr)) + : "=m"(*__hp(len, addr)) : "ir" ((unsigned short)val) : "memory"); return; @@ -324,7 +329,7 @@ void __uatomic_or(void *addr, unsigned long val, int len) { __asm__ __volatile__( "lock; orl %1, %0" - : "=m"(*__hp(addr)) + : "=m"(*__hp(len, addr)) : "ir" ((unsigned int)val) : "memory"); return; @@ -334,7 +339,7 @@ void __uatomic_or(void *addr, unsigned long val, int len) { __asm__ __volatile__( "lock; orq %1, %0" - : "=m"(*__hp(addr)) + : "=m"(*__hp(len, addr)) : "er" ((unsigned long)val) : "memory"); return; @@ -362,7 +367,7 @@ void __uatomic_add(void *addr, unsigned long val, int len) { __asm__ __volatile__( "lock; addb %1, %0" - : "=m"(*__hp(addr)) + : "=m"(*__hp(len, addr)) : "iq" ((unsigned char)val) : "memory"); return; @@ -371,7 +376,7 @@ void __uatomic_add(void *addr, unsigned long val, int len) { __asm__ __volatile__( "lock; addw %1, %0" - : "=m"(*__hp(addr)) + : "=m"(*__hp(len, addr)) : "ir" ((unsigned short)val) : "memory"); return; @@ -380,7 +385,7 @@ void __uatomic_add(void *addr, unsigned long val, int len) { __asm__ __volatile__( "lock; addl %1, %0" - : "=m"(*__hp(addr)) + : "=m"(*__hp(len, addr)) : "ir" ((unsigned int)val) : "memory"); return; @@ -390,7 +395,7 @@ void __uatomic_add(void *addr, unsigned long val, int len) { __asm__ __volatile__( "lock; addq %1, %0" - : "=m"(*__hp(addr)) + : "=m"(*__hp(len, addr)) : "er" ((unsigned long)val) : "memory"); return; @@ -419,7 +424,7 @@ void __uatomic_inc(void *addr, int len) { __asm__ __volatile__( "lock; incb %0" - : "=m"(*__hp(addr)) + : "=m"(*__hp(len, addr)) : : "memory"); return; @@ -428,7 +433,7 @@ void __uatomic_inc(void *addr, int len) { __asm__ __volatile__( "lock; incw %0" - : "=m"(*__hp(addr)) + : "=m"(*__hp(len, addr)) : : "memory"); return; @@ -437,7 +442,7 @@ void __uatomic_inc(void *addr, int len) { __asm__ __volatile__( "lock; incl %0" - : "=m"(*__hp(addr)) + : "=m"(*__hp(len, addr)) : : "memory"); return; @@ -447,7 +452,7 @@ void __uatomic_inc(void *addr, int len) { __asm__ __volatile__( "lock; incq %0" - : "=m"(*__hp(addr)) + : "=m"(*__hp(len, addr)) : : "memory"); return; @@ -472,7 +477,7 @@ void __uatomic_dec(void *addr, int len) { __asm__ __volatile__( "lock; decb %0" - : "=m"(*__hp(addr)) + : "=m"(*__hp(len, addr)) : : "memory"); return; @@ -481,7 +486,7 @@ void __uatomic_dec(void *addr, int len) { __asm__ __volatile__( "lock; decw %0" - : "=m"(*__hp(addr)) + : "=m"(*__hp(len, addr)) : : "memory"); return; @@ -490,7 +495,7 @@ void __uatomic_dec(void *addr, int len) { __asm__ __volatile__( "lock; decl %0" - : "=m"(*__hp(addr)) + : "=m"(*__hp(len, addr)) : : "memory"); return; @@ -500,7 +505,7 @@ void __uatomic_dec(void *addr, int len) { __asm__ __volatile__( "lock; decq %0" - : "=m"(*__hp(addr)) + : "=m"(*__hp(len, addr)) : : "memory"); return; @@ -517,7 +522,11 @@ void __uatomic_dec(void *addr, int len) #define _uatomic_dec(addr) (__uatomic_dec((addr), sizeof(*(addr)))) -#if ((CAA_BITS_PER_LONG != 64) && defined(CONFIG_RCU_COMPAT_ARCH)) +#if ((CAA_BITS_PER_LONG != 64) && defined(URCU_ARCH_I386)) + +/* For backwards compat */ +#define CONFIG_RCU_COMPAT_ARCH 1 + extern int __rcu_cas_avail; extern int __rcu_cas_init(void);