X-Git-Url: http://git.liburcu.org/?p=urcu.git;a=blobdiff_plain;f=urcu.h;h=af7d75cb7bc9e389ba9896a9207695d4feea9540;hp=01a4c6857d7e6fd31668dec57615be03e94ccc91;hb=7abf2d0da3b328572a17c9457b33c3890b0ba58b;hpb=1430ee0bdca4cb454d534ef7fc84af3e0692f26b diff --git a/urcu.h b/urcu.h index 01a4c68..af7d75c 100644 --- a/urcu.h +++ b/urcu.h @@ -6,175 +6,91 @@ * * Userspace RCU header * - * Copyright February 2009 - Mathieu Desnoyers + * Copyright (c) 2009 Mathieu Desnoyers + * Copyright (c) 2009 Paul E. McKenney, IBM Corporation. * - * Credits for Paul e. McKenney - * for inspiration coming from the Linux kernel RCU and rcu-preempt. + * LGPL-compatible code should include this header with : * - * The barrier, mb, rmb, wmb, atomic_inc, smp_read_barrier_depends, ACCESS_ONCE - * and rcu_dereference primitives come from the Linux kernel. + * #define _LGPL_SOURCE + * #include * - * Distributed under GPLv2 + * This library is free software; you can redistribute it and/or + * modify it under the terms of the GNU Lesser General Public + * License as published by the Free Software Foundation; either + * version 2.1 of the License, or (at your option) any later version. + * + * This library is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Lesser General Public License for more details. + * + * You should have received a copy of the GNU Lesser General Public + * License along with this library; if not, write to the Free Software + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + * + * IBM's contributions to this file may be relicensed under LGPLv2 or later. */ #include - -/* The "volatile" is due to gcc bugs */ -#define barrier() __asm__ __volatile__("": : :"memory") - -/* x86 32/64 specific */ -#define mb() asm volatile("mfence":::"memory") -#define rmb() asm volatile("lfence":::"memory") -#define wmb() asm volatile("sfence" ::: "memory") - -static inline void atomic_inc(int *v) -{ - asm volatile("lock; incl %0" - : "+m" (*v)); -} - -/* Nop everywhere except on alpha. */ -#define smp_read_barrier_depends() +#include /* - * Prevent the compiler from merging or refetching accesses. The compiler - * is also forbidden from reordering successive instances of ACCESS_ONCE(), - * but only when the compiler is aware of some particular ordering. One way - * to make the compiler aware of ordering is to put the two invocations of - * ACCESS_ONCE() in different C statements. + * Important ! * - * This macro does absolutely -nothing- to prevent the CPU from reordering, - * merging, or refetching absolutely anything at any time. Its main intended - * use is to mediate communication between process-level code and irq/NMI - * handlers, all running on the same CPU. + * Each thread containing read-side critical sections must be registered + * with rcu_register_thread() before calling rcu_read_lock(). + * rcu_unregister_thread() should be called before the thread exits. */ -#define ACCESS_ONCE(x) (*(volatile typeof(x) *)&(x)) -/** - * rcu_dereference - fetch an RCU-protected pointer in an - * RCU read-side critical section. This pointer may later - * be safely dereferenced. - * - * Inserts memory barriers on architectures that require them - * (currently only the Alpha), and, more importantly, documents - * exactly which pointers are protected by RCU. +#ifdef _LGPL_SOURCE + +#include + +/* + * Mappings for static use of the userspace RCU library. + * Should only be used in LGPL-compatible code. */ -#define rcu_dereference(p) ({ \ - typeof(p) _________p1 = ACCESS_ONCE(p); \ - smp_read_barrier_depends(); \ - (_________p1); \ - }) +#define rcu_dereference _rcu_dereference +#define rcu_read_lock _rcu_read_lock +#define rcu_read_unlock _rcu_read_unlock -#define SIGURCU SIGUSR1 +#define rcu_assign_pointer _rcu_assign_pointer +#define rcu_xchg_pointer _rcu_xchg_pointer +#define rcu_publish_content _rcu_publish_content -#ifdef DEBUG_YIELD -#include +#else /* !_LGPL_SOURCE */ -#define YIELD_READ (1 << 0) -#define YIELD_WRITE (1 << 1) +/* + * library wrappers to be used by non-LGPL compatible source code. + */ -extern unsigned int yield_active; -extern unsigned int __thread rand_yield; +extern void rcu_read_lock(void); +extern void rcu_read_unlock(void); -static inline void debug_yield_read(void) -{ - if (yield_active & YIELD_READ) - if (rand_r(&rand_yield) & 0x1) - sched_yield(); -} +extern void *rcu_dereference(void *p); -static inline void debug_yield_write(void) -{ - if (yield_active & YIELD_WRITE) - if (rand_r(&rand_yield) & 0x1) - sched_yield(); -} +extern void *rcu_assign_pointer_sym(void **p, void *v); -static inline void debug_yield_init(void) -{ - rand_yield = time(NULL) ^ pthread_self(); -} -#else -static inline void debug_yield_read(void) -{ -} +#define rcu_assign_pointer(p, v) \ + rcu_assign_pointer_sym((void **)(p), (v)) -static inline void debug_yield_write(void) -{ -} +extern void *rcu_xchg_pointer_sym(void **p, void *v); +#define rcu_xchg_pointer(p, v) \ + rcu_xchg_pointer_sym((void **)(p), (v)) -static inline void debug_yield_init(void) -{ +extern void *rcu_publish_content_sym(void **p, void *v); +#define rcu_publish_content(p, v) \ + rcu_publish_content_sym((void **)(p), (v)) -} -#endif +#endif /* !_LGPL_SOURCE */ -/* - * Limiting the nesting level to 256 to keep instructions small in the read - * fast-path. - */ -#define RCU_GP_COUNT (1U << 0) -#define RCU_GP_CTR_BIT (1U << 8) -#define RCU_GP_CTR_NEST_MASK (RCU_GP_CTR_BIT - 1) - -/* Global quiescent period counter with low-order bits unused. */ -extern int urcu_gp_ctr; - -extern int __thread urcu_active_readers; - -static inline int rcu_old_gp_ongoing(int *value) -{ - int v; - - if (value == NULL) - return 0; - debug_yield_write(); - v = ACCESS_ONCE(*value); - debug_yield_write(); - return (v & RCU_GP_CTR_NEST_MASK) && - ((v ^ ACCESS_ONCE(urcu_gp_ctr)) & RCU_GP_CTR_BIT); -} - -static inline void rcu_read_lock(void) -{ - int tmp; - - debug_yield_read(); - tmp = urcu_active_readers; - debug_yield_read(); - if (!(tmp & RCU_GP_CTR_NEST_MASK)) - urcu_active_readers = urcu_gp_ctr + RCU_GP_COUNT; - else - urcu_active_readers = tmp + RCU_GP_COUNT; - debug_yield_read(); - /* - * Increment active readers count before accessing the pointer. - * See force_mb_all_threads(). - */ - barrier(); - debug_yield_read(); -} - -static inline void rcu_read_unlock(void) -{ - debug_yield_read(); - barrier(); - debug_yield_read(); - /* - * Finish using rcu before decrementing the pointer. - * See force_mb_all_threads(). - */ - urcu_active_readers -= RCU_GP_COUNT; - debug_yield_read(); -} - -extern void *urcu_publish_content(void **ptr, void *new); +extern void synchronize_rcu(void); /* * Reader thread registration. */ -extern void urcu_register_thread(void); -extern void urcu_unregister_thread(void); +extern void rcu_register_thread(void); +extern void rcu_unregister_thread(void); #endif /* _URCU_H */