X-Git-Url: http://git.liburcu.org/?p=urcu.git;a=blobdiff_plain;f=urcu-defer.c;h=412eb30477908b2a60acff9dde794d9ecfc793f8;hp=00dd3c60b3bcf1a2ea7895b3612953a0203af6f6;hb=b55487c706982c8190e42545f7f74dfadad59548;hpb=804b437535f97b519db3e8715807950f5818cf27 diff --git a/urcu-defer.c b/urcu-defer.c index 00dd3c6..412eb30 100644 --- a/urcu-defer.c +++ b/urcu-defer.c @@ -116,19 +116,15 @@ static void rcu_defer_barrier_queue(struct defer_queue *queue, smp_rmb(); /* read head before q[]. */ p = LOAD_SHARED(queue->q[i++ & DEFER_QUEUE_MASK]); if (unlikely(DQ_IS_FCT_BIT(p))) { - //printf("%lu fct bit %p\n", i-1, p); DQ_CLEAR_FCT_BIT(p); queue->last_fct_out = p; p = LOAD_SHARED(queue->q[i++ & DEFER_QUEUE_MASK]); } else if (unlikely(p == DQ_FCT_MARK)) { - //printf("%lu fct mark %p\n", i-1, p); p = LOAD_SHARED(queue->q[i++ & DEFER_QUEUE_MASK]); queue->last_fct_out = p; p = LOAD_SHARED(queue->q[i++ & DEFER_QUEUE_MASK]); - }// else - //printf("%lu data %p\n", i-1, p); + } fct = queue->last_fct_out; - //printf("tid %lu %lu last_fct %p data %p\n", pthread_self(), i-1, fct, p); fct(p); } smp_mb(); /* push tail after having used q[] */ @@ -137,9 +133,12 @@ static void rcu_defer_barrier_queue(struct defer_queue *queue, static void _rcu_defer_barrier_thread(void) { - unsigned long head; + unsigned long head, num_items; head = defer_queue.head; + num_items = head - defer_queue.tail; + if (unlikely(!num_items)) + return; synchronize_rcu(); rcu_defer_barrier_queue(&defer_queue, head); } @@ -152,20 +151,44 @@ void rcu_defer_barrier_thread(void) internal_urcu_unlock(&urcu_defer_mutex); } +/* + * rcu_defer_barrier - Execute all queued rcu callbacks. + * + * Execute all RCU callbacks queued before rcu_defer_barrier() execution. + * All callbacks queued on the local thread prior to a rcu_defer_barrier() call + * are guaranteed to be executed. + * Callbacks queued by other threads concurrently with rcu_defer_barrier() + * execution are not guaranteed to be executed in the current batch (could + * be left for the next batch). These callbacks queued by other threads are only + * guaranteed to be executed if there is explicit synchronization between + * the thread adding to the queue and the thread issuing the defer_barrier call. + */ + void rcu_defer_barrier(void) { struct deferer_registry *index; + unsigned long num_items = 0; if (!registry) return; internal_urcu_lock(&urcu_defer_mutex); - for (index = registry; index < registry + num_deferers; index++) + for (index = registry; index < registry + num_deferers; index++) { index->last_head = LOAD_SHARED(index->defer_queue->head); + num_items += index->last_head - index->defer_queue->tail; + } + if (likely(!num_items)) { + /* + * We skip the grace period because there are no queued + * callbacks to execute. + */ + goto end; + } synchronize_rcu(); for (index = registry; index < registry + num_deferers; index++) rcu_defer_barrier_queue(index->defer_queue, index->last_head); +end: internal_urcu_unlock(&urcu_defer_mutex); }