X-Git-Url: https://git.liburcu.org/?p=urcu.git;a=blobdiff_plain;f=rculfhash.c;h=8ed9c10b2386c710c2791ed8db3565d4ea13bce5;hp=034c7f9053cfbd881496b592cbb3bcddf92a1d8f;hb=59e371e3c619d7f85698bd997ac556cea49632f3;hpb=adc0de68ccd78689659ed49d4b3d5d36c6720e20 diff --git a/rculfhash.c b/rculfhash.c index 034c7f9..8ed9c10 100644 --- a/rculfhash.c +++ b/rculfhash.c @@ -194,15 +194,13 @@ /* * The removed flag needs to be updated atomically with the pointer. * It indicates that no node must attach to the node scheduled for - * removal. The gc flag also needs to be updated atomically with the - * pointer. It indicates that node garbage collection must be performed. + * removal, and that node garbage collection must be performed. * The dummy flag does not require to be updated atomically with the * pointer, but it is added as a pointer low bit flag to save space. */ #define REMOVED_FLAG (1UL << 0) -#define GC_FLAG (1UL << 1) -#define DUMMY_FLAG (1UL << 2) -#define FLAGS_MASK ((1UL << 3) - 1) +#define DUMMY_FLAG (1UL << 1) +#define FLAGS_MASK ((1UL << 2) - 1) /* Value of the end pointer. Should not interact with flags. */ #define END_VALUE NULL @@ -248,7 +246,7 @@ struct cds_lfht { void (*cds_lfht_rcu_register_thread)(void); void (*cds_lfht_rcu_unregister_thread)(void); pthread_attr_t *resize_attr; /* Resize threads attributes */ - unsigned long count; /* global approximate item count */ + long count; /* global approximate item count */ struct ht_items_count *percpu_count; /* per-cpu item count */ }; @@ -277,11 +275,6 @@ struct cds_lfht_node *_cds_lfht_add(struct cds_lfht *ht, struct cds_lfht_node *node, enum add_mode mode, int dummy); -static -int _cds_lfht_del(struct cds_lfht *ht, unsigned long size, - struct cds_lfht_node *node, - int dummy_removal, int do_gc); - /* * Algorithm to reverse bits in a word by lookup table, extended to * 64-bit words. @@ -563,7 +556,7 @@ void ht_count_add(struct cds_lfht *ht, unsigned long size) return; percpu_count = uatomic_add_return(&ht->percpu_count[cpu].add, 1); if (unlikely(!(percpu_count & ((1UL << COUNT_COMMIT_ORDER) - 1)))) { - unsigned long count; + long count; dbg_printf("add percpu %lu\n", percpu_count); count = uatomic_add_return(&ht->count, @@ -572,7 +565,7 @@ void ht_count_add(struct cds_lfht *ht, unsigned long size) if (!(count & (count - 1))) { if ((count >> CHAIN_LEN_RESIZE_THRESHOLD) < size) return; - dbg_printf("add set global %lu\n", count); + dbg_printf("add set global %ld\n", count); cds_lfht_resize_lazy_count(ht, size, count >> (CHAIN_LEN_TARGET - 1)); } @@ -590,9 +583,9 @@ void ht_count_del(struct cds_lfht *ht, unsigned long size) cpu = ht_get_cpu(); if (unlikely(cpu < 0)) return; - percpu_count = uatomic_add_return(&ht->percpu_count[cpu].del, -1); + percpu_count = uatomic_add_return(&ht->percpu_count[cpu].del, 1); if (unlikely(!(percpu_count & ((1UL << COUNT_COMMIT_ORDER) - 1)))) { - unsigned long count; + long count; dbg_printf("del percpu %lu\n", percpu_count); count = uatomic_add_return(&ht->count, @@ -601,7 +594,13 @@ void ht_count_del(struct cds_lfht *ht, unsigned long size) if (!(count & (count - 1))) { if ((count >> CHAIN_LEN_RESIZE_THRESHOLD) >= size) return; - dbg_printf("del set global %lu\n", count); + dbg_printf("del set global %ld\n", count); + /* + * Don't shrink table if the number of nodes is below a + * certain threshold. + */ + if (count < (1UL << COUNT_COMMIT_ORDER) * (nr_cpus_mask + 1)) + return; cds_lfht_resize_lazy_count(ht, size, count >> (CHAIN_LEN_TARGET - 1)); } @@ -676,18 +675,6 @@ struct cds_lfht_node *flag_removed(struct cds_lfht_node *node) return (struct cds_lfht_node *) (((unsigned long) node) | REMOVED_FLAG); } -static -int is_gc(struct cds_lfht_node *node) -{ - return ((unsigned long) node) & GC_FLAG; -} - -static -struct cds_lfht_node *flag_gc(struct cds_lfht_node *node) -{ - return (struct cds_lfht_node *) (((unsigned long) node) | GC_FLAG); -} - static int is_dummy(struct cds_lfht_node *node) { @@ -743,10 +730,8 @@ void _cds_lfht_gc_bucket(struct cds_lfht_node *dummy, struct cds_lfht_node *node struct cds_lfht_node *iter_prev, *iter, *next, *new_next; assert(!is_dummy(dummy)); - assert(!is_gc(dummy)); assert(!is_removed(dummy)); assert(!is_dummy(node)); - assert(!is_gc(node)); assert(!is_removed(node)); for (;;) { iter_prev = dummy; @@ -766,12 +751,12 @@ void _cds_lfht_gc_bucket(struct cds_lfht_node *dummy, struct cds_lfht_node *node if (likely(clear_flag(iter)->p.reverse_hash > node->p.reverse_hash)) return; next = rcu_dereference(clear_flag(iter)->p.next); - if (likely(is_gc(next))) + if (likely(is_removed(next))) break; iter_prev = clear_flag(iter); iter = next; } - assert(!is_gc(iter)); + assert(!is_removed(iter)); if (is_dummy(iter)) new_next = flag_dummy(clear_flag(next)); else @@ -795,7 +780,6 @@ struct cds_lfht_node *_cds_lfht_add(struct cds_lfht *ht, unsigned long hash, index, order; assert(!is_dummy(node)); - assert(!is_gc(node)); assert(!is_removed(node)); if (!size) { assert(dummy); @@ -823,9 +807,8 @@ struct cds_lfht_node *_cds_lfht_add(struct cds_lfht *ht, if (likely(clear_flag(iter)->p.reverse_hash > node->p.reverse_hash)) goto insert; next = rcu_dereference(clear_flag(iter)->p.next); - if (unlikely(is_gc(next))) + if (unlikely(is_removed(next))) goto gc_node; - assert(!is_removed(next)); if ((mode == ADD_UNIQUE || mode == ADD_REPLACE) && !is_dummy(next) && !ht->compare_fct(node->key, node->key_len, @@ -848,8 +831,6 @@ struct cds_lfht_node *_cds_lfht_add(struct cds_lfht *ht, assert(node != clear_flag(iter)); assert(!is_removed(iter_prev)); assert(!is_removed(iter)); - assert(!is_gc(iter_prev)); - assert(!is_gc(iter)); assert(iter_prev != node); if (!dummy) node->p.next = clear_flag(iter); @@ -877,8 +858,6 @@ struct cds_lfht_node *_cds_lfht_add(struct cds_lfht *ht, assert(node != clear_flag(iter)); assert(!is_removed(iter_prev)); assert(!is_removed(iter)); - assert(!is_gc(iter_prev)); - assert(!is_gc(iter)); assert(iter_prev != node); assert(!dummy); node->p.next = clear_flag(iter); @@ -890,14 +869,13 @@ struct cds_lfht_node *_cds_lfht_add(struct cds_lfht *ht, * Here is the whole trick for lock-free replace: we add * the replacement node _after_ the node we want to * replace by atomically setting its next pointer at the - * same time we set its removal and gc flags. Given that + * same time we set its removal flag. Given that * the lookups/get next use an iterator aware of the * next pointer, they will either skip the old node due - * to the removal/gc flag and see the new node, or use - * the old new, but will not see the new one. + * to the removal flag and see the new node, or use + * the old node, but will not see the new one. */ new_node = flag_removed(new_node); - new_node = flag_gc(new_node); if (uatomic_cmpxchg(&iter_prev->p.next, iter, new_node) != iter) { continue; /* retry */ @@ -908,7 +886,6 @@ struct cds_lfht_node *_cds_lfht_add(struct cds_lfht *ht, gc_node: assert(!is_removed(iter)); - assert(!is_gc(iter)); if (is_dummy(iter)) new_next = flag_dummy(clear_flag(next)); else @@ -929,7 +906,7 @@ gc_end: static int _cds_lfht_del(struct cds_lfht *ht, unsigned long size, struct cds_lfht_node *node, - int dummy_removal, int do_gc) + int dummy_removal) { struct cds_lfht_node *dummy, *next, *old; struct _cds_lfht_node *lookup; @@ -938,7 +915,6 @@ int _cds_lfht_del(struct cds_lfht *ht, unsigned long size, /* logically delete the node */ assert(!is_dummy(node)); - assert(!is_gc(node)); assert(!is_removed(node)); old = rcu_dereference(node->p.next); do { @@ -952,17 +928,12 @@ int _cds_lfht_del(struct cds_lfht *ht, unsigned long size, else assert(!is_dummy(next)); new_next = flag_removed(next); - if (do_gc) - new_next = flag_gc(new_next); old = uatomic_cmpxchg(&node->p.next, next, new_next); } while (old != next); /* We performed the (logical) deletion. */ flagged = 1; - if (!do_gc) - goto end; - /* * Ensure that the node is not visible to readers anymore: lookup for * the node, and remove it (along with any other logically removed node) @@ -1169,7 +1140,7 @@ void remove_table_partition(struct cds_lfht *ht, unsigned long i, fini_node->p.reverse_hash = bit_reverse_ulong(!i ? 0 : (1UL << (i - 1)) + j); (void) _cds_lfht_del(ht, !i ? 0 : (1UL << (i - 1)), - fini_node, 1, 1); + fini_node, 1); if (CMM_LOAD_SHARED(ht->in_progress_destroy)) break; } @@ -1309,11 +1280,11 @@ void cds_lfht_lookup(struct cds_lfht *ht, void *key, size_t key_len, node = clear_flag(node); for (;;) { if (unlikely(is_end(node))) { - node = NULL; + node = next = NULL; break; } if (unlikely(node->p.reverse_hash > reverse_hash)) { - node = NULL; + node = next = NULL; break; } next = rcu_dereference(node->p.next); @@ -1345,11 +1316,11 @@ void cds_lfht_next(struct cds_lfht *ht, struct cds_lfht_iter *iter) for (;;) { if (unlikely(is_end(node))) { - node = NULL; + node = next = NULL; break; } if (unlikely(node->p.reverse_hash > reverse_hash)) { - node = NULL; + node = next = NULL; break; } next = rcu_dereference(node->p.next); @@ -1415,7 +1386,7 @@ int cds_lfht_del(struct cds_lfht *ht, struct cds_lfht_node *node) int ret; size = rcu_dereference(ht->t.size); - ret = _cds_lfht_del(ht, size, node, 0, 1); + ret = _cds_lfht_del(ht, size, node, 0); if (!ret) ht_count_del(ht, size); return ret; @@ -1436,7 +1407,6 @@ int cds_lfht_delete_dummy(struct cds_lfht *ht) if (!is_dummy(node)) return -EPERM; assert(!is_removed(node)); - assert(!is_gc(node)); } while (!is_end(node)); /* * size accessed without rcu_dereference because hash table is @@ -1482,13 +1452,25 @@ int cds_lfht_destroy(struct cds_lfht *ht, pthread_attr_t **attr) } void cds_lfht_count_nodes(struct cds_lfht *ht, + long *approx_before, unsigned long *count, - unsigned long *removed) + unsigned long *removed, + long *approx_after) { struct cds_lfht_node *node, *next; struct _cds_lfht_node *lookup; unsigned long nr_dummy = 0; + *approx_before = 0; + if (nr_cpus_mask >= 0) { + int i; + + for (i = 0; i < nr_cpus_mask + 1; i++) { + *approx_before += uatomic_read(&ht->percpu_count[i].add); + *approx_before -= uatomic_read(&ht->percpu_count[i].del); + } + } + *count = 0; *removed = 0; @@ -1497,9 +1479,11 @@ void cds_lfht_count_nodes(struct cds_lfht *ht, node = (struct cds_lfht_node *) lookup; do { next = rcu_dereference(node->p.next); - if (is_removed(next) || is_gc(next)) { - assert(!is_dummy(next)); - (*removed)++; + if (is_removed(next)) { + if (!is_dummy(next)) + (*removed)++; + else + (nr_dummy)++; } else if (!is_dummy(next)) (*count)++; else @@ -1507,6 +1491,15 @@ void cds_lfht_count_nodes(struct cds_lfht *ht, node = clear_flag(next); } while (!is_end(node)); dbg_printf("number of dummy nodes: %lu\n", nr_dummy); + *approx_after = 0; + if (nr_cpus_mask >= 0) { + int i; + + for (i = 0; i < nr_cpus_mask + 1; i++) { + *approx_after += uatomic_read(&ht->percpu_count[i].add); + *approx_after -= uatomic_read(&ht->percpu_count[i].del); + } + } } /* called with resize mutex held */