projects
/
urcu.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
rculfhash: use dbg_printf() for grow/shrink printout
[urcu.git]
/
rculfhash.c
diff --git
a/rculfhash.c
b/rculfhash.c
index 13274189887182f81a2a8ed5336db50a36e13c90..cf822fcd620e98b47d6a718582f1385887f34496 100644
(file)
--- a/
rculfhash.c
+++ b/
rculfhash.c
@@
-255,6
+255,7
@@
struct rcu_resize_work {
};
struct partition_resize_work {
};
struct partition_resize_work {
+ pthread_t thread_id;
struct cds_lfht *ht;
unsigned long i, start, len;
void (*fct)(struct cds_lfht *ht, unsigned long i,
struct cds_lfht *ht;
unsigned long i, start, len;
void (*fct)(struct cds_lfht *ht, unsigned long i,
@@
-1033,7
+1034,6
@@
void partition_resize_helper(struct cds_lfht *ht, unsigned long i,
struct partition_resize_work *work;
int thread, ret;
unsigned long nr_threads;
struct partition_resize_work *work;
int thread, ret;
unsigned long nr_threads;
- pthread_t *thread_id;
/*
* Note: nr_cpus_mask + 1 is always power of 2.
/*
* Note: nr_cpus_mask + 1 is always power of 2.
@@
-1048,7
+1048,6
@@
void partition_resize_helper(struct cds_lfht *ht, unsigned long i,
}
partition_len = len >> get_count_order_ulong(nr_threads);
work = calloc(nr_threads, sizeof(*work));
}
partition_len = len >> get_count_order_ulong(nr_threads);
work = calloc(nr_threads, sizeof(*work));
- thread_id = calloc(nr_threads, sizeof(*thread_id));
assert(work);
for (thread = 0; thread < nr_threads; thread++) {
work[thread].ht = ht;
assert(work);
for (thread = 0; thread < nr_threads; thread++) {
work[thread].ht = ht;
@@
-1056,16
+1055,15
@@
void partition_resize_helper(struct cds_lfht *ht, unsigned long i,
work[thread].len = partition_len;
work[thread].start = thread * partition_len;
work[thread].fct = fct;
work[thread].len = partition_len;
work[thread].start = thread * partition_len;
work[thread].fct = fct;
- ret = pthread_create(&
thread_id[thread]
, ht->resize_attr,
+ ret = pthread_create(&
(work[thread].thread_id)
, ht->resize_attr,
partition_resize_thread, &work[thread]);
assert(!ret);
}
for (thread = 0; thread < nr_threads; thread++) {
partition_resize_thread, &work[thread]);
assert(!ret);
}
for (thread = 0; thread < nr_threads; thread++) {
- ret = pthread_join(
thread_id[thread]
, NULL);
+ ret = pthread_join(
work[thread].thread_id
, NULL);
assert(!ret);
}
free(work);
assert(!ret);
}
free(work);
- free(thread_id);
}
/*
}
/*
@@
-1611,8
+1609,8
@@
void _do_cds_lfht_grow(struct cds_lfht *ht,
old_order = get_count_order_ulong(old_size) + 1;
new_order = get_count_order_ulong(new_size) + 1;
old_order = get_count_order_ulong(old_size) + 1;
new_order = get_count_order_ulong(new_size) + 1;
- printf("resize from %lu (order %lu) to %lu (order %lu) buckets\n",
- old_size, old_order, new_size, new_order);
+
dbg_
printf("resize from %lu (order %lu) to %lu (order %lu) buckets\n",
+
old_size, old_order, new_size, new_order);
assert(new_size > old_size);
init_table(ht, old_order, new_order - old_order);
}
assert(new_size > old_size);
init_table(ht, old_order, new_order - old_order);
}
@@
-1627,8
+1625,8
@@
void _do_cds_lfht_shrink(struct cds_lfht *ht,
new_size = max(new_size, MIN_TABLE_SIZE);
old_order = get_count_order_ulong(old_size) + 1;
new_order = get_count_order_ulong(new_size) + 1;
new_size = max(new_size, MIN_TABLE_SIZE);
old_order = get_count_order_ulong(old_size) + 1;
new_order = get_count_order_ulong(new_size) + 1;
- printf("resize from %lu (order %lu) to %lu (order %lu) buckets\n",
- old_size, old_order, new_size, new_order);
+
dbg_
printf("resize from %lu (order %lu) to %lu (order %lu) buckets\n",
+
old_size, old_order, new_size, new_order);
assert(new_size < old_size);
/* Remove and unlink all dummy nodes to remove. */
assert(new_size < old_size);
/* Remove and unlink all dummy nodes to remove. */
This page took
0.024828 seconds
and
4
git commands to generate.