X-Git-Url: https://git.liburcu.org/?a=blobdiff_plain;f=urcu.c;h=c6d7dbc9d83a011a7b7d2885f93460a268da6ea2;hb=d40fde2c4f6170c3ace770354d99acc9e9732ec5;hp=140cd19b98add54f276755e08b13bfad9e8c3d0c;hpb=27d65bc5db18f8563a9120924569acffb845ec4d;p=urcu.git diff --git a/urcu.c b/urcu.c index 140cd19..c6d7dbc 100644 --- a/urcu.c +++ b/urcu.c @@ -99,9 +99,9 @@ static void mutex_lock(pthread_mutex_t *mutex) perror("Error in pthread mutex lock"); exit(-1); } - if (rcu_reader.need_mb) { + if (LOAD_SHARED(rcu_reader.need_mb)) { smp_mb(); - rcu_reader.need_mb = 0; + _STORE_SHARED(rcu_reader.need_mb, 0); smp_mb(); } poll(NULL,0,10); @@ -155,8 +155,7 @@ static void force_mb_all_readers(void) * cache flush is enforced. */ list_for_each_entry(index, ®istry, head) { - index->need_mb = 1; - smp_mc(); /* write need_mb before sending the signal */ + STORE_SHARED(index->need_mb, 1); pthread_kill(index->tid, SIGRCU); } /* @@ -173,7 +172,7 @@ static void force_mb_all_readers(void) * the Linux Test Project (LTP). */ list_for_each_entry(index, ®istry, head) { - while (index->need_mb) { + while (LOAD_SHARED(index->need_mb)) { pthread_kill(index->tid, SIGRCU); poll(NULL, 0, 1); } @@ -209,13 +208,15 @@ void update_counter_and_wait(void) STORE_SHARED(rcu_gp_ctr, rcu_gp_ctr ^ RCU_GP_CTR_PHASE); /* - * Must commit qparity update to memory before waiting for other parity - * quiescent state. Failure to do so could result in the writer waiting - * forever while new readers are always accessing data (no progress). - * Ensured by STORE_SHARED and LOAD_SHARED. + * Must commit rcu_gp_ctr update to memory before waiting for quiescent + * state. Failure to do so could result in the writer waiting forever + * while new readers are always accessing data (no progress). Enforce + * compiler-order of store to rcu_gp_ctr before load rcu_reader ctr. */ + barrier(); /* + * * Adding a smp_mb() which is _not_ formally required, but makes the * model easier to understand. It does not have a big performance impact * anyway, given this is the write-side. @@ -234,7 +235,7 @@ void update_counter_and_wait(void) } list_for_each_entry_safe(index, tmp, ®istry, head) { - if (!rcu_old_gp_ongoing(&index->ctr)) + if (!rcu_gp_ongoing(&index->ctr)) list_move(&index->head, &qsreaders); } @@ -303,11 +304,12 @@ void synchronize_rcu(void) /* * Must finish waiting for quiescent state for parity 0 before - * committing qparity update to memory. Failure to do so could result in - * the writer waiting forever while new readers are always accessing - * data (no progress). - * Ensured by STORE_SHARED and LOAD_SHARED. + * committing next rcu_gp_ctr update to memory. Failure to do so could + * result in the writer waiting forever while new readers are always + * accessing data (no progress). Enforce compiler-order of load + * rcu_reader ctr before store to rcu_gp_ctr. */ + barrier(); /* * Adding a smp_mb() which is _not_ formally required, but makes the @@ -347,7 +349,7 @@ void rcu_register_thread(void) { rcu_reader.tid = pthread_self(); assert(rcu_reader.need_mb == 0); - assert(rcu_reader.ctr == 0); + assert(!(rcu_reader.ctr & RCU_GP_CTR_NEST_MASK)); mutex_lock(&rcu_gp_lock); rcu_init(); /* In case gcc does not support constructor attribute */ @@ -382,7 +384,7 @@ static void sigrcu_handler(int signo, siginfo_t *siginfo, void *context) * executed on. */ smp_mb(); - rcu_reader.need_mb = 0; + _STORE_SHARED(rcu_reader.need_mb, 0); smp_mb(); }