summary |
shortlog |
log |
commit | commitdiff |
tree
raw |
patch |
inline | side by side (from parent 1:
d486ecd)
Signed-off-by: Mathieu Desnoyers <mathieu.desnoyers@polymtl.ca>
{
assert(reader_data);
sig_done = 0;
{
assert(reader_data);
sig_done = 0;
- smp_mb(); /* write sig_done before sending the signals */
+ /*
+ * pthread_kill has a smp_mb(). But beware, we assume it performs
+ * a cache flush on architectures with non-coherent cache. Let's play
+ * safe and don't assume anything : we use smp_mc() to make sure the
+ * cache flush is enforced.
+ * smp_mb(); write sig_done before sending the signals
+ */
+ smp_mc(); /* write sig_done before sending the signals */
pthread_kill(tid, SIGURCU);
/*
* Wait for sighandler (and thus mb()) to execute on every thread.
pthread_kill(tid, SIGURCU);
/*
* Wait for sighandler (and thus mb()) to execute on every thread.
sig_done = 0;
/*
* pthread_kill has a smp_mb(). But beware, we assume it performs
sig_done = 0;
/*
* pthread_kill has a smp_mb(). But beware, we assume it performs
- * a cache flush on architectures with non-coherent cache.
+ * a cache flush on architectures with non-coherent cache. Let's play
+ * safe and don't assume anything : we use smp_mc() to make sure the
+ * cache flush is enforced.
* smp_mb(); write sig_done before sending the signals
*/
* smp_mb(); write sig_done before sending the signals
*/
+ smp_mc(); /* write sig_done before sending the signals */
for (index = reader_data; index < reader_data + num_readers; index++)
pthread_kill(index->tid, SIGURCU);
/*
for (index = reader_data; index < reader_data + num_readers; index++)
pthread_kill(index->tid, SIGURCU);
/*