X-Git-Url: http://git.liburcu.org/?a=blobdiff_plain;f=lib%2Fringbuffer%2Ffrontend_api.h;h=a6f2c6f4f16742a4b35dafaeafcc25eaa13e1a61;hb=620971f9806e90c0805f766e0e56e95bffa20613;hp=13975602380bf02e82979c071969be0eaa56612a;hpb=5fb66f07aae4884426f1706d0281bd242a38c2a7;p=lttng-modules.git diff --git a/lib/ringbuffer/frontend_api.h b/lib/ringbuffer/frontend_api.h index 13975602..a6f2c6f4 100644 --- a/lib/ringbuffer/frontend_api.h +++ b/lib/ringbuffer/frontend_api.h @@ -29,7 +29,8 @@ * See linux/ringbuffer/frontend.h for channel allocation and read-side API. */ -#include "../../wrapper/ringbuffer/frontend.h" +#include +#include #include #include @@ -73,7 +74,7 @@ static inline void lib_ring_buffer_put_cpu(const struct lib_ring_buffer_config *config) { barrier(); - __get_cpu_var(lib_ring_buffer_nesting)--; + (*lttng_this_cpu_ptr(&lib_ring_buffer_nesting))--; rcu_read_unlock_sched_notrace(); } @@ -125,6 +126,14 @@ int lib_ring_buffer_try_reserve(const struct lib_ring_buffer_config *config, * boundary. It's safe to write. */ *o_end = *o_begin + ctx->slot_size; + + if (unlikely((subbuf_offset(*o_end, chan)) == 0)) + /* + * The offset_end will fall at the very beginning of the next + * subbuffer. + */ + return 1; + return 0; } @@ -153,14 +162,14 @@ int lib_ring_buffer_reserve(const struct lib_ring_buffer_config *config, unsigned long o_begin, o_end, o_old; size_t before_hdr_pad = 0; - if (atomic_read(&chan->record_disabled)) + if (unlikely(atomic_read(&chan->record_disabled))) return -EAGAIN; if (config->alloc == RING_BUFFER_ALLOC_PER_CPU) buf = per_cpu_ptr(chan->backend.buf, ctx->cpu); else buf = chan->backend.buf; - if (atomic_read(&buf->record_disabled)) + if (unlikely(atomic_read(&buf->record_disabled))) return -EAGAIN; ctx->buf = buf; @@ -241,6 +250,7 @@ void lib_ring_buffer_commit(const struct lib_ring_buffer_config *config, unsigned long offset_end = ctx->buf_offset; unsigned long endidx = subbuf_index(offset_end - 1, chan); unsigned long commit_count; + struct commit_counters_hot *cc_hot = &buf->commit_hot[endidx]; /* * Must count record before incrementing the commit count. @@ -261,7 +271,7 @@ void lib_ring_buffer_commit(const struct lib_ring_buffer_config *config, } else smp_wmb(); - v_add(config, ctx->slot_size, &buf->commit_hot[endidx].cc); + v_add(config, ctx->slot_size, &cc_hot->cc); /* * commit count read can race with concurrent OOO commit count updates. @@ -281,17 +291,16 @@ void lib_ring_buffer_commit(const struct lib_ring_buffer_config *config, * count reaches back the reserve offset for a specific sub-buffer, * which is completely independent of the order. */ - commit_count = v_read(config, &buf->commit_hot[endidx].cc); + commit_count = v_read(config, &cc_hot->cc); lib_ring_buffer_check_deliver(config, buf, chan, offset_end - 1, - commit_count, endidx); + commit_count, endidx, ctx->tsc); /* * Update used size at each commit. It's needed only for extracting * ring_buffer buffers from vmcore, after crash. */ - lib_ring_buffer_write_commit_counter(config, buf, chan, endidx, - ctx->buf_offset, commit_count, - ctx->slot_size); + lib_ring_buffer_write_commit_counter(config, buf, chan, + offset_end, commit_count, cc_hot); } /**