Fix: don't perform extra flush on metadata channel
[lttng-modules.git] / lib / ringbuffer / frontend_internal.h
index 281aafd7e4f2bb5227771e17cb7c779af723a460..2bc540370631ab53364a7f8fe318e776b386aabc 100644 (file)
@@ -1,25 +1,37 @@
-#ifndef _LINUX_RING_BUFFER_FRONTEND_INTERNAL_H
-#define _LINUX_RING_BUFFER_FRONTEND_INTERNAL_H
+#ifndef _LIB_RING_BUFFER_FRONTEND_INTERNAL_H
+#define _LIB_RING_BUFFER_FRONTEND_INTERNAL_H
 
 /*
  * linux/ringbuffer/frontend_internal.h
  *
- * (C) Copyright 2005-2010 - Mathieu Desnoyers <mathieu.desnoyers@efficios.com>
- *
  * Ring Buffer Library Synchronization Header (internal helpers).
  *
+ * Copyright (C) 2005-2012 Mathieu Desnoyers <mathieu.desnoyers@efficios.com>
+ *
+ * This library is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Lesser General Public
+ * License as published by the Free Software Foundation; only
+ * version 2.1 of the License.
+ *
+ * This library is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * Lesser General Public License for more details.
+ *
+ * You should have received a copy of the GNU Lesser General Public
+ * License along with this library; if not, write to the Free Software
+ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
+ *
  * Author:
  *     Mathieu Desnoyers <mathieu.desnoyers@efficios.com>
  *
  * See ring_buffer_frontend.c for more information on wait-free algorithms.
- *
- * Dual LGPL v2.1/GPL v2 license.
  */
 
 #include "../../wrapper/ringbuffer/config.h"
 #include "../../wrapper/ringbuffer/backend_types.h"
 #include "../../wrapper/ringbuffer/frontend_types.h"
-#include "../../wrapper/prio_heap.h"   /* For per-CPU read-side iterator */
+#include "../../lib/prio_heap/lttng_prio_heap.h"       /* For per-CPU read-side iterator */
 
 /* Buffer offset macros */
 
@@ -144,6 +156,12 @@ extern
 void lib_ring_buffer_switch_slow(struct lib_ring_buffer *buf,
                                 enum switch_mode mode);
 
+extern
+void lib_ring_buffer_switch_remote(struct lib_ring_buffer *buf);
+
+extern
+void lib_ring_buffer_switch_remote_empty(struct lib_ring_buffer *buf);
+
 /* Buffer write helpers */
 
 static inline
@@ -275,17 +293,24 @@ int lib_ring_buffer_reserve_committed(const struct lib_ring_buffer_config *confi
                     - (commit_count & chan->commit_count_mask) == 0);
 }
 
+/*
+ * Receive end of subbuffer TSC as parameter. It has been read in the
+ * space reservation loop of either reserve or switch, which ensures it
+ * progresses monotonically with event records in the buffer. Therefore,
+ * it ensures that the end timestamp of a subbuffer is <= begin
+ * timestamp of the following subbuffers.
+ */
 static inline
 void lib_ring_buffer_check_deliver(const struct lib_ring_buffer_config *config,
                                   struct lib_ring_buffer *buf,
                                   struct channel *chan,
                                   unsigned long offset,
                                   unsigned long commit_count,
-                                  unsigned long idx)
+                                  unsigned long idx,
+                                  u64 tsc)
 {
        unsigned long old_commit_count = commit_count
                                         - chan->backend.subbuf_size;
-       u64 tsc;
 
        /* Check if all commits have been done */
        if (unlikely((buf_trunc(offset, chan) >> chan->backend.num_subbuf_order)
@@ -316,6 +341,12 @@ void lib_ring_buffer_check_deliver(const struct lib_ring_buffer_config *config,
                 * The subbuffer size is least 2 bytes (minimum size: 1 page).
                 * This guarantees that old_commit_count + 1 != commit_count.
                 */
+
+               /*
+                * Order prior updates to reserve count prior to the
+                * commit_cold cc_sb update.
+                */
+               smp_wmb();
                if (likely(v_cmpxchg(config, &buf->commit_cold[idx].cc_sb,
                                         old_commit_count, old_commit_count + 1)
                           == old_commit_count)) {
@@ -325,7 +356,6 @@ void lib_ring_buffer_check_deliver(const struct lib_ring_buffer_config *config,
                         * and any other writer trying to access this subbuffer
                         * in this state is required to drop records.
                         */
-                       tsc = config->cb.ring_buffer_clock_read(chan);
                        v_add(config,
                              subbuffer_get_records_count(config,
                                                          &buf->backend, idx),
@@ -358,6 +388,11 @@ void lib_ring_buffer_check_deliver(const struct lib_ring_buffer_config *config,
                        /* End of exclusive subbuffer access */
                        v_set(config, &buf->commit_cold[idx].cc_sb,
                              commit_count);
+                       /*
+                        * Order later updates to reserve count after
+                        * the commit_cold cc_sb update.
+                        */
+                       smp_wmb();
                        lib_ring_buffer_vmcore_check_deliver(config, buf,
                                                         commit_count, idx);
 
@@ -389,23 +424,20 @@ void lib_ring_buffer_write_commit_counter(const struct lib_ring_buffer_config *c
                                          struct channel *chan,
                                          unsigned long idx,
                                          unsigned long buf_offset,
-                                         unsigned long commit_count,
-                                         size_t slot_size)
+                                         unsigned long commit_count)
 {
-       unsigned long offset, commit_seq_old;
+       unsigned long commit_seq_old;
 
        if (config->oops != RING_BUFFER_OOPS_CONSISTENCY)
                return;
 
-       offset = buf_offset + slot_size;
-
        /*
         * subbuf_offset includes commit_count_mask. We can simply
         * compare the offsets within the subbuffer without caring about
         * buffer full/empty mismatch because offset is never zero here
         * (subbuffer header and record headers have non-zero length).
         */
-       if (unlikely(subbuf_offset(offset - commit_count, chan)))
+       if (unlikely(subbuf_offset(buf_offset - commit_count, chan)))
                return;
 
        commit_seq_old = v_read(config, &buf->commit_hot[idx].seq);
@@ -421,4 +453,4 @@ extern void lib_ring_buffer_free(struct lib_ring_buffer *buf);
 /* Keep track of trap nesting inside ring buffer code */
 DECLARE_PER_CPU(unsigned int, lib_ring_buffer_nesting);
 
-#endif /* _LINUX_RING_BUFFER_FRONTEND_INTERNAL_H */
+#endif /* _LIB_RING_BUFFER_FRONTEND_INTERNAL_H */
This page took 0.030186 seconds and 4 git commands to generate.