Namespace lttng/align.h as lttng/ust-align.h
[lttng-ust.git] / libringbuffer / ring_buffer_frontend.c
index 9721df1634585f028316d13f96b184af7e99cc08..23122865dfe4b4360da35a46438658881d9946ec 100644 (file)
@@ -1,23 +1,8 @@
 /*
- * ring_buffer_frontend.c
+ * SPDX-License-Identifier: LGPL-2.1-only
  *
  * Copyright (C) 2005-2012 Mathieu Desnoyers <mathieu.desnoyers@efficios.com>
  *
- * This library is free software; you can redistribute it and/or
- * modify it under the terms of the GNU Lesser General Public
- * License as published by the Free Software Foundation; only
- * version 2.1 of the License.
- *
- * This library is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- * Lesser General Public License for more details.
- *
- * You should have received a copy of the GNU Lesser General Public
- * License along with this library; if not, write to the Free Software
- * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
- *
- *
  * Ring buffer wait-free buffer synchronization. Producer-consumer and flight
  * recorder (overwrite) modes. See thesis:
  *
@@ -51,7 +36,6 @@
  *   - put_subbuf
  */
 
-#define _GNU_SOURCE
 #define _LGPL_SOURCE
 #include <sys/types.h>
 #include <sys/mman.h>
 #include <signal.h>
 #include <time.h>
 #include <stdbool.h>
+#include <stdint.h>
 #include <urcu/compiler.h>
 #include <urcu/ref.h>
 #include <urcu/tls-compat.h>
 #include <poll.h>
-#include <helper.h>
+#include <ust-helper.h>
+
+#include <lttng/ust-align.h>
+#include <lttng/ringbuffer-context.h>
 
 #include "smp.h"
-#include <lttng/ringbuffer-config.h>
+#include "ringbuffer-config.h"
 #include "vatomic.h"
 #include "backend.h"
 #include "frontend.h"
@@ -127,7 +115,7 @@ DEFINE_URCU_TLS(unsigned int, lib_ring_buffer_nesting);
 static pthread_mutex_t wakeup_fd_mutex = PTHREAD_MUTEX_INITIALIZER;
 
 static
-void lib_ring_buffer_print_errors(struct channel *chan,
+void lib_ring_buffer_print_errors(struct lttng_ust_lib_ring_buffer_channel *chan,
                                struct lttng_ust_lib_ring_buffer *buf, int cpu,
                                struct lttng_ust_shm_handle *handle);
 
@@ -159,7 +147,7 @@ void lttng_ust_ringbuffer_set_allow_blocking(void)
 }
 
 /* Get blocking timeout, in ms */
-static int lttng_ust_ringbuffer_get_timeout(struct channel *chan)
+static int lttng_ust_ringbuffer_get_timeout(struct lttng_ust_lib_ring_buffer_channel *chan)
 {
        if (!lttng_ust_allow_blocking)
                return 0;
@@ -178,7 +166,7 @@ static int lttng_ust_ringbuffer_get_timeout(struct channel *chan)
 void lib_ring_buffer_reset(struct lttng_ust_lib_ring_buffer *buf,
                           struct lttng_ust_shm_handle *handle)
 {
-       struct channel *chan;
+       struct lttng_ust_lib_ring_buffer_channel *chan;
        const struct lttng_ust_lib_ring_buffer_config *config;
        unsigned int i;
 
@@ -194,6 +182,7 @@ void lib_ring_buffer_reset(struct lttng_ust_lib_ring_buffer *buf,
        for (i = 0; i < chan->backend.num_subbuf; i++) {
                struct commit_counters_hot *cc_hot;
                struct commit_counters_cold *cc_cold;
+               uint64_t *ts_end;
 
                cc_hot = shmp_index(handle, buf->commit_hot, i);
                if (!cc_hot)
@@ -201,9 +190,13 @@ void lib_ring_buffer_reset(struct lttng_ust_lib_ring_buffer *buf,
                cc_cold = shmp_index(handle, buf->commit_cold, i);
                if (!cc_cold)
                        return;
+               ts_end = shmp_index(handle, buf->ts_end, i);
+               if (!ts_end)
+                       return;
                v_set(config, &cc_hot->cc, 0);
                v_set(config, &cc_hot->seq, 0);
                v_set(config, &cc_cold->cc_sb, 0);
+               *ts_end = 0;
        }
        uatomic_set(&buf->consumed, 0);
        uatomic_set(&buf->record_disabled, 0);
@@ -227,7 +220,7 @@ void lib_ring_buffer_reset(struct lttng_ust_lib_ring_buffer *buf,
  * be using the iterator concurrently with reset. The previous current iterator
  * record is reset.
  */
-void channel_reset(struct channel *chan)
+void channel_reset(struct lttng_ust_lib_ring_buffer_channel *chan)
 {
        /*
         * Reset iterators first. Will put the subbuffer if held for reading.
@@ -338,9 +331,10 @@ int lib_ring_buffer_create(struct lttng_ust_lib_ring_buffer *buf,
                           struct shm_object *shmobj)
 {
        const struct lttng_ust_lib_ring_buffer_config *config = &chanb->config;
-       struct channel *chan = caa_container_of(chanb, struct channel, backend);
+       struct lttng_ust_lib_ring_buffer_channel *chan = caa_container_of(chanb,
+                       struct lttng_ust_lib_ring_buffer_channel, backend);
        struct lttng_ust_lib_ring_buffer_backend_subbuffer *wsb;
-       struct channel *shmp_chan;
+       struct lttng_ust_lib_ring_buffer_channel *shmp_chan;
        struct commit_counters_hot *cc_hot;
        void *priv = channel_get_private(chan);
        size_t subbuf_header_size;
@@ -368,6 +362,16 @@ int lib_ring_buffer_create(struct lttng_ust_lib_ring_buffer *buf,
                goto free_commit;
        }
 
+       align_shm(shmobj, __alignof__(uint64_t));
+       set_shmp(buf->ts_end,
+                zalloc_shm(shmobj,
+                       sizeof(uint64_t) * chan->backend.num_subbuf));
+       if (!shmp(handle, buf->ts_end)) {
+               ret = -ENOMEM;
+               goto free_commit_cold;
+       }
+
+
        ret = lib_ring_buffer_backend_create(&buf->backend, &chan->backend,
                        cpu, handle, shmobj);
        if (ret) {
@@ -414,6 +418,8 @@ int lib_ring_buffer_create(struct lttng_ust_lib_ring_buffer *buf,
 
        /* Error handling */
 free_init:
+       /* ts_end will be freed by shm teardown */
+free_commit_cold:
        /* commit_cold will be freed by shm teardown */
 free_commit:
        /* commit_hot will be freed by shm teardown */
@@ -426,7 +432,7 @@ void lib_ring_buffer_channel_switch_timer(int sig, siginfo_t *si, void *uc)
 {
        const struct lttng_ust_lib_ring_buffer_config *config;
        struct lttng_ust_shm_handle *handle;
-       struct channel *chan;
+       struct lttng_ust_lib_ring_buffer_channel *chan;
        int cpu;
 
        assert(CMM_LOAD_SHARED(timer_signal.tid) == pthread_self());
@@ -470,7 +476,7 @@ end:
 static
 int lib_ring_buffer_poll_deliver(const struct lttng_ust_lib_ring_buffer_config *config,
                                 struct lttng_ust_lib_ring_buffer *buf,
-                                struct channel *chan,
+                                struct lttng_ust_lib_ring_buffer_channel *chan,
                                 struct lttng_ust_shm_handle *handle)
 {
        unsigned long consumed_old, consumed_idx, commit_count, write_offset;
@@ -579,7 +585,7 @@ void lib_ring_buffer_wakeup(struct lttng_ust_lib_ring_buffer *buf,
 }
 
 static
-void lib_ring_buffer_channel_do_read(struct channel *chan)
+void lib_ring_buffer_channel_do_read(struct lttng_ust_lib_ring_buffer_channel *chan)
 {
        const struct lttng_ust_lib_ring_buffer_config *config;
        struct lttng_ust_shm_handle *handle;
@@ -624,7 +630,7 @@ end:
 static
 void lib_ring_buffer_channel_read_timer(int sig, siginfo_t *si, void *uc)
 {
-       struct channel *chan;
+       struct lttng_ust_lib_ring_buffer_channel *chan;
 
        assert(CMM_LOAD_SHARED(timer_signal.tid) == pthread_self());
        chan = si->si_value.sival_ptr;
@@ -774,7 +780,7 @@ void lib_ring_buffer_wait_signal_thread_qs(unsigned int signr)
 }
 
 static
-void lib_ring_buffer_channel_switch_timer_start(struct channel *chan)
+void lib_ring_buffer_channel_switch_timer_start(struct lttng_ust_lib_ring_buffer_channel *chan)
 {
        struct sigevent sev;
        struct itimerspec its;
@@ -787,6 +793,7 @@ void lib_ring_buffer_channel_switch_timer_start(struct channel *chan)
 
        lib_ring_buffer_setup_timer_thread();
 
+       memset(&sev, 0, sizeof(sev));
        sev.sigev_notify = SIGEV_SIGNAL;
        sev.sigev_signo = LTTNG_UST_RB_SIG_FLUSH;
        sev.sigev_value.sival_ptr = chan;
@@ -807,7 +814,7 @@ void lib_ring_buffer_channel_switch_timer_start(struct channel *chan)
 }
 
 static
-void lib_ring_buffer_channel_switch_timer_stop(struct channel *chan)
+void lib_ring_buffer_channel_switch_timer_stop(struct lttng_ust_lib_ring_buffer_channel *chan)
 {
        int ret;
 
@@ -826,7 +833,7 @@ void lib_ring_buffer_channel_switch_timer_stop(struct channel *chan)
 }
 
 static
-void lib_ring_buffer_channel_read_timer_start(struct channel *chan)
+void lib_ring_buffer_channel_read_timer_start(struct lttng_ust_lib_ring_buffer_channel *chan)
 {
        const struct lttng_ust_lib_ring_buffer_config *config = &chan->backend.config;
        struct sigevent sev;
@@ -861,7 +868,7 @@ void lib_ring_buffer_channel_read_timer_start(struct channel *chan)
 }
 
 static
-void lib_ring_buffer_channel_read_timer_stop(struct channel *chan)
+void lib_ring_buffer_channel_read_timer_stop(struct lttng_ust_lib_ring_buffer_channel *chan)
 {
        const struct lttng_ust_lib_ring_buffer_config *config = &chan->backend.config;
        int ret;
@@ -887,14 +894,14 @@ void lib_ring_buffer_channel_read_timer_stop(struct channel *chan)
        chan->read_timer_enabled = 0;
 }
 
-static void channel_unregister_notifiers(struct channel *chan,
+static void channel_unregister_notifiers(struct lttng_ust_lib_ring_buffer_channel *chan,
                           struct lttng_ust_shm_handle *handle)
 {
        lib_ring_buffer_channel_switch_timer_stop(chan);
        lib_ring_buffer_channel_read_timer_stop(chan);
 }
 
-static void channel_print_errors(struct channel *chan,
+static void channel_print_errors(struct lttng_ust_lib_ring_buffer_channel *chan,
                struct lttng_ust_shm_handle *handle)
 {
        const struct lttng_ust_lib_ring_buffer_config *config =
@@ -917,7 +924,7 @@ static void channel_print_errors(struct channel *chan,
        }
 }
 
-static void channel_free(struct channel *chan,
+static void channel_free(struct lttng_ust_lib_ring_buffer_channel *chan,
                struct lttng_ust_shm_handle *handle,
                int consumer)
 {
@@ -963,7 +970,7 @@ struct lttng_ust_shm_handle *channel_create(const struct lttng_ust_lib_ring_buff
 {
        int ret;
        size_t shmsize, chansize;
-       struct channel *chan;
+       struct lttng_ust_lib_ring_buffer_channel *chan;
        struct lttng_ust_shm_handle *handle;
        struct shm_object *shmobj;
        unsigned int nr_streams;
@@ -1004,12 +1011,12 @@ struct lttng_ust_shm_handle *channel_create(const struct lttng_ust_lib_ring_buff
                goto error_table_alloc;
 
        /* Calculate the shm allocation layout */
-       shmsize = sizeof(struct channel);
-       shmsize += offset_align(shmsize, __alignof__(struct lttng_ust_lib_ring_buffer_shmp));
+       shmsize = sizeof(struct lttng_ust_lib_ring_buffer_channel);
+       shmsize += lttng_ust_offset_align(shmsize, __alignof__(struct lttng_ust_lib_ring_buffer_shmp));
        shmsize += sizeof(struct lttng_ust_lib_ring_buffer_shmp) * nr_streams;
        chansize = shmsize;
        if (priv_data_align)
-               shmsize += offset_align(shmsize, priv_data_align);
+               shmsize += lttng_ust_offset_align(shmsize, priv_data_align);
        shmsize += priv_data_size;
 
        /* Allocate normal memory for channel (not shared) */
@@ -1017,7 +1024,7 @@ struct lttng_ust_shm_handle *channel_create(const struct lttng_ust_lib_ring_buff
                        -1, -1);
        if (!shmobj)
                goto error_append;
-       /* struct channel is at object 0, offset 0 (hardcoded) */
+       /* struct lttng_ust_lib_ring_buffer_channel is at object 0, offset 0 (hardcoded) */
        set_shmp(handle->chan, zalloc_shm(shmobj, chansize));
        assert(handle->chan._ref.index == 0);
        assert(handle->chan._ref.offset == 0);
@@ -1089,7 +1096,7 @@ struct lttng_ust_shm_handle *channel_handle_create(void *data,
                        memory_map_size, wakeup_fd);
        if (!object)
                goto error_table_object;
-       /* struct channel is at object 0, offset 0 (hardcoded) */
+       /* struct lttng_ust_lib_ring_buffer_channel is at object 0, offset 0 (hardcoded) */
        handle->chan._ref.index = 0;
        handle->chan._ref.offset = 0;
        return handle;
@@ -1123,7 +1130,7 @@ unsigned int channel_handle_get_nr_streams(struct lttng_ust_shm_handle *handle)
 }
 
 static
-void channel_release(struct channel *chan, struct lttng_ust_shm_handle *handle,
+void channel_release(struct lttng_ust_lib_ring_buffer_channel *chan, struct lttng_ust_shm_handle *handle,
                int consumer)
 {
        channel_free(chan, handle, consumer);
@@ -1137,9 +1144,9 @@ void channel_release(struct channel *chan, struct lttng_ust_shm_handle *handle,
  * Call "destroy" callback, finalize channels, decrement the channel
  * reference count. Note that when readers have completed data
  * consumption of finalized channels, get_subbuf() will return -ENODATA.
- * They should release their handle at that point. 
+ * They should release their handle at that point.
  */
-void channel_destroy(struct channel *chan, struct lttng_ust_shm_handle *handle,
+void channel_destroy(struct lttng_ust_lib_ring_buffer_channel *chan, struct lttng_ust_shm_handle *handle,
                int consumer)
 {
        if (consumer) {
@@ -1164,7 +1171,7 @@ void channel_destroy(struct channel *chan, struct lttng_ust_shm_handle *handle,
 
 struct lttng_ust_lib_ring_buffer *channel_get_ring_buffer(
                                        const struct lttng_ust_lib_ring_buffer_config *config,
-                                       struct channel *chan, int cpu,
+                                       struct lttng_ust_lib_ring_buffer_channel *chan, int cpu,
                                        struct lttng_ust_shm_handle *handle,
                                        int *shm_fd, int *wait_fd,
                                        int *wakeup_fd,
@@ -1188,7 +1195,7 @@ struct lttng_ust_lib_ring_buffer *channel_get_ring_buffer(
 }
 
 int ring_buffer_channel_close_wait_fd(const struct lttng_ust_lib_ring_buffer_config *config,
-                       struct channel *chan,
+                       struct lttng_ust_lib_ring_buffer_channel *chan,
                        struct lttng_ust_shm_handle *handle)
 {
        struct shm_ref *ref;
@@ -1198,7 +1205,7 @@ int ring_buffer_channel_close_wait_fd(const struct lttng_ust_lib_ring_buffer_con
 }
 
 int ring_buffer_channel_close_wakeup_fd(const struct lttng_ust_lib_ring_buffer_config *config,
-                       struct channel *chan,
+                       struct lttng_ust_lib_ring_buffer_channel *chan,
                        struct lttng_ust_shm_handle *handle)
 {
        struct shm_ref *ref;
@@ -1208,7 +1215,7 @@ int ring_buffer_channel_close_wakeup_fd(const struct lttng_ust_lib_ring_buffer_c
 }
 
 int ring_buffer_stream_close_wait_fd(const struct lttng_ust_lib_ring_buffer_config *config,
-                       struct channel *chan,
+                       struct lttng_ust_lib_ring_buffer_channel *chan,
                        struct lttng_ust_shm_handle *handle,
                        int cpu)
 {
@@ -1225,7 +1232,7 @@ int ring_buffer_stream_close_wait_fd(const struct lttng_ust_lib_ring_buffer_conf
 }
 
 int ring_buffer_stream_close_wakeup_fd(const struct lttng_ust_lib_ring_buffer_config *config,
-                       struct channel *chan,
+                       struct lttng_ust_lib_ring_buffer_channel *chan,
                        struct lttng_ust_shm_handle *handle,
                        int cpu)
 {
@@ -1257,7 +1264,7 @@ int lib_ring_buffer_open_read(struct lttng_ust_lib_ring_buffer *buf,
 void lib_ring_buffer_release_read(struct lttng_ust_lib_ring_buffer *buf,
                                  struct lttng_ust_shm_handle *handle)
 {
-       struct channel *chan = shmp(handle, buf->backend.chan);
+       struct lttng_ust_lib_ring_buffer_channel *chan = shmp(handle, buf->backend.chan);
 
        if (!chan)
                return;
@@ -1280,7 +1287,7 @@ int lib_ring_buffer_snapshot(struct lttng_ust_lib_ring_buffer *buf,
                             unsigned long *consumed, unsigned long *produced,
                             struct lttng_ust_shm_handle *handle)
 {
-       struct channel *chan;
+       struct lttng_ust_lib_ring_buffer_channel *chan;
        const struct lttng_ust_lib_ring_buffer_config *config;
        unsigned long consumed_cur, write_offset;
        int finalized;
@@ -1345,7 +1352,7 @@ int lib_ring_buffer_snapshot_sample_positions(
                             unsigned long *consumed, unsigned long *produced,
                             struct lttng_ust_shm_handle *handle)
 {
-       struct channel *chan;
+       struct lttng_ust_lib_ring_buffer_channel *chan;
        const struct lttng_ust_lib_ring_buffer_config *config;
 
        chan = shmp(handle, buf->backend.chan);
@@ -1376,7 +1383,7 @@ void lib_ring_buffer_move_consumer(struct lttng_ust_lib_ring_buffer *buf,
                                   struct lttng_ust_shm_handle *handle)
 {
        struct lttng_ust_lib_ring_buffer_backend *bufb = &buf->backend;
-       struct channel *chan;
+       struct lttng_ust_lib_ring_buffer_channel *chan;
        unsigned long consumed;
 
        chan = shmp(handle, bufb->chan);
@@ -1407,7 +1414,7 @@ int lib_ring_buffer_get_subbuf(struct lttng_ust_lib_ring_buffer *buf,
                               unsigned long consumed,
                               struct lttng_ust_shm_handle *handle)
 {
-       struct channel *chan;
+       struct lttng_ust_lib_ring_buffer_channel *chan;
        const struct lttng_ust_lib_ring_buffer_config *config;
        unsigned long consumed_cur, consumed_idx, commit_count, write_offset;
        int ret, finalized, nr_retry = LTTNG_UST_RING_BUFFER_GET_RETRY;
@@ -1571,7 +1578,7 @@ void lib_ring_buffer_put_subbuf(struct lttng_ust_lib_ring_buffer *buf,
                                struct lttng_ust_shm_handle *handle)
 {
        struct lttng_ust_lib_ring_buffer_backend *bufb = &buf->backend;
-       struct channel *chan;
+       struct lttng_ust_lib_ring_buffer_channel *chan;
        const struct lttng_ust_lib_ring_buffer_config *config;
        unsigned long sb_bindex, consumed_idx, consumed;
        struct lttng_ust_lib_ring_buffer_backend_pages_shmp *rpages;
@@ -1638,7 +1645,7 @@ void lib_ring_buffer_put_subbuf(struct lttng_ust_lib_ring_buffer *buf,
  */
 static
 void lib_ring_buffer_print_subbuffer_errors(struct lttng_ust_lib_ring_buffer *buf,
-                                           struct channel *chan,
+                                           struct lttng_ust_lib_ring_buffer_channel *chan,
                                            unsigned long cons_offset,
                                            int cpu,
                                            struct lttng_ust_shm_handle *handle)
@@ -1673,7 +1680,7 @@ void lib_ring_buffer_print_subbuffer_errors(struct lttng_ust_lib_ring_buffer *bu
 
 static
 void lib_ring_buffer_print_buffer_errors(struct lttng_ust_lib_ring_buffer *buf,
-                                        struct channel *chan,
+                                        struct lttng_ust_lib_ring_buffer_channel *chan,
                                         void *priv, int cpu,
                                         struct lttng_ust_shm_handle *handle)
 {
@@ -1703,7 +1710,7 @@ void lib_ring_buffer_print_buffer_errors(struct lttng_ust_lib_ring_buffer *buf,
 }
 
 static
-void lib_ring_buffer_print_errors(struct channel *chan,
+void lib_ring_buffer_print_errors(struct lttng_ust_lib_ring_buffer_channel *chan,
                                struct lttng_ust_lib_ring_buffer *buf, int cpu,
                                struct lttng_ust_shm_handle *handle)
 {
@@ -1745,7 +1752,7 @@ void lib_ring_buffer_print_errors(struct channel *chan,
  */
 static
 void lib_ring_buffer_switch_old_start(struct lttng_ust_lib_ring_buffer *buf,
-                                     struct channel *chan,
+                                     struct lttng_ust_lib_ring_buffer_channel *chan,
                                      struct switch_offsets *offsets,
                                      uint64_t tsc,
                                      struct lttng_ust_shm_handle *handle)
@@ -1786,7 +1793,7 @@ void lib_ring_buffer_switch_old_start(struct lttng_ust_lib_ring_buffer *buf,
  */
 static
 void lib_ring_buffer_switch_old_end(struct lttng_ust_lib_ring_buffer *buf,
-                                   struct channel *chan,
+                                   struct lttng_ust_lib_ring_buffer_channel *chan,
                                    struct switch_offsets *offsets,
                                    uint64_t tsc,
                                    struct lttng_ust_shm_handle *handle)
@@ -1795,15 +1802,29 @@ void lib_ring_buffer_switch_old_end(struct lttng_ust_lib_ring_buffer *buf,
        unsigned long oldidx = subbuf_index(offsets->old - 1, chan);
        unsigned long commit_count, padding_size, data_size;
        struct commit_counters_hot *cc_hot;
+       uint64_t *ts_end;
 
        data_size = subbuf_offset(offsets->old - 1, chan) + 1;
        padding_size = chan->backend.subbuf_size - data_size;
        subbuffer_set_data_size(config, &buf->backend, oldidx, data_size,
                                handle);
 
+       ts_end = shmp_index(handle, buf->ts_end, oldidx);
+       if (!ts_end)
+               return;
        /*
-        * Order all writes to buffer before the commit count update that will
-        * determine that the subbuffer is full.
+        * This is the last space reservation in that sub-buffer before
+        * it gets delivered. This provides exclusive access to write to
+        * this sub-buffer's ts_end. There are also no concurrent
+        * readers of that ts_end because delivery of that sub-buffer is
+        * postponed until the commit counter is incremented for the
+        * current space reservation.
+        */
+       *ts_end = tsc;
+
+       /*
+        * Order all writes to buffer and store to ts_end before the commit
+        * count update that will determine that the subbuffer is full.
         */
        cmm_smp_wmb();
        cc_hot = shmp_index(handle, buf->commit_hot, oldidx);
@@ -1827,7 +1848,7 @@ void lib_ring_buffer_switch_old_end(struct lttng_ust_lib_ring_buffer *buf,
  */
 static
 void lib_ring_buffer_switch_new_start(struct lttng_ust_lib_ring_buffer *buf,
-                                     struct channel *chan,
+                                     struct lttng_ust_lib_ring_buffer_channel *chan,
                                      struct switch_offsets *offsets,
                                      uint64_t tsc,
                                      struct lttng_ust_shm_handle *handle)
@@ -1867,18 +1888,31 @@ void lib_ring_buffer_switch_new_start(struct lttng_ust_lib_ring_buffer *buf,
  */
 static
 void lib_ring_buffer_switch_new_end(struct lttng_ust_lib_ring_buffer *buf,
-                                   struct channel *chan,
+                                   struct lttng_ust_lib_ring_buffer_channel *chan,
                                    struct switch_offsets *offsets,
                                    uint64_t tsc,
                                    struct lttng_ust_shm_handle *handle)
 {
        const struct lttng_ust_lib_ring_buffer_config *config = &chan->backend.config;
        unsigned long endidx, data_size;
+       uint64_t *ts_end;
 
        endidx = subbuf_index(offsets->end - 1, chan);
        data_size = subbuf_offset(offsets->end - 1, chan) + 1;
        subbuffer_set_data_size(config, &buf->backend, endidx, data_size,
                                handle);
+       ts_end = shmp_index(handle, buf->ts_end, endidx);
+       if (!ts_end)
+               return;
+       /*
+        * This is the last space reservation in that sub-buffer before
+        * it gets delivered. This provides exclusive access to write to
+        * this sub-buffer's ts_end. There are also no concurrent
+        * readers of that ts_end because delivery of that sub-buffer is
+        * postponed until the commit counter is incremented for the
+        * current space reservation.
+        */
+       *ts_end = tsc;
 }
 
 /*
@@ -1889,7 +1923,7 @@ void lib_ring_buffer_switch_new_end(struct lttng_ust_lib_ring_buffer *buf,
 static
 int lib_ring_buffer_try_switch_slow(enum switch_mode mode,
                                    struct lttng_ust_lib_ring_buffer *buf,
-                                   struct channel *chan,
+                                   struct lttng_ust_lib_ring_buffer_channel *chan,
                                    struct switch_offsets *offsets,
                                    uint64_t *tsc,
                                    struct lttng_ust_shm_handle *handle)
@@ -2005,7 +2039,7 @@ int lib_ring_buffer_try_switch_slow(enum switch_mode mode,
 void lib_ring_buffer_switch_slow(struct lttng_ust_lib_ring_buffer *buf, enum switch_mode mode,
                                 struct lttng_ust_shm_handle *handle)
 {
-       struct channel *chan;
+       struct lttng_ust_lib_ring_buffer_channel *chan;
        const struct lttng_ust_lib_ring_buffer_config *config;
        struct switch_offsets offsets;
        unsigned long oldidx;
@@ -2084,7 +2118,7 @@ bool handle_blocking_retry(int *timeout_left_ms)
  */
 static
 int lib_ring_buffer_try_reserve_slow(struct lttng_ust_lib_ring_buffer *buf,
-                                    struct channel *chan,
+                                    struct lttng_ust_lib_ring_buffer_channel *chan,
                                     struct switch_offsets *offsets,
                                     struct lttng_ust_lib_ring_buffer_ctx *ctx,
                                     void *client_ctx)
@@ -2278,7 +2312,7 @@ retry:
 int lib_ring_buffer_reserve_slow(struct lttng_ust_lib_ring_buffer_ctx *ctx,
                void *client_ctx)
 {
-       struct channel *chan = ctx->chan;
+       struct lttng_ust_lib_ring_buffer_channel *chan = ctx->chan;
        struct lttng_ust_shm_handle *handle = ctx->handle;
        const struct lttng_ust_lib_ring_buffer_config *config = &chan->backend.config;
        struct lttng_ust_lib_ring_buffer *buf;
@@ -2396,7 +2430,7 @@ void deliver_count_events(const struct lttng_ust_lib_ring_buffer_config *config,
 
 void lib_ring_buffer_check_deliver_slow(const struct lttng_ust_lib_ring_buffer_config *config,
                                   struct lttng_ust_lib_ring_buffer *buf,
-                                  struct channel *chan,
+                                  struct lttng_ust_lib_ring_buffer_channel *chan,
                                   unsigned long offset,
                                   unsigned long commit_count,
                                   unsigned long idx,
@@ -2445,14 +2479,26 @@ void lib_ring_buffer_check_deliver_slow(const struct lttng_ust_lib_ring_buffer_c
        if (caa_likely(v_cmpxchg(config, &cc_cold->cc_sb,
                                 old_commit_count, old_commit_count + 1)
                   == old_commit_count)) {
+               uint64_t *ts_end;
+
                /*
                 * Start of exclusive subbuffer access. We are
                 * guaranteed to be the last writer in this subbuffer
                 * and any other writer trying to access this subbuffer
                 * in this state is required to drop records.
+                *
+                * We can read the ts_end for the current sub-buffer
+                * which has been saved by the very last space
+                * reservation for the current sub-buffer.
+                *
+                * Order increment of commit counter before reading ts_end.
                 */
+               cmm_smp_mb();
+               ts_end = shmp_index(handle, buf->ts_end, idx);
+               if (!ts_end)
+                       return;
                deliver_count_events(config, buf, idx, handle);
-               config->cb.buffer_end(buf, tsc, idx,
+               config->cb.buffer_end(buf, *ts_end, idx,
                                      lib_ring_buffer_get_data_size(config,
                                                                buf,
                                                                idx,
This page took 0.031935 seconds and 4 git commands to generate.