shm: introduce shmp_index
[lttng-ust.git] / libringbuffer / backend_internal.h
index 30e32098af0de0832461dcdfb3df31c5f1b54ff8..ba0497f9f089f3cb6714f2e59603524eab5c725b 100644 (file)
@@ -190,8 +190,8 @@ void subbuffer_count_record(const struct lib_ring_buffer_config *config,
 {
        unsigned long sb_bindex;
 
-       sb_bindex = subbuffer_id_get_index(config, shmp(handle, bufb->buf_wsb)[idx].id);
-       v_inc(config, &shmp(handle, (shmp(handle, bufb->array)[sb_bindex]).shmp)->records_commit);
+       sb_bindex = subbuffer_id_get_index(config, shmp_index(handle, bufb->buf_wsb, idx)->id);
+       v_inc(config, &shmp(handle, shmp_index(handle, bufb->array, sb_bindex)->shmp)->records_commit);
 }
 
 /*
@@ -207,9 +207,9 @@ void subbuffer_consume_record(const struct lib_ring_buffer_config *config,
 
        sb_bindex = subbuffer_id_get_index(config, bufb->buf_rsb.id);
        CHAN_WARN_ON(shmp(handle, bufb->chan),
-                    !v_read(config, &shmp(handle, (shmp(handle, bufb->array)[sb_bindex]).shmp)->records_unread));
+                    !v_read(config, &shmp(handle, shmp_index(handle, bufb->array, sb_bindex)->shmp)->records_unread));
        /* Non-atomic decrement protected by exclusive subbuffer access */
-       _v_dec(config, &shmp(handle, (shmp(handle, bufb->array)[sb_bindex]).shmp)->records_unread);
+       _v_dec(config, &shmp(handle, shmp_index(handle, bufb->array, sb_bindex)->shmp)->records_unread);
        v_inc(config, &bufb->records_read);
 }
 
@@ -222,8 +222,8 @@ unsigned long subbuffer_get_records_count(
 {
        unsigned long sb_bindex;
 
-       sb_bindex = subbuffer_id_get_index(config, shmp(handle, bufb->buf_wsb)[idx].id);
-       return v_read(config, &shmp(handle, (shmp(handle, bufb->array)[sb_bindex]).shmp)->records_commit);
+       sb_bindex = subbuffer_id_get_index(config, shmp_index(handle, bufb->buf_wsb, idx)->id);
+       return v_read(config, &shmp(handle, shmp_index(handle, bufb->array, sb_bindex)->shmp)->records_commit);
 }
 
 /*
@@ -242,8 +242,8 @@ unsigned long subbuffer_count_records_overrun(
        struct lib_ring_buffer_backend_pages_shmp *pages;
        unsigned long overruns, sb_bindex;
 
-       sb_bindex = subbuffer_id_get_index(config, shmp(handle, bufb->buf_wsb)[idx].id);
-       pages = &shmp(handle, bufb->array)[sb_bindex];
+       sb_bindex = subbuffer_id_get_index(config, shmp_index(handle, bufb->buf_wsb, idx)->id);
+       pages = shmp_index(handle, bufb->array, sb_bindex);
        overruns = v_read(config, &shmp(handle, pages->shmp)->records_unread);
        v_set(config, &shmp(handle, pages->shmp)->records_unread,
              v_read(config, &shmp(handle, pages->shmp)->records_commit));
@@ -262,8 +262,8 @@ void subbuffer_set_data_size(const struct lib_ring_buffer_config *config,
        struct lib_ring_buffer_backend_pages_shmp *pages;
        unsigned long sb_bindex;
 
-       sb_bindex = subbuffer_id_get_index(config, shmp(handle, bufb->buf_wsb)[idx].id);
-       pages = &shmp(handle, bufb->array)[sb_bindex];
+       sb_bindex = subbuffer_id_get_index(config, shmp_index(handle, bufb->buf_wsb, idx)->id);
+       pages = shmp_index(handle, bufb->array, sb_bindex);
        shmp(handle, pages->shmp)->data_size = data_size;
 }
 
@@ -277,7 +277,7 @@ unsigned long subbuffer_get_read_data_size(
        unsigned long sb_bindex;
 
        sb_bindex = subbuffer_id_get_index(config, bufb->buf_rsb.id);
-       pages = &shmp(handle, bufb->array)[sb_bindex];
+       pages = shmp_index(handle, bufb->array, sb_bindex);
        return shmp(handle, pages->shmp)->data_size;
 }
 
@@ -291,8 +291,8 @@ unsigned long subbuffer_get_data_size(
        struct lib_ring_buffer_backend_pages_shmp *pages;
        unsigned long sb_bindex;
 
-       sb_bindex = subbuffer_id_get_index(config, shmp(handle, bufb->buf_wsb)[idx].id);
-       pages = &shmp(handle, bufb->array)[sb_bindex];
+       sb_bindex = subbuffer_id_get_index(config, shmp_index(handle, bufb->buf_wsb, idx)->id);
+       pages = shmp_index(handle, bufb->array, sb_bindex);
        return shmp(handle, pages->shmp)->data_size;
 }
 
@@ -315,7 +315,7 @@ void lib_ring_buffer_clear_noref(const struct lib_ring_buffer_config *config,
         * Performing a volatile access to read the sb_pages, because we want to
         * read a coherent version of the pointer and the associated noref flag.
         */
-       id = CMM_ACCESS_ONCE(shmp(handle, bufb->buf_wsb)[idx].id);
+       id = CMM_ACCESS_ONCE(shmp_index(handle, bufb->buf_wsb, idx)->id);
        for (;;) {
                /* This check is called on the fast path for each record. */
                if (likely(!subbuffer_id_is_noref(config, id))) {
@@ -329,7 +329,7 @@ void lib_ring_buffer_clear_noref(const struct lib_ring_buffer_config *config,
                }
                new_id = id;
                subbuffer_id_clear_noref(config, &new_id);
-               new_id = uatomic_cmpxchg(&shmp(handle, bufb->buf_wsb)[idx].id, id, new_id);
+               new_id = uatomic_cmpxchg(&shmp_index(handle, bufb->buf_wsb, idx)->id, id, new_id);
                if (likely(new_id == id))
                        break;
                id = new_id;
@@ -361,13 +361,13 @@ void lib_ring_buffer_set_noref_offset(const struct lib_ring_buffer_config *confi
         * readers of the noref flag.
         */
        CHAN_WARN_ON(shmp(handle, bufb->chan),
-                    subbuffer_id_is_noref(config, shmp(handle, bufb->buf_wsb)[idx].id));
+                    subbuffer_id_is_noref(config, shmp_index(handle, bufb->buf_wsb, idx)->id));
        /*
         * Memory barrier that ensures counter stores are ordered before set
         * noref and offset.
         */
        cmm_smp_mb();
-       subbuffer_id_set_noref_offset(config, &shmp(handle, bufb->buf_wsb)[idx].id, offset);
+       subbuffer_id_set_noref_offset(config, &shmp_index(handle, bufb->buf_wsb, idx)->id, offset);
 }
 
 /**
@@ -390,7 +390,7 @@ int update_read_sb_index(const struct lib_ring_buffer_config *config,
                 * old_wpage, because the value read will be confirmed by the
                 * following cmpxchg().
                 */
-               old_id = shmp(handle, bufb->buf_wsb)[consumed_idx].id;
+               old_id = shmp_index(handle, bufb->buf_wsb, consumed_idx)->id;
                if (unlikely(!subbuffer_id_is_noref(config, old_id)))
                        return -EAGAIN;
                /*
@@ -404,14 +404,14 @@ int update_read_sb_index(const struct lib_ring_buffer_config *config,
                             !subbuffer_id_is_noref(config, bufb->buf_rsb.id));
                subbuffer_id_set_noref_offset(config, &bufb->buf_rsb.id,
                                              consumed_count);
-               new_id = uatomic_cmpxchg(&shmp(handle, bufb->buf_wsb)[consumed_idx].id, old_id,
+               new_id = uatomic_cmpxchg(&shmp_index(handle, bufb->buf_wsb, consumed_idx)->id, old_id,
                                 bufb->buf_rsb.id);
                if (unlikely(old_id != new_id))
                        return -EAGAIN;
                bufb->buf_rsb.id = new_id;
        } else {
                /* No page exchange, use the writer page directly */
-               bufb->buf_rsb.id = shmp(handle, bufb->buf_wsb)[consumed_idx].id;
+               bufb->buf_rsb.id = shmp_index(handle, bufb->buf_wsb, consumed_idx)->id;
        }
        return 0;
 }
This page took 0.027067 seconds and 4 git commands to generate.