Store the instance id and packet_seq_num in indexes
[lttng-tools.git] / src / common / ust-consumer / ust-consumer.c
index d45707b763deaf653b3ed75f65c80a3f5ed42b55..5686fbd09fe58966cebf3f3be70bf8ec9acd339e 100644 (file)
@@ -16,7 +16,6 @@
  * 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
  */
 
-#define _GNU_SOURCE
 #define _LGPL_SOURCE
 #include <assert.h>
 #include <lttng/ust-ctl.h>
 #include <common/relayd/relayd.h>
 #include <common/compat/fcntl.h>
 #include <common/compat/endian.h>
-#include <common/consumer-metadata-cache.h>
-#include <common/consumer-stream.h>
-#include <common/consumer-timer.h>
+#include <common/consumer/consumer-metadata-cache.h>
+#include <common/consumer/consumer-stream.h>
+#include <common/consumer/consumer-timer.h>
 #include <common/utils.h>
 #include <common/index/index.h>
 
 #include "ust-consumer.h"
 
-#define UINT_MAX_STR_LEN 11    /* includes \0 */
+#define INT_MAX_STR_LEN 12     /* includes \0 */
 
 extern struct lttng_consumer_global_data consumer_data;
 extern int consumer_poll_timeout;
@@ -248,14 +247,14 @@ error:
 static
 int get_stream_shm_path(char *stream_shm_path, const char *shm_path, int cpu)
 {
-       char cpu_nr[UINT_MAX_STR_LEN];  /* unsigned int max len */
+       char cpu_nr[INT_MAX_STR_LEN];  /* int max len */
        int ret;
 
        strncpy(stream_shm_path, shm_path, PATH_MAX);
        stream_shm_path[PATH_MAX - 1] = '\0';
-       ret = snprintf(cpu_nr, UINT_MAX_STR_LEN, "%u", cpu);
-       if (ret != 1) {
-               ret = -1;
+       ret = snprintf(cpu_nr, INT_MAX_STR_LEN, "%i", cpu);
+       if (ret < 0) {
+               PERROR("snprintf");
                goto end;
        }
        strncat(stream_shm_path, cpu_nr,
@@ -1030,6 +1029,8 @@ static int snapshot_channel(uint64_t key, char *path, uint64_t relayd_id,
        DBG("UST consumer snapshot channel %" PRIu64, key);
 
        cds_list_for_each_entry(stream, &channel->streams.head, send_node) {
+               /* Are we at a position _before_ the first available packet ? */
+               bool before_first_packet = true;
 
                health_code_update();
 
@@ -1096,6 +1097,7 @@ static int snapshot_channel(uint64_t key, char *path, uint64_t relayd_id,
                while (consumed_pos < produced_pos) {
                        ssize_t read_len;
                        unsigned long len, padded_len;
+                       int lost_packet = 0;
 
                        health_code_update();
 
@@ -1109,6 +1111,15 @@ static int snapshot_channel(uint64_t key, char *path, uint64_t relayd_id,
                                }
                                DBG("UST consumer get subbuf failed. Skipping it.");
                                consumed_pos += stream->max_sb_size;
+
+                               /*
+                                * Start accounting lost packets only when we
+                                * already have extracted packets (to match the
+                                * content of the final snapshot).
+                                */
+                               if (!before_first_packet) {
+                                       lost_packet = 1;
+                               }
                                continue;
                        }
 
@@ -1144,6 +1155,16 @@ static int snapshot_channel(uint64_t key, char *path, uint64_t relayd_id,
                                goto error_close_stream;
                        }
                        consumed_pos += stream->max_sb_size;
+
+                       /*
+                        * Only account lost packets located between
+                        * succesfully extracted packets (do not account before
+                        * and after since they are not visible in the
+                        * resulting snapshot).
+                        */
+                       stream->chan->lost_packets += lost_packet;
+                       lost_packet = 0;
+                       before_first_packet = false;
                }
 
                /* Simply close the stream so we can use it on the next snapshot. */
@@ -1667,6 +1688,107 @@ int lttng_ustconsumer_recv_cmd(struct lttng_consumer_local_data *ctx,
                health_code_update();
                break;
        }
+       case LTTNG_CONSUMER_DISCARDED_EVENTS:
+       {
+               uint64_t ret;
+               struct lttng_ht_iter iter;
+               struct lttng_ht *ht;
+               struct lttng_consumer_stream *stream;
+               uint64_t id = msg.u.discarded_events.session_id;
+               uint64_t key = msg.u.discarded_events.channel_key;
+
+               DBG("UST consumer discarded events command for session id %"
+                               PRIu64, id);
+               rcu_read_lock();
+               pthread_mutex_lock(&consumer_data.lock);
+
+               ht = consumer_data.stream_list_ht;
+
+               /*
+                * We only need a reference to the channel, but they are not
+                * directly indexed, so we just use the first matching stream
+                * to extract the information we need, we default to 0 if not
+                * found (no events are dropped if the channel is not yet in
+                * use).
+                */
+               ret = 0;
+               cds_lfht_for_each_entry_duplicate(ht->ht,
+                               ht->hash_fct(&id, lttng_ht_seed),
+                               ht->match_fct, &id,
+                               &iter.iter, stream, node_session_id.node) {
+                       if (stream->chan->key == key) {
+                               ret = stream->chan->discarded_events;
+                               break;
+                       }
+               }
+               pthread_mutex_unlock(&consumer_data.lock);
+               rcu_read_unlock();
+
+               DBG("UST consumer discarded events command for session id %"
+                               PRIu64 ", channel key %" PRIu64, id, key);
+
+               health_code_update();
+
+               /* Send back returned value to session daemon */
+               ret = lttcomm_send_unix_sock(sock, &ret, sizeof(ret));
+               if (ret < 0) {
+                       PERROR("send discarded events");
+                       goto error_fatal;
+               }
+
+               break;
+       }
+       case LTTNG_CONSUMER_LOST_PACKETS:
+       {
+               int ret;
+               uint64_t lost_packets;
+               struct lttng_ht_iter iter;
+               struct lttng_ht *ht;
+               struct lttng_consumer_stream *stream;
+               uint64_t id = msg.u.lost_packets.session_id;
+               uint64_t key = msg.u.lost_packets.channel_key;
+
+               DBG("UST consumer lost packets command for session id %"
+                               PRIu64, id);
+               rcu_read_lock();
+               pthread_mutex_lock(&consumer_data.lock);
+
+               ht = consumer_data.stream_list_ht;
+
+               /*
+                * We only need a reference to the channel, but they are not
+                * directly indexed, so we just use the first matching stream
+                * to extract the information we need, we default to 0 if not
+                * found (no packets lost if the channel is not yet in use).
+                */
+               lost_packets = 0;
+               cds_lfht_for_each_entry_duplicate(ht->ht,
+                               ht->hash_fct(&id, lttng_ht_seed),
+                               ht->match_fct, &id,
+                               &iter.iter, stream, node_session_id.node) {
+                       if (stream->chan->key == key) {
+                               lost_packets = stream->chan->lost_packets;
+                               break;
+                       }
+               }
+               pthread_mutex_unlock(&consumer_data.lock);
+               rcu_read_unlock();
+
+               DBG("UST consumer lost packets command for session id %"
+                               PRIu64 ", channel key %" PRIu64, id, key);
+
+               health_code_update();
+
+               /* Send back returned value to session daemon */
+               ret = lttcomm_send_unix_sock(sock, &lost_packets,
+                               sizeof(lost_packets));
+               if (ret < 0) {
+                       PERROR("send lost packets");
+                       goto error_fatal;
+               }
+
+               break;
+       }
        default:
                break;
        }
@@ -1809,6 +1931,16 @@ int lttng_ustconsumer_get_current_timestamp(
        return ustctl_get_current_timestamp(stream->ustream, ts);
 }
 
+int lttng_ustconsumer_get_sequence_number(
+               struct lttng_consumer_stream *stream, uint64_t *seq)
+{
+       assert(stream);
+       assert(stream->ustream);
+       assert(seq);
+
+       return ustctl_get_sequence_number(stream->ustream, seq);
+}
+
 /*
  * Called when the stream signal the consumer that it has hang up.
  */
@@ -1947,6 +2079,20 @@ static int get_index_values(struct ctf_packet_index *index,
        }
        index->stream_id = htobe64(index->stream_id);
 
+       ret = ustctl_get_instance_id(ustream, &index->stream_instance_id);
+       if (ret < 0) {
+               PERROR("ustctl_get_instance_id");
+               goto error;
+       }
+       index->stream_instance_id = htobe64(index->stream_instance_id);
+
+       ret = ustctl_get_sequence_number(ustream, &index->packet_seq_num);
+       if (ret < 0) {
+               PERROR("ustctl_get_sequence_number");
+               goto error;
+       }
+       index->packet_seq_num = htobe64(index->packet_seq_num);
+
 error:
        return ret;
 }
@@ -2105,6 +2251,61 @@ end:
        return ret;
 }
 
+static
+int update_stream_stats(struct lttng_consumer_stream *stream)
+{
+       int ret;
+       uint64_t seq, discarded;
+
+       ret = ustctl_get_sequence_number(stream->ustream, &seq);
+       if (ret < 0) {
+               PERROR("ustctl_get_sequence_number");
+               goto end;
+       }
+       /*
+        * Start the sequence when we extract the first packet in case we don't
+        * start at 0 (for example if a consumer is not connected to the
+        * session immediately after the beginning).
+        */
+       if (stream->last_sequence_number == -1ULL) {
+               stream->last_sequence_number = seq;
+       } else if (seq > stream->last_sequence_number) {
+               stream->chan->lost_packets += seq -
+                               stream->last_sequence_number - 1;
+       } else {
+               /* seq <= last_sequence_number */
+               ERR("Sequence number inconsistent : prev = %" PRIu64
+                               ", current = %" PRIu64,
+                               stream->last_sequence_number, seq);
+               ret = -1;
+               goto end;
+       }
+       stream->last_sequence_number = seq;
+
+       ret = ustctl_get_events_discarded(stream->ustream, &discarded);
+       if (ret < 0) {
+               PERROR("kernctl_get_events_discarded");
+               goto end;
+       }
+       if (discarded < stream->last_discarded_events) {
+               /*
+                * Overflow has occured. We assume only one wrap-around
+                * has occured.
+                */
+               stream->chan->discarded_events +=
+                               (1ULL << (CAA_BITS_PER_LONG - 1)) -
+                               stream->last_discarded_events + discarded;
+       } else {
+               stream->chan->discarded_events += discarded -
+                               stream->last_discarded_events;
+       }
+       stream->last_discarded_events = discarded;
+       ret = 0;
+
+end:
+       return ret;
+}
+
 /*
  * Read subbuffer from the given stream.
  *
@@ -2188,6 +2389,13 @@ retry:
                if (ret < 0) {
                        goto end;
                }
+
+               /* Update the stream's sequence and discarded events count. */
+               ret = update_stream_stats(stream);
+               if (ret < 0) {
+                       PERROR("kernctl_get_events_discarded");
+                       goto end;
+               }
        } else {
                write_index = 0;
        }
This page took 0.02717 seconds and 4 git commands to generate.