Fix: update back the metadata len sent on failure
[lttng-tools.git] / src / bin / lttng-sessiond / ust-app.c
1 /*
2 * Copyright (C) 2011 - David Goulet <david.goulet@polymtl.ca>
3 *
4 * This program is free software; you can redistribute it and/or modify
5 * it under the terms of the GNU General Public License, version 2 only,
6 * as published by the Free Software Foundation.
7 *
8 * This program is distributed in the hope that it will be useful,
9 * but WITHOUT ANY WARRANTY; without even the implied warranty of
10 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
11 * GNU General Public License for more details.
12 *
13 * You should have received a copy of the GNU General Public License along
14 * with this program; if not, write to the Free Software Foundation, Inc.,
15 * 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
16 */
17
18 #define _GNU_SOURCE
19 #include <errno.h>
20 #include <inttypes.h>
21 #include <pthread.h>
22 #include <stdio.h>
23 #include <stdlib.h>
24 #include <string.h>
25 #include <sys/stat.h>
26 #include <sys/types.h>
27 #include <unistd.h>
28 #include <urcu/compiler.h>
29 #include <lttng/ust-error.h>
30 #include <signal.h>
31
32 #include <common/common.h>
33 #include <common/sessiond-comm/sessiond-comm.h>
34
35 #include "buffer-registry.h"
36 #include "fd-limit.h"
37 #include "health-sessiond.h"
38 #include "ust-app.h"
39 #include "ust-consumer.h"
40 #include "ust-ctl.h"
41 #include "utils.h"
42
43 /* Next available channel key. Access under next_channel_key_lock. */
44 static uint64_t _next_channel_key;
45 static pthread_mutex_t next_channel_key_lock = PTHREAD_MUTEX_INITIALIZER;
46
47 /* Next available session ID. Access under next_session_id_lock. */
48 static uint64_t _next_session_id;
49 static pthread_mutex_t next_session_id_lock = PTHREAD_MUTEX_INITIALIZER;
50
51 /*
52 * Return the incremented value of next_channel_key.
53 */
54 static uint64_t get_next_channel_key(void)
55 {
56 uint64_t ret;
57
58 pthread_mutex_lock(&next_channel_key_lock);
59 ret = ++_next_channel_key;
60 pthread_mutex_unlock(&next_channel_key_lock);
61 return ret;
62 }
63
64 /*
65 * Return the atomically incremented value of next_session_id.
66 */
67 static uint64_t get_next_session_id(void)
68 {
69 uint64_t ret;
70
71 pthread_mutex_lock(&next_session_id_lock);
72 ret = ++_next_session_id;
73 pthread_mutex_unlock(&next_session_id_lock);
74 return ret;
75 }
76
77 static void copy_channel_attr_to_ustctl(
78 struct ustctl_consumer_channel_attr *attr,
79 struct lttng_ust_channel_attr *uattr)
80 {
81 /* Copy event attributes since the layout is different. */
82 attr->subbuf_size = uattr->subbuf_size;
83 attr->num_subbuf = uattr->num_subbuf;
84 attr->overwrite = uattr->overwrite;
85 attr->switch_timer_interval = uattr->switch_timer_interval;
86 attr->read_timer_interval = uattr->read_timer_interval;
87 attr->output = uattr->output;
88 }
89
90 /*
91 * Match function for the hash table lookup.
92 *
93 * It matches an ust app event based on three attributes which are the event
94 * name, the filter bytecode and the loglevel.
95 */
96 static int ht_match_ust_app_event(struct cds_lfht_node *node, const void *_key)
97 {
98 struct ust_app_event *event;
99 const struct ust_app_ht_key *key;
100
101 assert(node);
102 assert(_key);
103
104 event = caa_container_of(node, struct ust_app_event, node.node);
105 key = _key;
106
107 /* Match the 4 elements of the key: name, filter, loglevel, exclusions */
108
109 /* Event name */
110 if (strncmp(event->attr.name, key->name, sizeof(event->attr.name)) != 0) {
111 goto no_match;
112 }
113
114 /* Event loglevel. */
115 if (event->attr.loglevel != key->loglevel) {
116 if (event->attr.loglevel_type == LTTNG_UST_LOGLEVEL_ALL
117 && key->loglevel == 0 && event->attr.loglevel == -1) {
118 /*
119 * Match is accepted. This is because on event creation, the
120 * loglevel is set to -1 if the event loglevel type is ALL so 0 and
121 * -1 are accepted for this loglevel type since 0 is the one set by
122 * the API when receiving an enable event.
123 */
124 } else {
125 goto no_match;
126 }
127 }
128
129 /* One of the filters is NULL, fail. */
130 if ((key->filter && !event->filter) || (!key->filter && event->filter)) {
131 goto no_match;
132 }
133
134 if (key->filter && event->filter) {
135 /* Both filters exists, check length followed by the bytecode. */
136 if (event->filter->len != key->filter->len ||
137 memcmp(event->filter->data, key->filter->data,
138 event->filter->len) != 0) {
139 goto no_match;
140 }
141 }
142
143 /* One of the exclusions is NULL, fail. */
144 if ((key->exclusion && !event->exclusion) || (!key->exclusion && event->exclusion)) {
145 goto no_match;
146 }
147
148 if (key->exclusion && event->exclusion) {
149 /* Both exclusions exists, check count followed by the names. */
150 if (event->exclusion->count != key->exclusion->count ||
151 memcmp(event->exclusion->names, key->exclusion->names,
152 event->exclusion->count * LTTNG_UST_SYM_NAME_LEN) != 0) {
153 goto no_match;
154 }
155 }
156
157
158 /* Match. */
159 return 1;
160
161 no_match:
162 return 0;
163 }
164
165 /*
166 * Unique add of an ust app event in the given ht. This uses the custom
167 * ht_match_ust_app_event match function and the event name as hash.
168 */
169 static void add_unique_ust_app_event(struct ust_app_channel *ua_chan,
170 struct ust_app_event *event)
171 {
172 struct cds_lfht_node *node_ptr;
173 struct ust_app_ht_key key;
174 struct lttng_ht *ht;
175
176 assert(ua_chan);
177 assert(ua_chan->events);
178 assert(event);
179
180 ht = ua_chan->events;
181 key.name = event->attr.name;
182 key.filter = event->filter;
183 key.loglevel = event->attr.loglevel;
184 key.exclusion = event->exclusion;
185
186 node_ptr = cds_lfht_add_unique(ht->ht,
187 ht->hash_fct(event->node.key, lttng_ht_seed),
188 ht_match_ust_app_event, &key, &event->node.node);
189 assert(node_ptr == &event->node.node);
190 }
191
192 /*
193 * Close the notify socket from the given RCU head object. This MUST be called
194 * through a call_rcu().
195 */
196 static void close_notify_sock_rcu(struct rcu_head *head)
197 {
198 int ret;
199 struct ust_app_notify_sock_obj *obj =
200 caa_container_of(head, struct ust_app_notify_sock_obj, head);
201
202 /* Must have a valid fd here. */
203 assert(obj->fd >= 0);
204
205 ret = close(obj->fd);
206 if (ret) {
207 ERR("close notify sock %d RCU", obj->fd);
208 }
209 lttng_fd_put(LTTNG_FD_APPS, 1);
210
211 free(obj);
212 }
213
214 /*
215 * Return the session registry according to the buffer type of the given
216 * session.
217 *
218 * A registry per UID object MUST exists before calling this function or else
219 * it assert() if not found. RCU read side lock must be acquired.
220 */
221 static struct ust_registry_session *get_session_registry(
222 struct ust_app_session *ua_sess)
223 {
224 struct ust_registry_session *registry = NULL;
225
226 assert(ua_sess);
227
228 switch (ua_sess->buffer_type) {
229 case LTTNG_BUFFER_PER_PID:
230 {
231 struct buffer_reg_pid *reg_pid = buffer_reg_pid_find(ua_sess->id);
232 if (!reg_pid) {
233 goto error;
234 }
235 registry = reg_pid->registry->reg.ust;
236 break;
237 }
238 case LTTNG_BUFFER_PER_UID:
239 {
240 struct buffer_reg_uid *reg_uid = buffer_reg_uid_find(
241 ua_sess->tracing_id, ua_sess->bits_per_long, ua_sess->uid);
242 if (!reg_uid) {
243 goto error;
244 }
245 registry = reg_uid->registry->reg.ust;
246 break;
247 }
248 default:
249 assert(0);
250 };
251
252 error:
253 return registry;
254 }
255
256 /*
257 * Delete ust context safely. RCU read lock must be held before calling
258 * this function.
259 */
260 static
261 void delete_ust_app_ctx(int sock, struct ust_app_ctx *ua_ctx)
262 {
263 int ret;
264
265 assert(ua_ctx);
266
267 if (ua_ctx->obj) {
268 ret = ustctl_release_object(sock, ua_ctx->obj);
269 if (ret < 0 && ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
270 ERR("UST app sock %d release ctx obj handle %d failed with ret %d",
271 sock, ua_ctx->obj->handle, ret);
272 }
273 free(ua_ctx->obj);
274 }
275 free(ua_ctx);
276 }
277
278 /*
279 * Delete ust app event safely. RCU read lock must be held before calling
280 * this function.
281 */
282 static
283 void delete_ust_app_event(int sock, struct ust_app_event *ua_event)
284 {
285 int ret;
286
287 assert(ua_event);
288
289 free(ua_event->filter);
290 if (ua_event->exclusion != NULL)
291 free(ua_event->exclusion);
292 if (ua_event->obj != NULL) {
293 ret = ustctl_release_object(sock, ua_event->obj);
294 if (ret < 0 && ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
295 ERR("UST app sock %d release event obj failed with ret %d",
296 sock, ret);
297 }
298 free(ua_event->obj);
299 }
300 free(ua_event);
301 }
302
303 /*
304 * Release ust data object of the given stream.
305 *
306 * Return 0 on success or else a negative value.
307 */
308 static int release_ust_app_stream(int sock, struct ust_app_stream *stream)
309 {
310 int ret = 0;
311
312 assert(stream);
313
314 if (stream->obj) {
315 ret = ustctl_release_object(sock, stream->obj);
316 if (ret < 0 && ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
317 ERR("UST app sock %d release stream obj failed with ret %d",
318 sock, ret);
319 }
320 lttng_fd_put(LTTNG_FD_APPS, 2);
321 free(stream->obj);
322 }
323
324 return ret;
325 }
326
327 /*
328 * Delete ust app stream safely. RCU read lock must be held before calling
329 * this function.
330 */
331 static
332 void delete_ust_app_stream(int sock, struct ust_app_stream *stream)
333 {
334 assert(stream);
335
336 (void) release_ust_app_stream(sock, stream);
337 free(stream);
338 }
339
340 /*
341 * We need to execute ht_destroy outside of RCU read-side critical
342 * section and outside of call_rcu thread, so we postpone its execution
343 * using ht_cleanup_push. It is simpler than to change the semantic of
344 * the many callers of delete_ust_app_session().
345 */
346 static
347 void delete_ust_app_channel_rcu(struct rcu_head *head)
348 {
349 struct ust_app_channel *ua_chan =
350 caa_container_of(head, struct ust_app_channel, rcu_head);
351
352 ht_cleanup_push(ua_chan->ctx);
353 ht_cleanup_push(ua_chan->events);
354 free(ua_chan);
355 }
356
357 /*
358 * Delete ust app channel safely. RCU read lock must be held before calling
359 * this function.
360 */
361 static
362 void delete_ust_app_channel(int sock, struct ust_app_channel *ua_chan,
363 struct ust_app *app)
364 {
365 int ret;
366 struct lttng_ht_iter iter;
367 struct ust_app_event *ua_event;
368 struct ust_app_ctx *ua_ctx;
369 struct ust_app_stream *stream, *stmp;
370 struct ust_registry_session *registry;
371
372 assert(ua_chan);
373
374 DBG3("UST app deleting channel %s", ua_chan->name);
375
376 /* Wipe stream */
377 cds_list_for_each_entry_safe(stream, stmp, &ua_chan->streams.head, list) {
378 cds_list_del(&stream->list);
379 delete_ust_app_stream(sock, stream);
380 }
381
382 /* Wipe context */
383 cds_lfht_for_each_entry(ua_chan->ctx->ht, &iter.iter, ua_ctx, node.node) {
384 cds_list_del(&ua_ctx->list);
385 ret = lttng_ht_del(ua_chan->ctx, &iter);
386 assert(!ret);
387 delete_ust_app_ctx(sock, ua_ctx);
388 }
389
390 /* Wipe events */
391 cds_lfht_for_each_entry(ua_chan->events->ht, &iter.iter, ua_event,
392 node.node) {
393 ret = lttng_ht_del(ua_chan->events, &iter);
394 assert(!ret);
395 delete_ust_app_event(sock, ua_event);
396 }
397
398 if (ua_chan->session->buffer_type == LTTNG_BUFFER_PER_PID) {
399 /* Wipe and free registry from session registry. */
400 registry = get_session_registry(ua_chan->session);
401 if (registry) {
402 ust_registry_channel_del_free(registry, ua_chan->key);
403 }
404 }
405
406 if (ua_chan->obj != NULL) {
407 /* Remove channel from application UST object descriptor. */
408 iter.iter.node = &ua_chan->ust_objd_node.node;
409 lttng_ht_del(app->ust_objd, &iter);
410 ret = ustctl_release_object(sock, ua_chan->obj);
411 if (ret < 0 && ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
412 ERR("UST app sock %d release channel obj failed with ret %d",
413 sock, ret);
414 }
415 lttng_fd_put(LTTNG_FD_APPS, 1);
416 free(ua_chan->obj);
417 }
418 call_rcu(&ua_chan->rcu_head, delete_ust_app_channel_rcu);
419 }
420
421 /*
422 * Push metadata to consumer socket.
423 *
424 * The socket lock MUST be acquired.
425 * The ust app session lock MUST be acquired.
426 *
427 * On success, return the len of metadata pushed or else a negative value.
428 */
429 ssize_t ust_app_push_metadata(struct ust_registry_session *registry,
430 struct consumer_socket *socket, int send_zero_data)
431 {
432 int ret;
433 char *metadata_str = NULL;
434 size_t len, offset;
435 ssize_t ret_val;
436
437 assert(registry);
438 assert(socket);
439
440 /*
441 * On a push metadata error either the consumer is dead or the metadata
442 * channel has been destroyed because its endpoint might have died (e.g:
443 * relayd). If so, the metadata closed flag is set to 1 so we deny pushing
444 * metadata again which is not valid anymore on the consumer side.
445 *
446 * The ust app session mutex locked allows us to make this check without
447 * the registry lock.
448 */
449 if (registry->metadata_closed) {
450 return -EPIPE;
451 }
452
453 pthread_mutex_lock(&registry->lock);
454
455 offset = registry->metadata_len_sent;
456 len = registry->metadata_len - registry->metadata_len_sent;
457 if (len == 0) {
458 DBG3("No metadata to push for metadata key %" PRIu64,
459 registry->metadata_key);
460 ret_val = len;
461 if (send_zero_data) {
462 DBG("No metadata to push");
463 goto push_data;
464 }
465 goto end;
466 }
467
468 /* Allocate only what we have to send. */
469 metadata_str = zmalloc(len);
470 if (!metadata_str) {
471 PERROR("zmalloc ust app metadata string");
472 ret_val = -ENOMEM;
473 goto error;
474 }
475 /* Copy what we haven't send out. */
476 memcpy(metadata_str, registry->metadata + offset, len);
477 registry->metadata_len_sent += len;
478
479 push_data:
480 pthread_mutex_unlock(&registry->lock);
481 ret = consumer_push_metadata(socket, registry->metadata_key,
482 metadata_str, len, offset);
483 if (ret < 0) {
484 /*
485 * There is an acceptable race here between the registry metadata key
486 * assignment and the creation on the consumer. The session daemon can
487 * concurrently push metadata for this registry while being created on
488 * the consumer since the metadata key of the registry is assigned
489 * *before* it is setup to avoid the consumer to ask for metadata that
490 * could possibly be not found in the session daemon.
491 *
492 * The metadata will get pushed either by the session being stopped or
493 * the consumer requesting metadata if that race is triggered.
494 */
495 if (ret == -LTTCOMM_CONSUMERD_CHANNEL_FAIL) {
496 ret = 0;
497 }
498
499 /* Update back the actual metadata len sent since it failed here. */
500 pthread_mutex_lock(&registry->lock);
501 registry->metadata_len_sent -= len;
502 pthread_mutex_unlock(&registry->lock);
503 ret_val = ret;
504 goto error_push;
505 }
506
507 free(metadata_str);
508 return len;
509
510 end:
511 error:
512 pthread_mutex_unlock(&registry->lock);
513 error_push:
514 free(metadata_str);
515 return ret_val;
516 }
517
518 /*
519 * For a given application and session, push metadata to consumer. The session
520 * lock MUST be acquired here before calling this.
521 * Either sock or consumer is required : if sock is NULL, the default
522 * socket to send the metadata is retrieved from consumer, if sock
523 * is not NULL we use it to send the metadata.
524 *
525 * Return 0 on success else a negative error.
526 */
527 static int push_metadata(struct ust_registry_session *registry,
528 struct consumer_output *consumer)
529 {
530 int ret_val;
531 ssize_t ret;
532 struct consumer_socket *socket;
533
534 assert(registry);
535 assert(consumer);
536
537 rcu_read_lock();
538
539 /*
540 * Means that no metadata was assigned to the session. This can happens if
541 * no start has been done previously.
542 */
543 if (!registry->metadata_key) {
544 ret_val = 0;
545 goto end_rcu_unlock;
546 }
547
548 /* Get consumer socket to use to push the metadata.*/
549 socket = consumer_find_socket_by_bitness(registry->bits_per_long,
550 consumer);
551 if (!socket) {
552 ret_val = -1;
553 goto error_rcu_unlock;
554 }
555
556 /*
557 * TODO: Currently, we hold the socket lock around sampling of the next
558 * metadata segment to ensure we send metadata over the consumer socket in
559 * the correct order. This makes the registry lock nest inside the socket
560 * lock.
561 *
562 * Please note that this is a temporary measure: we should move this lock
563 * back into ust_consumer_push_metadata() when the consumer gets the
564 * ability to reorder the metadata it receives.
565 */
566 pthread_mutex_lock(socket->lock);
567 ret = ust_app_push_metadata(registry, socket, 0);
568 pthread_mutex_unlock(socket->lock);
569 if (ret < 0) {
570 ret_val = ret;
571 goto error_rcu_unlock;
572 }
573
574 rcu_read_unlock();
575 return 0;
576
577 error_rcu_unlock:
578 /*
579 * On error, flag the registry that the metadata is closed. We were unable
580 * to push anything and this means that either the consumer is not
581 * responding or the metadata cache has been destroyed on the consumer.
582 */
583 registry->metadata_closed = 1;
584 end_rcu_unlock:
585 rcu_read_unlock();
586 return ret_val;
587 }
588
589 /*
590 * Send to the consumer a close metadata command for the given session. Once
591 * done, the metadata channel is deleted and the session metadata pointer is
592 * nullified. The session lock MUST be acquired here unless the application is
593 * in the destroy path.
594 *
595 * Return 0 on success else a negative value.
596 */
597 static int close_metadata(struct ust_registry_session *registry,
598 struct consumer_output *consumer)
599 {
600 int ret;
601 struct consumer_socket *socket;
602
603 assert(registry);
604 assert(consumer);
605
606 rcu_read_lock();
607
608 if (!registry->metadata_key || registry->metadata_closed) {
609 ret = 0;
610 goto end;
611 }
612
613 /* Get consumer socket to use to push the metadata.*/
614 socket = consumer_find_socket_by_bitness(registry->bits_per_long,
615 consumer);
616 if (!socket) {
617 ret = -1;
618 goto error;
619 }
620
621 ret = consumer_close_metadata(socket, registry->metadata_key);
622 if (ret < 0) {
623 goto error;
624 }
625
626 error:
627 /*
628 * Metadata closed. Even on error this means that the consumer is not
629 * responding or not found so either way a second close should NOT be emit
630 * for this registry.
631 */
632 registry->metadata_closed = 1;
633 end:
634 rcu_read_unlock();
635 return ret;
636 }
637
638 /*
639 * We need to execute ht_destroy outside of RCU read-side critical
640 * section and outside of call_rcu thread, so we postpone its execution
641 * using ht_cleanup_push. It is simpler than to change the semantic of
642 * the many callers of delete_ust_app_session().
643 */
644 static
645 void delete_ust_app_session_rcu(struct rcu_head *head)
646 {
647 struct ust_app_session *ua_sess =
648 caa_container_of(head, struct ust_app_session, rcu_head);
649
650 ht_cleanup_push(ua_sess->channels);
651 free(ua_sess);
652 }
653
654 /*
655 * Delete ust app session safely. RCU read lock must be held before calling
656 * this function.
657 */
658 static
659 void delete_ust_app_session(int sock, struct ust_app_session *ua_sess,
660 struct ust_app *app)
661 {
662 int ret;
663 struct lttng_ht_iter iter;
664 struct ust_app_channel *ua_chan;
665 struct ust_registry_session *registry;
666
667 assert(ua_sess);
668
669 pthread_mutex_lock(&ua_sess->lock);
670
671 registry = get_session_registry(ua_sess);
672 if (registry && !registry->metadata_closed) {
673 /* Push metadata for application before freeing the application. */
674 (void) push_metadata(registry, ua_sess->consumer);
675
676 /*
677 * Don't ask to close metadata for global per UID buffers. Close
678 * metadata only on destroy trace session in this case. Also, the
679 * previous push metadata could have flag the metadata registry to
680 * close so don't send a close command if closed.
681 */
682 if (ua_sess->buffer_type != LTTNG_BUFFER_PER_UID &&
683 !registry->metadata_closed) {
684 /* And ask to close it for this session registry. */
685 (void) close_metadata(registry, ua_sess->consumer);
686 }
687 }
688
689 cds_lfht_for_each_entry(ua_sess->channels->ht, &iter.iter, ua_chan,
690 node.node) {
691 ret = lttng_ht_del(ua_sess->channels, &iter);
692 assert(!ret);
693 delete_ust_app_channel(sock, ua_chan, app);
694 }
695
696 /* In case of per PID, the registry is kept in the session. */
697 if (ua_sess->buffer_type == LTTNG_BUFFER_PER_PID) {
698 struct buffer_reg_pid *reg_pid = buffer_reg_pid_find(ua_sess->id);
699 if (reg_pid) {
700 buffer_reg_pid_remove(reg_pid);
701 buffer_reg_pid_destroy(reg_pid);
702 }
703 }
704
705 if (ua_sess->handle != -1) {
706 ret = ustctl_release_handle(sock, ua_sess->handle);
707 if (ret < 0 && ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
708 ERR("UST app sock %d release session handle failed with ret %d",
709 sock, ret);
710 }
711 }
712 pthread_mutex_unlock(&ua_sess->lock);
713
714 call_rcu(&ua_sess->rcu_head, delete_ust_app_session_rcu);
715 }
716
717 /*
718 * Delete a traceable application structure from the global list. Never call
719 * this function outside of a call_rcu call.
720 *
721 * RCU read side lock should _NOT_ be held when calling this function.
722 */
723 static
724 void delete_ust_app(struct ust_app *app)
725 {
726 int ret, sock;
727 struct ust_app_session *ua_sess, *tmp_ua_sess;
728
729 /* Delete ust app sessions info */
730 sock = app->sock;
731 app->sock = -1;
732
733 /* Wipe sessions */
734 cds_list_for_each_entry_safe(ua_sess, tmp_ua_sess, &app->teardown_head,
735 teardown_node) {
736 /* Free every object in the session and the session. */
737 rcu_read_lock();
738 delete_ust_app_session(sock, ua_sess, app);
739 rcu_read_unlock();
740 }
741
742 ht_cleanup_push(app->sessions);
743 ht_cleanup_push(app->ust_objd);
744
745 /*
746 * Wait until we have deleted the application from the sock hash table
747 * before closing this socket, otherwise an application could re-use the
748 * socket ID and race with the teardown, using the same hash table entry.
749 *
750 * It's OK to leave the close in call_rcu. We want it to stay unique for
751 * all RCU readers that could run concurrently with unregister app,
752 * therefore we _need_ to only close that socket after a grace period. So
753 * it should stay in this RCU callback.
754 *
755 * This close() is a very important step of the synchronization model so
756 * every modification to this function must be carefully reviewed.
757 */
758 ret = close(sock);
759 if (ret) {
760 PERROR("close");
761 }
762 lttng_fd_put(LTTNG_FD_APPS, 1);
763
764 DBG2("UST app pid %d deleted", app->pid);
765 free(app);
766 }
767
768 /*
769 * URCU intermediate call to delete an UST app.
770 */
771 static
772 void delete_ust_app_rcu(struct rcu_head *head)
773 {
774 struct lttng_ht_node_ulong *node =
775 caa_container_of(head, struct lttng_ht_node_ulong, head);
776 struct ust_app *app =
777 caa_container_of(node, struct ust_app, pid_n);
778
779 DBG3("Call RCU deleting app PID %d", app->pid);
780 delete_ust_app(app);
781 }
782
783 /*
784 * Delete the session from the application ht and delete the data structure by
785 * freeing every object inside and releasing them.
786 */
787 static void destroy_app_session(struct ust_app *app,
788 struct ust_app_session *ua_sess)
789 {
790 int ret;
791 struct lttng_ht_iter iter;
792
793 assert(app);
794 assert(ua_sess);
795
796 iter.iter.node = &ua_sess->node.node;
797 ret = lttng_ht_del(app->sessions, &iter);
798 if (ret) {
799 /* Already scheduled for teardown. */
800 goto end;
801 }
802
803 /* Once deleted, free the data structure. */
804 delete_ust_app_session(app->sock, ua_sess, app);
805
806 end:
807 return;
808 }
809
810 /*
811 * Alloc new UST app session.
812 */
813 static
814 struct ust_app_session *alloc_ust_app_session(struct ust_app *app)
815 {
816 struct ust_app_session *ua_sess;
817
818 /* Init most of the default value by allocating and zeroing */
819 ua_sess = zmalloc(sizeof(struct ust_app_session));
820 if (ua_sess == NULL) {
821 PERROR("malloc");
822 goto error_free;
823 }
824
825 ua_sess->handle = -1;
826 ua_sess->channels = lttng_ht_new(0, LTTNG_HT_TYPE_STRING);
827 pthread_mutex_init(&ua_sess->lock, NULL);
828
829 return ua_sess;
830
831 error_free:
832 return NULL;
833 }
834
835 /*
836 * Alloc new UST app channel.
837 */
838 static
839 struct ust_app_channel *alloc_ust_app_channel(char *name,
840 struct ust_app_session *ua_sess,
841 struct lttng_ust_channel_attr *attr)
842 {
843 struct ust_app_channel *ua_chan;
844
845 /* Init most of the default value by allocating and zeroing */
846 ua_chan = zmalloc(sizeof(struct ust_app_channel));
847 if (ua_chan == NULL) {
848 PERROR("malloc");
849 goto error;
850 }
851
852 /* Setup channel name */
853 strncpy(ua_chan->name, name, sizeof(ua_chan->name));
854 ua_chan->name[sizeof(ua_chan->name) - 1] = '\0';
855
856 ua_chan->enabled = 1;
857 ua_chan->handle = -1;
858 ua_chan->session = ua_sess;
859 ua_chan->key = get_next_channel_key();
860 ua_chan->ctx = lttng_ht_new(0, LTTNG_HT_TYPE_ULONG);
861 ua_chan->events = lttng_ht_new(0, LTTNG_HT_TYPE_STRING);
862 lttng_ht_node_init_str(&ua_chan->node, ua_chan->name);
863
864 CDS_INIT_LIST_HEAD(&ua_chan->streams.head);
865 CDS_INIT_LIST_HEAD(&ua_chan->ctx_list);
866
867 /* Copy attributes */
868 if (attr) {
869 /* Translate from lttng_ust_channel to ustctl_consumer_channel_attr. */
870 ua_chan->attr.subbuf_size = attr->subbuf_size;
871 ua_chan->attr.num_subbuf = attr->num_subbuf;
872 ua_chan->attr.overwrite = attr->overwrite;
873 ua_chan->attr.switch_timer_interval = attr->switch_timer_interval;
874 ua_chan->attr.read_timer_interval = attr->read_timer_interval;
875 ua_chan->attr.output = attr->output;
876 }
877 /* By default, the channel is a per cpu channel. */
878 ua_chan->attr.type = LTTNG_UST_CHAN_PER_CPU;
879
880 DBG3("UST app channel %s allocated", ua_chan->name);
881
882 return ua_chan;
883
884 error:
885 return NULL;
886 }
887
888 /*
889 * Allocate and initialize a UST app stream.
890 *
891 * Return newly allocated stream pointer or NULL on error.
892 */
893 struct ust_app_stream *ust_app_alloc_stream(void)
894 {
895 struct ust_app_stream *stream = NULL;
896
897 stream = zmalloc(sizeof(*stream));
898 if (stream == NULL) {
899 PERROR("zmalloc ust app stream");
900 goto error;
901 }
902
903 /* Zero could be a valid value for a handle so flag it to -1. */
904 stream->handle = -1;
905
906 error:
907 return stream;
908 }
909
910 /*
911 * Alloc new UST app event.
912 */
913 static
914 struct ust_app_event *alloc_ust_app_event(char *name,
915 struct lttng_ust_event *attr)
916 {
917 struct ust_app_event *ua_event;
918
919 /* Init most of the default value by allocating and zeroing */
920 ua_event = zmalloc(sizeof(struct ust_app_event));
921 if (ua_event == NULL) {
922 PERROR("malloc");
923 goto error;
924 }
925
926 ua_event->enabled = 1;
927 strncpy(ua_event->name, name, sizeof(ua_event->name));
928 ua_event->name[sizeof(ua_event->name) - 1] = '\0';
929 lttng_ht_node_init_str(&ua_event->node, ua_event->name);
930
931 /* Copy attributes */
932 if (attr) {
933 memcpy(&ua_event->attr, attr, sizeof(ua_event->attr));
934 }
935
936 DBG3("UST app event %s allocated", ua_event->name);
937
938 return ua_event;
939
940 error:
941 return NULL;
942 }
943
944 /*
945 * Alloc new UST app context.
946 */
947 static
948 struct ust_app_ctx *alloc_ust_app_ctx(struct lttng_ust_context *uctx)
949 {
950 struct ust_app_ctx *ua_ctx;
951
952 ua_ctx = zmalloc(sizeof(struct ust_app_ctx));
953 if (ua_ctx == NULL) {
954 goto error;
955 }
956
957 CDS_INIT_LIST_HEAD(&ua_ctx->list);
958
959 if (uctx) {
960 memcpy(&ua_ctx->ctx, uctx, sizeof(ua_ctx->ctx));
961 }
962
963 DBG3("UST app context %d allocated", ua_ctx->ctx.ctx);
964
965 error:
966 return ua_ctx;
967 }
968
969 /*
970 * Allocate a filter and copy the given original filter.
971 *
972 * Return allocated filter or NULL on error.
973 */
974 static struct lttng_ust_filter_bytecode *alloc_copy_ust_app_filter(
975 struct lttng_ust_filter_bytecode *orig_f)
976 {
977 struct lttng_ust_filter_bytecode *filter = NULL;
978
979 /* Copy filter bytecode */
980 filter = zmalloc(sizeof(*filter) + orig_f->len);
981 if (!filter) {
982 PERROR("zmalloc alloc ust app filter");
983 goto error;
984 }
985
986 memcpy(filter, orig_f, sizeof(*filter) + orig_f->len);
987
988 error:
989 return filter;
990 }
991
992 /*
993 * Find an ust_app using the sock and return it. RCU read side lock must be
994 * held before calling this helper function.
995 */
996 struct ust_app *ust_app_find_by_sock(int sock)
997 {
998 struct lttng_ht_node_ulong *node;
999 struct lttng_ht_iter iter;
1000
1001 lttng_ht_lookup(ust_app_ht_by_sock, (void *)((unsigned long) sock), &iter);
1002 node = lttng_ht_iter_get_node_ulong(&iter);
1003 if (node == NULL) {
1004 DBG2("UST app find by sock %d not found", sock);
1005 goto error;
1006 }
1007
1008 return caa_container_of(node, struct ust_app, sock_n);
1009
1010 error:
1011 return NULL;
1012 }
1013
1014 /*
1015 * Find an ust_app using the notify sock and return it. RCU read side lock must
1016 * be held before calling this helper function.
1017 */
1018 static struct ust_app *find_app_by_notify_sock(int sock)
1019 {
1020 struct lttng_ht_node_ulong *node;
1021 struct lttng_ht_iter iter;
1022
1023 lttng_ht_lookup(ust_app_ht_by_notify_sock, (void *)((unsigned long) sock),
1024 &iter);
1025 node = lttng_ht_iter_get_node_ulong(&iter);
1026 if (node == NULL) {
1027 DBG2("UST app find by notify sock %d not found", sock);
1028 goto error;
1029 }
1030
1031 return caa_container_of(node, struct ust_app, notify_sock_n);
1032
1033 error:
1034 return NULL;
1035 }
1036
1037 /*
1038 * Lookup for an ust app event based on event name, filter bytecode and the
1039 * event loglevel.
1040 *
1041 * Return an ust_app_event object or NULL on error.
1042 */
1043 static struct ust_app_event *find_ust_app_event(struct lttng_ht *ht,
1044 char *name, struct lttng_ust_filter_bytecode *filter, int loglevel,
1045 const struct lttng_event_exclusion *exclusion)
1046 {
1047 struct lttng_ht_iter iter;
1048 struct lttng_ht_node_str *node;
1049 struct ust_app_event *event = NULL;
1050 struct ust_app_ht_key key;
1051
1052 assert(name);
1053 assert(ht);
1054
1055 /* Setup key for event lookup. */
1056 key.name = name;
1057 key.filter = filter;
1058 key.loglevel = loglevel;
1059 /* lttng_event_exclusion and lttng_ust_event_exclusion structures are similar */
1060 key.exclusion = (struct lttng_ust_event_exclusion *)exclusion;
1061
1062 /* Lookup using the event name as hash and a custom match fct. */
1063 cds_lfht_lookup(ht->ht, ht->hash_fct((void *) name, lttng_ht_seed),
1064 ht_match_ust_app_event, &key, &iter.iter);
1065 node = lttng_ht_iter_get_node_str(&iter);
1066 if (node == NULL) {
1067 goto end;
1068 }
1069
1070 event = caa_container_of(node, struct ust_app_event, node);
1071
1072 end:
1073 return event;
1074 }
1075
1076 /*
1077 * Create the channel context on the tracer.
1078 *
1079 * Called with UST app session lock held.
1080 */
1081 static
1082 int create_ust_channel_context(struct ust_app_channel *ua_chan,
1083 struct ust_app_ctx *ua_ctx, struct ust_app *app)
1084 {
1085 int ret;
1086
1087 health_code_update();
1088
1089 ret = ustctl_add_context(app->sock, &ua_ctx->ctx,
1090 ua_chan->obj, &ua_ctx->obj);
1091 if (ret < 0) {
1092 if (ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
1093 ERR("UST app create channel context failed for app (pid: %d) "
1094 "with ret %d", app->pid, ret);
1095 } else {
1096 /*
1097 * This is normal behavior, an application can die during the
1098 * creation process. Don't report an error so the execution can
1099 * continue normally.
1100 */
1101 ret = 0;
1102 DBG3("UST app disable event failed. Application is dead.");
1103 }
1104 goto error;
1105 }
1106
1107 ua_ctx->handle = ua_ctx->obj->handle;
1108
1109 DBG2("UST app context handle %d created successfully for channel %s",
1110 ua_ctx->handle, ua_chan->name);
1111
1112 error:
1113 health_code_update();
1114 return ret;
1115 }
1116
1117 /*
1118 * Set the filter on the tracer.
1119 */
1120 static
1121 int set_ust_event_filter(struct ust_app_event *ua_event,
1122 struct ust_app *app)
1123 {
1124 int ret;
1125
1126 health_code_update();
1127
1128 if (!ua_event->filter) {
1129 ret = 0;
1130 goto error;
1131 }
1132
1133 ret = ustctl_set_filter(app->sock, ua_event->filter,
1134 ua_event->obj);
1135 if (ret < 0) {
1136 if (ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
1137 ERR("UST app event %s filter failed for app (pid: %d) "
1138 "with ret %d", ua_event->attr.name, app->pid, ret);
1139 } else {
1140 /*
1141 * This is normal behavior, an application can die during the
1142 * creation process. Don't report an error so the execution can
1143 * continue normally.
1144 */
1145 ret = 0;
1146 DBG3("UST app filter event failed. Application is dead.");
1147 }
1148 goto error;
1149 }
1150
1151 DBG2("UST filter set successfully for event %s", ua_event->name);
1152
1153 error:
1154 health_code_update();
1155 return ret;
1156 }
1157
1158 /*
1159 * Set event exclusions on the tracer.
1160 */
1161 static
1162 int set_ust_event_exclusion(struct ust_app_event *ua_event,
1163 struct ust_app *app)
1164 {
1165 int ret;
1166
1167 health_code_update();
1168
1169 if (!ua_event->exclusion || !ua_event->exclusion->count) {
1170 ret = 0;
1171 goto error;
1172 }
1173
1174 ret = ustctl_set_exclusion(app->sock, ua_event->exclusion,
1175 ua_event->obj);
1176 if (ret < 0) {
1177 if (ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
1178 ERR("UST app event %s exclusions failed for app (pid: %d) "
1179 "with ret %d", ua_event->attr.name, app->pid, ret);
1180 } else {
1181 /*
1182 * This is normal behavior, an application can die during the
1183 * creation process. Don't report an error so the execution can
1184 * continue normally.
1185 */
1186 ret = 0;
1187 DBG3("UST app event exclusion failed. Application is dead.");
1188 }
1189 goto error;
1190 }
1191
1192 DBG2("UST exclusion set successfully for event %s", ua_event->name);
1193
1194 error:
1195 health_code_update();
1196 return ret;
1197 }
1198
1199 /*
1200 * Disable the specified event on to UST tracer for the UST session.
1201 */
1202 static int disable_ust_event(struct ust_app *app,
1203 struct ust_app_session *ua_sess, struct ust_app_event *ua_event)
1204 {
1205 int ret;
1206
1207 health_code_update();
1208
1209 ret = ustctl_disable(app->sock, ua_event->obj);
1210 if (ret < 0) {
1211 if (ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
1212 ERR("UST app event %s disable failed for app (pid: %d) "
1213 "and session handle %d with ret %d",
1214 ua_event->attr.name, app->pid, ua_sess->handle, ret);
1215 } else {
1216 /*
1217 * This is normal behavior, an application can die during the
1218 * creation process. Don't report an error so the execution can
1219 * continue normally.
1220 */
1221 ret = 0;
1222 DBG3("UST app disable event failed. Application is dead.");
1223 }
1224 goto error;
1225 }
1226
1227 DBG2("UST app event %s disabled successfully for app (pid: %d)",
1228 ua_event->attr.name, app->pid);
1229
1230 error:
1231 health_code_update();
1232 return ret;
1233 }
1234
1235 /*
1236 * Disable the specified channel on to UST tracer for the UST session.
1237 */
1238 static int disable_ust_channel(struct ust_app *app,
1239 struct ust_app_session *ua_sess, struct ust_app_channel *ua_chan)
1240 {
1241 int ret;
1242
1243 health_code_update();
1244
1245 ret = ustctl_disable(app->sock, ua_chan->obj);
1246 if (ret < 0) {
1247 if (ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
1248 ERR("UST app channel %s disable failed for app (pid: %d) "
1249 "and session handle %d with ret %d",
1250 ua_chan->name, app->pid, ua_sess->handle, ret);
1251 } else {
1252 /*
1253 * This is normal behavior, an application can die during the
1254 * creation process. Don't report an error so the execution can
1255 * continue normally.
1256 */
1257 ret = 0;
1258 DBG3("UST app disable channel failed. Application is dead.");
1259 }
1260 goto error;
1261 }
1262
1263 DBG2("UST app channel %s disabled successfully for app (pid: %d)",
1264 ua_chan->name, app->pid);
1265
1266 error:
1267 health_code_update();
1268 return ret;
1269 }
1270
1271 /*
1272 * Enable the specified channel on to UST tracer for the UST session.
1273 */
1274 static int enable_ust_channel(struct ust_app *app,
1275 struct ust_app_session *ua_sess, struct ust_app_channel *ua_chan)
1276 {
1277 int ret;
1278
1279 health_code_update();
1280
1281 ret = ustctl_enable(app->sock, ua_chan->obj);
1282 if (ret < 0) {
1283 if (ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
1284 ERR("UST app channel %s enable failed for app (pid: %d) "
1285 "and session handle %d with ret %d",
1286 ua_chan->name, app->pid, ua_sess->handle, ret);
1287 } else {
1288 /*
1289 * This is normal behavior, an application can die during the
1290 * creation process. Don't report an error so the execution can
1291 * continue normally.
1292 */
1293 ret = 0;
1294 DBG3("UST app enable channel failed. Application is dead.");
1295 }
1296 goto error;
1297 }
1298
1299 ua_chan->enabled = 1;
1300
1301 DBG2("UST app channel %s enabled successfully for app (pid: %d)",
1302 ua_chan->name, app->pid);
1303
1304 error:
1305 health_code_update();
1306 return ret;
1307 }
1308
1309 /*
1310 * Enable the specified event on to UST tracer for the UST session.
1311 */
1312 static int enable_ust_event(struct ust_app *app,
1313 struct ust_app_session *ua_sess, struct ust_app_event *ua_event)
1314 {
1315 int ret;
1316
1317 health_code_update();
1318
1319 ret = ustctl_enable(app->sock, ua_event->obj);
1320 if (ret < 0) {
1321 if (ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
1322 ERR("UST app event %s enable failed for app (pid: %d) "
1323 "and session handle %d with ret %d",
1324 ua_event->attr.name, app->pid, ua_sess->handle, ret);
1325 } else {
1326 /*
1327 * This is normal behavior, an application can die during the
1328 * creation process. Don't report an error so the execution can
1329 * continue normally.
1330 */
1331 ret = 0;
1332 DBG3("UST app enable event failed. Application is dead.");
1333 }
1334 goto error;
1335 }
1336
1337 DBG2("UST app event %s enabled successfully for app (pid: %d)",
1338 ua_event->attr.name, app->pid);
1339
1340 error:
1341 health_code_update();
1342 return ret;
1343 }
1344
1345 /*
1346 * Send channel and stream buffer to application.
1347 *
1348 * Return 0 on success. On error, a negative value is returned.
1349 */
1350 static int send_channel_pid_to_ust(struct ust_app *app,
1351 struct ust_app_session *ua_sess, struct ust_app_channel *ua_chan)
1352 {
1353 int ret;
1354 struct ust_app_stream *stream, *stmp;
1355
1356 assert(app);
1357 assert(ua_sess);
1358 assert(ua_chan);
1359
1360 health_code_update();
1361
1362 DBG("UST app sending channel %s to UST app sock %d", ua_chan->name,
1363 app->sock);
1364
1365 /* Send channel to the application. */
1366 ret = ust_consumer_send_channel_to_ust(app, ua_sess, ua_chan);
1367 if (ret < 0) {
1368 goto error;
1369 }
1370
1371 health_code_update();
1372
1373 /* Send all streams to application. */
1374 cds_list_for_each_entry_safe(stream, stmp, &ua_chan->streams.head, list) {
1375 ret = ust_consumer_send_stream_to_ust(app, ua_chan, stream);
1376 if (ret < 0) {
1377 goto error;
1378 }
1379 /* We don't need the stream anymore once sent to the tracer. */
1380 cds_list_del(&stream->list);
1381 delete_ust_app_stream(-1, stream);
1382 }
1383 /* Flag the channel that it is sent to the application. */
1384 ua_chan->is_sent = 1;
1385
1386 error:
1387 health_code_update();
1388 return ret;
1389 }
1390
1391 /*
1392 * Create the specified event onto the UST tracer for a UST session.
1393 *
1394 * Should be called with session mutex held.
1395 */
1396 static
1397 int create_ust_event(struct ust_app *app, struct ust_app_session *ua_sess,
1398 struct ust_app_channel *ua_chan, struct ust_app_event *ua_event)
1399 {
1400 int ret = 0;
1401
1402 health_code_update();
1403
1404 /* Create UST event on tracer */
1405 ret = ustctl_create_event(app->sock, &ua_event->attr, ua_chan->obj,
1406 &ua_event->obj);
1407 if (ret < 0) {
1408 if (ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
1409 ERR("Error ustctl create event %s for app pid: %d with ret %d",
1410 ua_event->attr.name, app->pid, ret);
1411 } else {
1412 /*
1413 * This is normal behavior, an application can die during the
1414 * creation process. Don't report an error so the execution can
1415 * continue normally.
1416 */
1417 ret = 0;
1418 DBG3("UST app create event failed. Application is dead.");
1419 }
1420 goto error;
1421 }
1422
1423 ua_event->handle = ua_event->obj->handle;
1424
1425 DBG2("UST app event %s created successfully for pid:%d",
1426 ua_event->attr.name, app->pid);
1427
1428 health_code_update();
1429
1430 /* Set filter if one is present. */
1431 if (ua_event->filter) {
1432 ret = set_ust_event_filter(ua_event, app);
1433 if (ret < 0) {
1434 goto error;
1435 }
1436 }
1437
1438 /* Set exclusions for the event */
1439 if (ua_event->exclusion) {
1440 ret = set_ust_event_exclusion(ua_event, app);
1441 if (ret < 0) {
1442 goto error;
1443 }
1444 }
1445
1446 /* If event not enabled, disable it on the tracer */
1447 if (ua_event->enabled == 0) {
1448 ret = disable_ust_event(app, ua_sess, ua_event);
1449 if (ret < 0) {
1450 /*
1451 * If we hit an EPERM, something is wrong with our disable call. If
1452 * we get an EEXIST, there is a problem on the tracer side since we
1453 * just created it.
1454 */
1455 switch (ret) {
1456 case -LTTNG_UST_ERR_PERM:
1457 /* Code flow problem */
1458 assert(0);
1459 case -LTTNG_UST_ERR_EXIST:
1460 /* It's OK for our use case. */
1461 ret = 0;
1462 break;
1463 default:
1464 break;
1465 }
1466 goto error;
1467 }
1468 }
1469
1470 error:
1471 health_code_update();
1472 return ret;
1473 }
1474
1475 /*
1476 * Copy data between an UST app event and a LTT event.
1477 */
1478 static void shadow_copy_event(struct ust_app_event *ua_event,
1479 struct ltt_ust_event *uevent)
1480 {
1481 size_t exclusion_alloc_size;
1482
1483 strncpy(ua_event->name, uevent->attr.name, sizeof(ua_event->name));
1484 ua_event->name[sizeof(ua_event->name) - 1] = '\0';
1485
1486 ua_event->enabled = uevent->enabled;
1487
1488 /* Copy event attributes */
1489 memcpy(&ua_event->attr, &uevent->attr, sizeof(ua_event->attr));
1490
1491 /* Copy filter bytecode */
1492 if (uevent->filter) {
1493 ua_event->filter = alloc_copy_ust_app_filter(uevent->filter);
1494 /* Filter might be NULL here in case of ENONEM. */
1495 }
1496
1497 /* Copy exclusion data */
1498 if (uevent->exclusion) {
1499 exclusion_alloc_size = sizeof(struct lttng_ust_event_exclusion) +
1500 LTTNG_UST_SYM_NAME_LEN * uevent->exclusion->count;
1501 ua_event->exclusion = zmalloc(exclusion_alloc_size);
1502 if (ua_event->exclusion == NULL) {
1503 PERROR("malloc");
1504 } else {
1505 memcpy(ua_event->exclusion, uevent->exclusion,
1506 exclusion_alloc_size);
1507 }
1508 }
1509 }
1510
1511 /*
1512 * Copy data between an UST app channel and a LTT channel.
1513 */
1514 static void shadow_copy_channel(struct ust_app_channel *ua_chan,
1515 struct ltt_ust_channel *uchan)
1516 {
1517 struct lttng_ht_iter iter;
1518 struct ltt_ust_event *uevent;
1519 struct ltt_ust_context *uctx;
1520 struct ust_app_event *ua_event;
1521 struct ust_app_ctx *ua_ctx;
1522
1523 DBG2("UST app shadow copy of channel %s started", ua_chan->name);
1524
1525 strncpy(ua_chan->name, uchan->name, sizeof(ua_chan->name));
1526 ua_chan->name[sizeof(ua_chan->name) - 1] = '\0';
1527
1528 ua_chan->tracefile_size = uchan->tracefile_size;
1529 ua_chan->tracefile_count = uchan->tracefile_count;
1530
1531 /* Copy event attributes since the layout is different. */
1532 ua_chan->attr.subbuf_size = uchan->attr.subbuf_size;
1533 ua_chan->attr.num_subbuf = uchan->attr.num_subbuf;
1534 ua_chan->attr.overwrite = uchan->attr.overwrite;
1535 ua_chan->attr.switch_timer_interval = uchan->attr.switch_timer_interval;
1536 ua_chan->attr.read_timer_interval = uchan->attr.read_timer_interval;
1537 ua_chan->attr.output = uchan->attr.output;
1538 /*
1539 * Note that the attribute channel type is not set since the channel on the
1540 * tracing registry side does not have this information.
1541 */
1542
1543 ua_chan->enabled = uchan->enabled;
1544 ua_chan->tracing_channel_id = uchan->id;
1545
1546 cds_list_for_each_entry(uctx, &uchan->ctx_list, list) {
1547 ua_ctx = alloc_ust_app_ctx(&uctx->ctx);
1548 if (ua_ctx == NULL) {
1549 continue;
1550 }
1551 lttng_ht_node_init_ulong(&ua_ctx->node,
1552 (unsigned long) ua_ctx->ctx.ctx);
1553 lttng_ht_add_unique_ulong(ua_chan->ctx, &ua_ctx->node);
1554 cds_list_add_tail(&ua_ctx->list, &ua_chan->ctx_list);
1555 }
1556
1557 /* Copy all events from ltt ust channel to ust app channel */
1558 cds_lfht_for_each_entry(uchan->events->ht, &iter.iter, uevent, node.node) {
1559 ua_event = find_ust_app_event(ua_chan->events, uevent->attr.name,
1560 uevent->filter, uevent->attr.loglevel, uevent->exclusion);
1561 if (ua_event == NULL) {
1562 DBG2("UST event %s not found on shadow copy channel",
1563 uevent->attr.name);
1564 ua_event = alloc_ust_app_event(uevent->attr.name, &uevent->attr);
1565 if (ua_event == NULL) {
1566 continue;
1567 }
1568 shadow_copy_event(ua_event, uevent);
1569 add_unique_ust_app_event(ua_chan, ua_event);
1570 }
1571 }
1572
1573 DBG3("UST app shadow copy of channel %s done", ua_chan->name);
1574 }
1575
1576 /*
1577 * Copy data between a UST app session and a regular LTT session.
1578 */
1579 static void shadow_copy_session(struct ust_app_session *ua_sess,
1580 struct ltt_ust_session *usess, struct ust_app *app)
1581 {
1582 struct lttng_ht_node_str *ua_chan_node;
1583 struct lttng_ht_iter iter;
1584 struct ltt_ust_channel *uchan;
1585 struct ust_app_channel *ua_chan;
1586 time_t rawtime;
1587 struct tm *timeinfo;
1588 char datetime[16];
1589 int ret;
1590
1591 /* Get date and time for unique app path */
1592 time(&rawtime);
1593 timeinfo = localtime(&rawtime);
1594 strftime(datetime, sizeof(datetime), "%Y%m%d-%H%M%S", timeinfo);
1595
1596 DBG2("Shadow copy of session handle %d", ua_sess->handle);
1597
1598 ua_sess->tracing_id = usess->id;
1599 ua_sess->id = get_next_session_id();
1600 ua_sess->uid = app->uid;
1601 ua_sess->gid = app->gid;
1602 ua_sess->euid = usess->uid;
1603 ua_sess->egid = usess->gid;
1604 ua_sess->buffer_type = usess->buffer_type;
1605 ua_sess->bits_per_long = app->bits_per_long;
1606 /* There is only one consumer object per session possible. */
1607 ua_sess->consumer = usess->consumer;
1608 ua_sess->output_traces = usess->output_traces;
1609 ua_sess->live_timer_interval = usess->live_timer_interval;
1610
1611 switch (ua_sess->buffer_type) {
1612 case LTTNG_BUFFER_PER_PID:
1613 ret = snprintf(ua_sess->path, sizeof(ua_sess->path),
1614 DEFAULT_UST_TRACE_PID_PATH "/%s-%d-%s", app->name, app->pid,
1615 datetime);
1616 break;
1617 case LTTNG_BUFFER_PER_UID:
1618 ret = snprintf(ua_sess->path, sizeof(ua_sess->path),
1619 DEFAULT_UST_TRACE_UID_PATH, ua_sess->uid, app->bits_per_long);
1620 break;
1621 default:
1622 assert(0);
1623 goto error;
1624 }
1625 if (ret < 0) {
1626 PERROR("asprintf UST shadow copy session");
1627 assert(0);
1628 goto error;
1629 }
1630
1631 /* Iterate over all channels in global domain. */
1632 cds_lfht_for_each_entry(usess->domain_global.channels->ht, &iter.iter,
1633 uchan, node.node) {
1634 struct lttng_ht_iter uiter;
1635
1636 lttng_ht_lookup(ua_sess->channels, (void *)uchan->name, &uiter);
1637 ua_chan_node = lttng_ht_iter_get_node_str(&uiter);
1638 if (ua_chan_node != NULL) {
1639 /* Session exist. Contiuing. */
1640 continue;
1641 }
1642
1643 DBG2("Channel %s not found on shadow session copy, creating it",
1644 uchan->name);
1645 ua_chan = alloc_ust_app_channel(uchan->name, ua_sess, &uchan->attr);
1646 if (ua_chan == NULL) {
1647 /* malloc failed FIXME: Might want to do handle ENOMEM .. */
1648 continue;
1649 }
1650 shadow_copy_channel(ua_chan, uchan);
1651 /*
1652 * The concept of metadata channel does not exist on the tracing
1653 * registry side of the session daemon so this can only be a per CPU
1654 * channel and not metadata.
1655 */
1656 ua_chan->attr.type = LTTNG_UST_CHAN_PER_CPU;
1657
1658 lttng_ht_add_unique_str(ua_sess->channels, &ua_chan->node);
1659 }
1660
1661 error:
1662 return;
1663 }
1664
1665 /*
1666 * Lookup sesison wrapper.
1667 */
1668 static
1669 void __lookup_session_by_app(struct ltt_ust_session *usess,
1670 struct ust_app *app, struct lttng_ht_iter *iter)
1671 {
1672 /* Get right UST app session from app */
1673 lttng_ht_lookup(app->sessions, &usess->id, iter);
1674 }
1675
1676 /*
1677 * Return ust app session from the app session hashtable using the UST session
1678 * id.
1679 */
1680 static struct ust_app_session *lookup_session_by_app(
1681 struct ltt_ust_session *usess, struct ust_app *app)
1682 {
1683 struct lttng_ht_iter iter;
1684 struct lttng_ht_node_u64 *node;
1685
1686 __lookup_session_by_app(usess, app, &iter);
1687 node = lttng_ht_iter_get_node_u64(&iter);
1688 if (node == NULL) {
1689 goto error;
1690 }
1691
1692 return caa_container_of(node, struct ust_app_session, node);
1693
1694 error:
1695 return NULL;
1696 }
1697
1698 /*
1699 * Setup buffer registry per PID for the given session and application. If none
1700 * is found, a new one is created, added to the global registry and
1701 * initialized. If regp is valid, it's set with the newly created object.
1702 *
1703 * Return 0 on success or else a negative value.
1704 */
1705 static int setup_buffer_reg_pid(struct ust_app_session *ua_sess,
1706 struct ust_app *app, struct buffer_reg_pid **regp)
1707 {
1708 int ret = 0;
1709 struct buffer_reg_pid *reg_pid;
1710
1711 assert(ua_sess);
1712 assert(app);
1713
1714 rcu_read_lock();
1715
1716 reg_pid = buffer_reg_pid_find(ua_sess->id);
1717 if (!reg_pid) {
1718 /*
1719 * This is the create channel path meaning that if there is NO
1720 * registry available, we have to create one for this session.
1721 */
1722 ret = buffer_reg_pid_create(ua_sess->id, &reg_pid);
1723 if (ret < 0) {
1724 goto error;
1725 }
1726 buffer_reg_pid_add(reg_pid);
1727 } else {
1728 goto end;
1729 }
1730
1731 /* Initialize registry. */
1732 ret = ust_registry_session_init(&reg_pid->registry->reg.ust, app,
1733 app->bits_per_long, app->uint8_t_alignment,
1734 app->uint16_t_alignment, app->uint32_t_alignment,
1735 app->uint64_t_alignment, app->long_alignment,
1736 app->byte_order, app->version.major,
1737 app->version.minor);
1738 if (ret < 0) {
1739 goto error;
1740 }
1741
1742 DBG3("UST app buffer registry per PID created successfully");
1743
1744 end:
1745 if (regp) {
1746 *regp = reg_pid;
1747 }
1748 error:
1749 rcu_read_unlock();
1750 return ret;
1751 }
1752
1753 /*
1754 * Setup buffer registry per UID for the given session and application. If none
1755 * is found, a new one is created, added to the global registry and
1756 * initialized. If regp is valid, it's set with the newly created object.
1757 *
1758 * Return 0 on success or else a negative value.
1759 */
1760 static int setup_buffer_reg_uid(struct ltt_ust_session *usess,
1761 struct ust_app *app, struct buffer_reg_uid **regp)
1762 {
1763 int ret = 0;
1764 struct buffer_reg_uid *reg_uid;
1765
1766 assert(usess);
1767 assert(app);
1768
1769 rcu_read_lock();
1770
1771 reg_uid = buffer_reg_uid_find(usess->id, app->bits_per_long, app->uid);
1772 if (!reg_uid) {
1773 /*
1774 * This is the create channel path meaning that if there is NO
1775 * registry available, we have to create one for this session.
1776 */
1777 ret = buffer_reg_uid_create(usess->id, app->bits_per_long, app->uid,
1778 LTTNG_DOMAIN_UST, &reg_uid);
1779 if (ret < 0) {
1780 goto error;
1781 }
1782 buffer_reg_uid_add(reg_uid);
1783 } else {
1784 goto end;
1785 }
1786
1787 /* Initialize registry. */
1788 ret = ust_registry_session_init(&reg_uid->registry->reg.ust, NULL,
1789 app->bits_per_long, app->uint8_t_alignment,
1790 app->uint16_t_alignment, app->uint32_t_alignment,
1791 app->uint64_t_alignment, app->long_alignment,
1792 app->byte_order, app->version.major,
1793 app->version.minor);
1794 if (ret < 0) {
1795 goto error;
1796 }
1797 /* Add node to teardown list of the session. */
1798 cds_list_add(&reg_uid->lnode, &usess->buffer_reg_uid_list);
1799
1800 DBG3("UST app buffer registry per UID created successfully");
1801
1802 end:
1803 if (regp) {
1804 *regp = reg_uid;
1805 }
1806 error:
1807 rcu_read_unlock();
1808 return ret;
1809 }
1810
1811 /*
1812 * Create a session on the tracer side for the given app.
1813 *
1814 * On success, ua_sess_ptr is populated with the session pointer or else left
1815 * untouched. If the session was created, is_created is set to 1. On error,
1816 * it's left untouched. Note that ua_sess_ptr is mandatory but is_created can
1817 * be NULL.
1818 *
1819 * Returns 0 on success or else a negative code which is either -ENOMEM or
1820 * -ENOTCONN which is the default code if the ustctl_create_session fails.
1821 */
1822 static int create_ust_app_session(struct ltt_ust_session *usess,
1823 struct ust_app *app, struct ust_app_session **ua_sess_ptr,
1824 int *is_created)
1825 {
1826 int ret, created = 0;
1827 struct ust_app_session *ua_sess;
1828
1829 assert(usess);
1830 assert(app);
1831 assert(ua_sess_ptr);
1832
1833 health_code_update();
1834
1835 ua_sess = lookup_session_by_app(usess, app);
1836 if (ua_sess == NULL) {
1837 DBG2("UST app pid: %d session id %" PRIu64 " not found, creating it",
1838 app->pid, usess->id);
1839 ua_sess = alloc_ust_app_session(app);
1840 if (ua_sess == NULL) {
1841 /* Only malloc can failed so something is really wrong */
1842 ret = -ENOMEM;
1843 goto error;
1844 }
1845 shadow_copy_session(ua_sess, usess, app);
1846 created = 1;
1847 }
1848
1849 switch (usess->buffer_type) {
1850 case LTTNG_BUFFER_PER_PID:
1851 /* Init local registry. */
1852 ret = setup_buffer_reg_pid(ua_sess, app, NULL);
1853 if (ret < 0) {
1854 goto error;
1855 }
1856 break;
1857 case LTTNG_BUFFER_PER_UID:
1858 /* Look for a global registry. If none exists, create one. */
1859 ret = setup_buffer_reg_uid(usess, app, NULL);
1860 if (ret < 0) {
1861 goto error;
1862 }
1863 break;
1864 default:
1865 assert(0);
1866 ret = -EINVAL;
1867 goto error;
1868 }
1869
1870 health_code_update();
1871
1872 if (ua_sess->handle == -1) {
1873 ret = ustctl_create_session(app->sock);
1874 if (ret < 0) {
1875 if (ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
1876 ERR("Creating session for app pid %d with ret %d",
1877 app->pid, ret);
1878 } else {
1879 DBG("UST app creating session failed. Application is dead");
1880 /*
1881 * This is normal behavior, an application can die during the
1882 * creation process. Don't report an error so the execution can
1883 * continue normally. This will get flagged ENOTCONN and the
1884 * caller will handle it.
1885 */
1886 ret = 0;
1887 }
1888 delete_ust_app_session(-1, ua_sess, app);
1889 if (ret != -ENOMEM) {
1890 /*
1891 * Tracer is probably gone or got an internal error so let's
1892 * behave like it will soon unregister or not usable.
1893 */
1894 ret = -ENOTCONN;
1895 }
1896 goto error;
1897 }
1898
1899 ua_sess->handle = ret;
1900
1901 /* Add ust app session to app's HT */
1902 lttng_ht_node_init_u64(&ua_sess->node,
1903 ua_sess->tracing_id);
1904 lttng_ht_add_unique_u64(app->sessions, &ua_sess->node);
1905
1906 DBG2("UST app session created successfully with handle %d", ret);
1907 }
1908
1909 *ua_sess_ptr = ua_sess;
1910 if (is_created) {
1911 *is_created = created;
1912 }
1913
1914 /* Everything went well. */
1915 ret = 0;
1916
1917 error:
1918 health_code_update();
1919 return ret;
1920 }
1921
1922 /*
1923 * Create a context for the channel on the tracer.
1924 *
1925 * Called with UST app session lock held and a RCU read side lock.
1926 */
1927 static
1928 int create_ust_app_channel_context(struct ust_app_session *ua_sess,
1929 struct ust_app_channel *ua_chan, struct lttng_ust_context *uctx,
1930 struct ust_app *app)
1931 {
1932 int ret = 0;
1933 struct lttng_ht_iter iter;
1934 struct lttng_ht_node_ulong *node;
1935 struct ust_app_ctx *ua_ctx;
1936
1937 DBG2("UST app adding context to channel %s", ua_chan->name);
1938
1939 lttng_ht_lookup(ua_chan->ctx, (void *)((unsigned long)uctx->ctx), &iter);
1940 node = lttng_ht_iter_get_node_ulong(&iter);
1941 if (node != NULL) {
1942 ret = -EEXIST;
1943 goto error;
1944 }
1945
1946 ua_ctx = alloc_ust_app_ctx(uctx);
1947 if (ua_ctx == NULL) {
1948 /* malloc failed */
1949 ret = -1;
1950 goto error;
1951 }
1952
1953 lttng_ht_node_init_ulong(&ua_ctx->node, (unsigned long) ua_ctx->ctx.ctx);
1954 lttng_ht_add_unique_ulong(ua_chan->ctx, &ua_ctx->node);
1955 cds_list_add_tail(&ua_ctx->list, &ua_chan->ctx_list);
1956
1957 ret = create_ust_channel_context(ua_chan, ua_ctx, app);
1958 if (ret < 0) {
1959 goto error;
1960 }
1961
1962 error:
1963 return ret;
1964 }
1965
1966 /*
1967 * Enable on the tracer side a ust app event for the session and channel.
1968 *
1969 * Called with UST app session lock held.
1970 */
1971 static
1972 int enable_ust_app_event(struct ust_app_session *ua_sess,
1973 struct ust_app_event *ua_event, struct ust_app *app)
1974 {
1975 int ret;
1976
1977 ret = enable_ust_event(app, ua_sess, ua_event);
1978 if (ret < 0) {
1979 goto error;
1980 }
1981
1982 ua_event->enabled = 1;
1983
1984 error:
1985 return ret;
1986 }
1987
1988 /*
1989 * Disable on the tracer side a ust app event for the session and channel.
1990 */
1991 static int disable_ust_app_event(struct ust_app_session *ua_sess,
1992 struct ust_app_event *ua_event, struct ust_app *app)
1993 {
1994 int ret;
1995
1996 ret = disable_ust_event(app, ua_sess, ua_event);
1997 if (ret < 0) {
1998 goto error;
1999 }
2000
2001 ua_event->enabled = 0;
2002
2003 error:
2004 return ret;
2005 }
2006
2007 /*
2008 * Lookup ust app channel for session and disable it on the tracer side.
2009 */
2010 static
2011 int disable_ust_app_channel(struct ust_app_session *ua_sess,
2012 struct ust_app_channel *ua_chan, struct ust_app *app)
2013 {
2014 int ret;
2015
2016 ret = disable_ust_channel(app, ua_sess, ua_chan);
2017 if (ret < 0) {
2018 goto error;
2019 }
2020
2021 ua_chan->enabled = 0;
2022
2023 error:
2024 return ret;
2025 }
2026
2027 /*
2028 * Lookup ust app channel for session and enable it on the tracer side. This
2029 * MUST be called with a RCU read side lock acquired.
2030 */
2031 static int enable_ust_app_channel(struct ust_app_session *ua_sess,
2032 struct ltt_ust_channel *uchan, struct ust_app *app)
2033 {
2034 int ret = 0;
2035 struct lttng_ht_iter iter;
2036 struct lttng_ht_node_str *ua_chan_node;
2037 struct ust_app_channel *ua_chan;
2038
2039 lttng_ht_lookup(ua_sess->channels, (void *)uchan->name, &iter);
2040 ua_chan_node = lttng_ht_iter_get_node_str(&iter);
2041 if (ua_chan_node == NULL) {
2042 DBG2("Unable to find channel %s in ust session id %" PRIu64,
2043 uchan->name, ua_sess->tracing_id);
2044 goto error;
2045 }
2046
2047 ua_chan = caa_container_of(ua_chan_node, struct ust_app_channel, node);
2048
2049 ret = enable_ust_channel(app, ua_sess, ua_chan);
2050 if (ret < 0) {
2051 goto error;
2052 }
2053
2054 error:
2055 return ret;
2056 }
2057
2058 /*
2059 * Ask the consumer to create a channel and get it if successful.
2060 *
2061 * Return 0 on success or else a negative value.
2062 */
2063 static int do_consumer_create_channel(struct ltt_ust_session *usess,
2064 struct ust_app_session *ua_sess, struct ust_app_channel *ua_chan,
2065 int bitness, struct ust_registry_session *registry)
2066 {
2067 int ret;
2068 unsigned int nb_fd = 0;
2069 struct consumer_socket *socket;
2070
2071 assert(usess);
2072 assert(ua_sess);
2073 assert(ua_chan);
2074 assert(registry);
2075
2076 rcu_read_lock();
2077 health_code_update();
2078
2079 /* Get the right consumer socket for the application. */
2080 socket = consumer_find_socket_by_bitness(bitness, usess->consumer);
2081 if (!socket) {
2082 ret = -EINVAL;
2083 goto error;
2084 }
2085
2086 health_code_update();
2087
2088 /* Need one fd for the channel. */
2089 ret = lttng_fd_get(LTTNG_FD_APPS, 1);
2090 if (ret < 0) {
2091 ERR("Exhausted number of available FD upon create channel");
2092 goto error;
2093 }
2094
2095 /*
2096 * Ask consumer to create channel. The consumer will return the number of
2097 * stream we have to expect.
2098 */
2099 ret = ust_consumer_ask_channel(ua_sess, ua_chan, usess->consumer, socket,
2100 registry);
2101 if (ret < 0) {
2102 goto error_ask;
2103 }
2104
2105 /*
2106 * Compute the number of fd needed before receiving them. It must be 2 per
2107 * stream (2 being the default value here).
2108 */
2109 nb_fd = DEFAULT_UST_STREAM_FD_NUM * ua_chan->expected_stream_count;
2110
2111 /* Reserve the amount of file descriptor we need. */
2112 ret = lttng_fd_get(LTTNG_FD_APPS, nb_fd);
2113 if (ret < 0) {
2114 ERR("Exhausted number of available FD upon create channel");
2115 goto error_fd_get_stream;
2116 }
2117
2118 health_code_update();
2119
2120 /*
2121 * Now get the channel from the consumer. This call wil populate the stream
2122 * list of that channel and set the ust objects.
2123 */
2124 if (usess->consumer->enabled) {
2125 ret = ust_consumer_get_channel(socket, ua_chan);
2126 if (ret < 0) {
2127 goto error_destroy;
2128 }
2129 }
2130
2131 rcu_read_unlock();
2132 return 0;
2133
2134 error_destroy:
2135 lttng_fd_put(LTTNG_FD_APPS, nb_fd);
2136 error_fd_get_stream:
2137 /*
2138 * Initiate a destroy channel on the consumer since we had an error
2139 * handling it on our side. The return value is of no importance since we
2140 * already have a ret value set by the previous error that we need to
2141 * return.
2142 */
2143 (void) ust_consumer_destroy_channel(socket, ua_chan);
2144 error_ask:
2145 lttng_fd_put(LTTNG_FD_APPS, 1);
2146 error:
2147 health_code_update();
2148 rcu_read_unlock();
2149 return ret;
2150 }
2151
2152 /*
2153 * Duplicate the ust data object of the ust app stream and save it in the
2154 * buffer registry stream.
2155 *
2156 * Return 0 on success or else a negative value.
2157 */
2158 static int duplicate_stream_object(struct buffer_reg_stream *reg_stream,
2159 struct ust_app_stream *stream)
2160 {
2161 int ret;
2162
2163 assert(reg_stream);
2164 assert(stream);
2165
2166 /* Reserve the amount of file descriptor we need. */
2167 ret = lttng_fd_get(LTTNG_FD_APPS, 2);
2168 if (ret < 0) {
2169 ERR("Exhausted number of available FD upon duplicate stream");
2170 goto error;
2171 }
2172
2173 /* Duplicate object for stream once the original is in the registry. */
2174 ret = ustctl_duplicate_ust_object_data(&stream->obj,
2175 reg_stream->obj.ust);
2176 if (ret < 0) {
2177 ERR("Duplicate stream obj from %p to %p failed with ret %d",
2178 reg_stream->obj.ust, stream->obj, ret);
2179 lttng_fd_put(LTTNG_FD_APPS, 2);
2180 goto error;
2181 }
2182 stream->handle = stream->obj->handle;
2183
2184 error:
2185 return ret;
2186 }
2187
2188 /*
2189 * Duplicate the ust data object of the ust app. channel and save it in the
2190 * buffer registry channel.
2191 *
2192 * Return 0 on success or else a negative value.
2193 */
2194 static int duplicate_channel_object(struct buffer_reg_channel *reg_chan,
2195 struct ust_app_channel *ua_chan)
2196 {
2197 int ret;
2198
2199 assert(reg_chan);
2200 assert(ua_chan);
2201
2202 /* Need two fds for the channel. */
2203 ret = lttng_fd_get(LTTNG_FD_APPS, 1);
2204 if (ret < 0) {
2205 ERR("Exhausted number of available FD upon duplicate channel");
2206 goto error_fd_get;
2207 }
2208
2209 /* Duplicate object for stream once the original is in the registry. */
2210 ret = ustctl_duplicate_ust_object_data(&ua_chan->obj, reg_chan->obj.ust);
2211 if (ret < 0) {
2212 ERR("Duplicate channel obj from %p to %p failed with ret: %d",
2213 reg_chan->obj.ust, ua_chan->obj, ret);
2214 goto error;
2215 }
2216 ua_chan->handle = ua_chan->obj->handle;
2217
2218 return 0;
2219
2220 error:
2221 lttng_fd_put(LTTNG_FD_APPS, 1);
2222 error_fd_get:
2223 return ret;
2224 }
2225
2226 /*
2227 * For a given channel buffer registry, setup all streams of the given ust
2228 * application channel.
2229 *
2230 * Return 0 on success or else a negative value.
2231 */
2232 static int setup_buffer_reg_streams(struct buffer_reg_channel *reg_chan,
2233 struct ust_app_channel *ua_chan)
2234 {
2235 int ret = 0;
2236 struct ust_app_stream *stream, *stmp;
2237
2238 assert(reg_chan);
2239 assert(ua_chan);
2240
2241 DBG2("UST app setup buffer registry stream");
2242
2243 /* Send all streams to application. */
2244 cds_list_for_each_entry_safe(stream, stmp, &ua_chan->streams.head, list) {
2245 struct buffer_reg_stream *reg_stream;
2246
2247 ret = buffer_reg_stream_create(&reg_stream);
2248 if (ret < 0) {
2249 goto error;
2250 }
2251
2252 /*
2253 * Keep original pointer and nullify it in the stream so the delete
2254 * stream call does not release the object.
2255 */
2256 reg_stream->obj.ust = stream->obj;
2257 stream->obj = NULL;
2258 buffer_reg_stream_add(reg_stream, reg_chan);
2259
2260 /* We don't need the streams anymore. */
2261 cds_list_del(&stream->list);
2262 delete_ust_app_stream(-1, stream);
2263 }
2264
2265 error:
2266 return ret;
2267 }
2268
2269 /*
2270 * Create a buffer registry channel for the given session registry and
2271 * application channel object. If regp pointer is valid, it's set with the
2272 * created object. Important, the created object is NOT added to the session
2273 * registry hash table.
2274 *
2275 * Return 0 on success else a negative value.
2276 */
2277 static int create_buffer_reg_channel(struct buffer_reg_session *reg_sess,
2278 struct ust_app_channel *ua_chan, struct buffer_reg_channel **regp)
2279 {
2280 int ret;
2281 struct buffer_reg_channel *reg_chan = NULL;
2282
2283 assert(reg_sess);
2284 assert(ua_chan);
2285
2286 DBG2("UST app creating buffer registry channel for %s", ua_chan->name);
2287
2288 /* Create buffer registry channel. */
2289 ret = buffer_reg_channel_create(ua_chan->tracing_channel_id, &reg_chan);
2290 if (ret < 0) {
2291 goto error_create;
2292 }
2293 assert(reg_chan);
2294 reg_chan->consumer_key = ua_chan->key;
2295 reg_chan->subbuf_size = ua_chan->attr.subbuf_size;
2296
2297 /* Create and add a channel registry to session. */
2298 ret = ust_registry_channel_add(reg_sess->reg.ust,
2299 ua_chan->tracing_channel_id);
2300 if (ret < 0) {
2301 goto error;
2302 }
2303 buffer_reg_channel_add(reg_sess, reg_chan);
2304
2305 if (regp) {
2306 *regp = reg_chan;
2307 }
2308
2309 return 0;
2310
2311 error:
2312 /* Safe because the registry channel object was not added to any HT. */
2313 buffer_reg_channel_destroy(reg_chan, LTTNG_DOMAIN_UST);
2314 error_create:
2315 return ret;
2316 }
2317
2318 /*
2319 * Setup buffer registry channel for the given session registry and application
2320 * channel object. If regp pointer is valid, it's set with the created object.
2321 *
2322 * Return 0 on success else a negative value.
2323 */
2324 static int setup_buffer_reg_channel(struct buffer_reg_session *reg_sess,
2325 struct ust_app_channel *ua_chan, struct buffer_reg_channel *reg_chan)
2326 {
2327 int ret;
2328
2329 assert(reg_sess);
2330 assert(reg_chan);
2331 assert(ua_chan);
2332 assert(ua_chan->obj);
2333
2334 DBG2("UST app setup buffer registry channel for %s", ua_chan->name);
2335
2336 /* Setup all streams for the registry. */
2337 ret = setup_buffer_reg_streams(reg_chan, ua_chan);
2338 if (ret < 0) {
2339 goto error;
2340 }
2341
2342 reg_chan->obj.ust = ua_chan->obj;
2343 ua_chan->obj = NULL;
2344
2345 return 0;
2346
2347 error:
2348 buffer_reg_channel_remove(reg_sess, reg_chan);
2349 buffer_reg_channel_destroy(reg_chan, LTTNG_DOMAIN_UST);
2350 return ret;
2351 }
2352
2353 /*
2354 * Send buffer registry channel to the application.
2355 *
2356 * Return 0 on success else a negative value.
2357 */
2358 static int send_channel_uid_to_ust(struct buffer_reg_channel *reg_chan,
2359 struct ust_app *app, struct ust_app_session *ua_sess,
2360 struct ust_app_channel *ua_chan)
2361 {
2362 int ret;
2363 struct buffer_reg_stream *reg_stream;
2364
2365 assert(reg_chan);
2366 assert(app);
2367 assert(ua_sess);
2368 assert(ua_chan);
2369
2370 DBG("UST app sending buffer registry channel to ust sock %d", app->sock);
2371
2372 ret = duplicate_channel_object(reg_chan, ua_chan);
2373 if (ret < 0) {
2374 goto error;
2375 }
2376
2377 /* Send channel to the application. */
2378 ret = ust_consumer_send_channel_to_ust(app, ua_sess, ua_chan);
2379 if (ret < 0) {
2380 goto error;
2381 }
2382
2383 health_code_update();
2384
2385 /* Send all streams to application. */
2386 pthread_mutex_lock(&reg_chan->stream_list_lock);
2387 cds_list_for_each_entry(reg_stream, &reg_chan->streams, lnode) {
2388 struct ust_app_stream stream;
2389
2390 ret = duplicate_stream_object(reg_stream, &stream);
2391 if (ret < 0) {
2392 goto error_stream_unlock;
2393 }
2394
2395 ret = ust_consumer_send_stream_to_ust(app, ua_chan, &stream);
2396 if (ret < 0) {
2397 (void) release_ust_app_stream(-1, &stream);
2398 goto error_stream_unlock;
2399 }
2400
2401 /*
2402 * The return value is not important here. This function will output an
2403 * error if needed.
2404 */
2405 (void) release_ust_app_stream(-1, &stream);
2406 }
2407 ua_chan->is_sent = 1;
2408
2409 error_stream_unlock:
2410 pthread_mutex_unlock(&reg_chan->stream_list_lock);
2411 error:
2412 return ret;
2413 }
2414
2415 /*
2416 * Create and send to the application the created buffers with per UID buffers.
2417 *
2418 * Return 0 on success else a negative value.
2419 */
2420 static int create_channel_per_uid(struct ust_app *app,
2421 struct ltt_ust_session *usess, struct ust_app_session *ua_sess,
2422 struct ust_app_channel *ua_chan)
2423 {
2424 int ret;
2425 struct buffer_reg_uid *reg_uid;
2426 struct buffer_reg_channel *reg_chan;
2427
2428 assert(app);
2429 assert(usess);
2430 assert(ua_sess);
2431 assert(ua_chan);
2432
2433 DBG("UST app creating channel %s with per UID buffers", ua_chan->name);
2434
2435 reg_uid = buffer_reg_uid_find(usess->id, app->bits_per_long, app->uid);
2436 /*
2437 * The session creation handles the creation of this global registry
2438 * object. If none can be find, there is a code flow problem or a
2439 * teardown race.
2440 */
2441 assert(reg_uid);
2442
2443 reg_chan = buffer_reg_channel_find(ua_chan->tracing_channel_id,
2444 reg_uid);
2445 if (!reg_chan) {
2446 /* Create the buffer registry channel object. */
2447 ret = create_buffer_reg_channel(reg_uid->registry, ua_chan, &reg_chan);
2448 if (ret < 0) {
2449 goto error;
2450 }
2451 assert(reg_chan);
2452
2453 /*
2454 * Create the buffers on the consumer side. This call populates the
2455 * ust app channel object with all streams and data object.
2456 */
2457 ret = do_consumer_create_channel(usess, ua_sess, ua_chan,
2458 app->bits_per_long, reg_uid->registry->reg.ust);
2459 if (ret < 0) {
2460 /*
2461 * Let's remove the previously created buffer registry channel so
2462 * it's not visible anymore in the session registry.
2463 */
2464 ust_registry_channel_del_free(reg_uid->registry->reg.ust,
2465 ua_chan->tracing_channel_id);
2466 buffer_reg_channel_remove(reg_uid->registry, reg_chan);
2467 buffer_reg_channel_destroy(reg_chan, LTTNG_DOMAIN_UST);
2468 goto error;
2469 }
2470
2471 /*
2472 * Setup the streams and add it to the session registry.
2473 */
2474 ret = setup_buffer_reg_channel(reg_uid->registry, ua_chan, reg_chan);
2475 if (ret < 0) {
2476 goto error;
2477 }
2478
2479 }
2480
2481 /* Send buffers to the application. */
2482 ret = send_channel_uid_to_ust(reg_chan, app, ua_sess, ua_chan);
2483 if (ret < 0) {
2484 goto error;
2485 }
2486
2487 error:
2488 return ret;
2489 }
2490
2491 /*
2492 * Create and send to the application the created buffers with per PID buffers.
2493 *
2494 * Return 0 on success else a negative value.
2495 */
2496 static int create_channel_per_pid(struct ust_app *app,
2497 struct ltt_ust_session *usess, struct ust_app_session *ua_sess,
2498 struct ust_app_channel *ua_chan)
2499 {
2500 int ret;
2501 struct ust_registry_session *registry;
2502
2503 assert(app);
2504 assert(usess);
2505 assert(ua_sess);
2506 assert(ua_chan);
2507
2508 DBG("UST app creating channel %s with per PID buffers", ua_chan->name);
2509
2510 rcu_read_lock();
2511
2512 registry = get_session_registry(ua_sess);
2513 assert(registry);
2514
2515 /* Create and add a new channel registry to session. */
2516 ret = ust_registry_channel_add(registry, ua_chan->key);
2517 if (ret < 0) {
2518 goto error;
2519 }
2520
2521 /* Create and get channel on the consumer side. */
2522 ret = do_consumer_create_channel(usess, ua_sess, ua_chan,
2523 app->bits_per_long, registry);
2524 if (ret < 0) {
2525 goto error;
2526 }
2527
2528 ret = send_channel_pid_to_ust(app, ua_sess, ua_chan);
2529 if (ret < 0) {
2530 goto error;
2531 }
2532
2533 error:
2534 rcu_read_unlock();
2535 return ret;
2536 }
2537
2538 /*
2539 * From an already allocated ust app channel, create the channel buffers if
2540 * need and send it to the application. This MUST be called with a RCU read
2541 * side lock acquired.
2542 *
2543 * Return 0 on success or else a negative value.
2544 */
2545 static int do_create_channel(struct ust_app *app,
2546 struct ltt_ust_session *usess, struct ust_app_session *ua_sess,
2547 struct ust_app_channel *ua_chan)
2548 {
2549 int ret;
2550
2551 assert(app);
2552 assert(usess);
2553 assert(ua_sess);
2554 assert(ua_chan);
2555
2556 /* Handle buffer type before sending the channel to the application. */
2557 switch (usess->buffer_type) {
2558 case LTTNG_BUFFER_PER_UID:
2559 {
2560 ret = create_channel_per_uid(app, usess, ua_sess, ua_chan);
2561 if (ret < 0) {
2562 goto error;
2563 }
2564 break;
2565 }
2566 case LTTNG_BUFFER_PER_PID:
2567 {
2568 ret = create_channel_per_pid(app, usess, ua_sess, ua_chan);
2569 if (ret < 0) {
2570 goto error;
2571 }
2572 break;
2573 }
2574 default:
2575 assert(0);
2576 ret = -EINVAL;
2577 goto error;
2578 }
2579
2580 /* Initialize ust objd object using the received handle and add it. */
2581 lttng_ht_node_init_ulong(&ua_chan->ust_objd_node, ua_chan->handle);
2582 lttng_ht_add_unique_ulong(app->ust_objd, &ua_chan->ust_objd_node);
2583
2584 /* If channel is not enabled, disable it on the tracer */
2585 if (!ua_chan->enabled) {
2586 ret = disable_ust_channel(app, ua_sess, ua_chan);
2587 if (ret < 0) {
2588 goto error;
2589 }
2590 }
2591
2592 error:
2593 return ret;
2594 }
2595
2596 /*
2597 * Create UST app channel and create it on the tracer. Set ua_chanp of the
2598 * newly created channel if not NULL.
2599 *
2600 * Called with UST app session lock and RCU read-side lock held.
2601 *
2602 * Return 0 on success or else a negative value.
2603 */
2604 static int create_ust_app_channel(struct ust_app_session *ua_sess,
2605 struct ltt_ust_channel *uchan, struct ust_app *app,
2606 enum lttng_ust_chan_type type, struct ltt_ust_session *usess,
2607 struct ust_app_channel **ua_chanp)
2608 {
2609 int ret = 0;
2610 struct lttng_ht_iter iter;
2611 struct lttng_ht_node_str *ua_chan_node;
2612 struct ust_app_channel *ua_chan;
2613
2614 /* Lookup channel in the ust app session */
2615 lttng_ht_lookup(ua_sess->channels, (void *)uchan->name, &iter);
2616 ua_chan_node = lttng_ht_iter_get_node_str(&iter);
2617 if (ua_chan_node != NULL) {
2618 ua_chan = caa_container_of(ua_chan_node, struct ust_app_channel, node);
2619 goto end;
2620 }
2621
2622 ua_chan = alloc_ust_app_channel(uchan->name, ua_sess, &uchan->attr);
2623 if (ua_chan == NULL) {
2624 /* Only malloc can fail here */
2625 ret = -ENOMEM;
2626 goto error_alloc;
2627 }
2628 shadow_copy_channel(ua_chan, uchan);
2629
2630 /* Set channel type. */
2631 ua_chan->attr.type = type;
2632
2633 ret = do_create_channel(app, usess, ua_sess, ua_chan);
2634 if (ret < 0) {
2635 goto error;
2636 }
2637
2638 DBG2("UST app create channel %s for PID %d completed", ua_chan->name,
2639 app->pid);
2640
2641 /* Only add the channel if successful on the tracer side. */
2642 lttng_ht_add_unique_str(ua_sess->channels, &ua_chan->node);
2643
2644 end:
2645 if (ua_chanp) {
2646 *ua_chanp = ua_chan;
2647 }
2648
2649 /* Everything went well. */
2650 return 0;
2651
2652 error:
2653 delete_ust_app_channel(ua_chan->is_sent ? app->sock : -1, ua_chan, app);
2654 error_alloc:
2655 return ret;
2656 }
2657
2658 /*
2659 * Create UST app event and create it on the tracer side.
2660 *
2661 * Called with ust app session mutex held.
2662 */
2663 static
2664 int create_ust_app_event(struct ust_app_session *ua_sess,
2665 struct ust_app_channel *ua_chan, struct ltt_ust_event *uevent,
2666 struct ust_app *app)
2667 {
2668 int ret = 0;
2669 struct ust_app_event *ua_event;
2670
2671 /* Get event node */
2672 ua_event = find_ust_app_event(ua_chan->events, uevent->attr.name,
2673 uevent->filter, uevent->attr.loglevel, uevent->exclusion);
2674 if (ua_event != NULL) {
2675 ret = -EEXIST;
2676 goto end;
2677 }
2678
2679 /* Does not exist so create one */
2680 ua_event = alloc_ust_app_event(uevent->attr.name, &uevent->attr);
2681 if (ua_event == NULL) {
2682 /* Only malloc can failed so something is really wrong */
2683 ret = -ENOMEM;
2684 goto end;
2685 }
2686 shadow_copy_event(ua_event, uevent);
2687
2688 /* Create it on the tracer side */
2689 ret = create_ust_event(app, ua_sess, ua_chan, ua_event);
2690 if (ret < 0) {
2691 /* Not found previously means that it does not exist on the tracer */
2692 assert(ret != -LTTNG_UST_ERR_EXIST);
2693 goto error;
2694 }
2695
2696 add_unique_ust_app_event(ua_chan, ua_event);
2697
2698 DBG2("UST app create event %s for PID %d completed", ua_event->name,
2699 app->pid);
2700
2701 end:
2702 return ret;
2703
2704 error:
2705 /* Valid. Calling here is already in a read side lock */
2706 delete_ust_app_event(-1, ua_event);
2707 return ret;
2708 }
2709
2710 /*
2711 * Create UST metadata and open it on the tracer side.
2712 *
2713 * Called with UST app session lock held and RCU read side lock.
2714 */
2715 static int create_ust_app_metadata(struct ust_app_session *ua_sess,
2716 struct ust_app *app, struct consumer_output *consumer,
2717 struct ustctl_consumer_channel_attr *attr)
2718 {
2719 int ret = 0;
2720 struct ust_app_channel *metadata;
2721 struct consumer_socket *socket;
2722 struct ust_registry_session *registry;
2723
2724 assert(ua_sess);
2725 assert(app);
2726 assert(consumer);
2727
2728 registry = get_session_registry(ua_sess);
2729 assert(registry);
2730
2731 /* Metadata already exists for this registry or it was closed previously */
2732 if (registry->metadata_key || registry->metadata_closed) {
2733 ret = 0;
2734 goto error;
2735 }
2736
2737 /* Allocate UST metadata */
2738 metadata = alloc_ust_app_channel(DEFAULT_METADATA_NAME, ua_sess, NULL);
2739 if (!metadata) {
2740 /* malloc() failed */
2741 ret = -ENOMEM;
2742 goto error;
2743 }
2744
2745 if (!attr) {
2746 /* Set default attributes for metadata. */
2747 metadata->attr.overwrite = DEFAULT_CHANNEL_OVERWRITE;
2748 metadata->attr.subbuf_size = default_get_metadata_subbuf_size();
2749 metadata->attr.num_subbuf = DEFAULT_METADATA_SUBBUF_NUM;
2750 metadata->attr.switch_timer_interval = DEFAULT_METADATA_SWITCH_TIMER;
2751 metadata->attr.read_timer_interval = DEFAULT_METADATA_READ_TIMER;
2752 metadata->attr.output = LTTNG_UST_MMAP;
2753 metadata->attr.type = LTTNG_UST_CHAN_METADATA;
2754 } else {
2755 memcpy(&metadata->attr, attr, sizeof(metadata->attr));
2756 metadata->attr.output = LTTNG_UST_MMAP;
2757 metadata->attr.type = LTTNG_UST_CHAN_METADATA;
2758 }
2759
2760 /* Need one fd for the channel. */
2761 ret = lttng_fd_get(LTTNG_FD_APPS, 1);
2762 if (ret < 0) {
2763 ERR("Exhausted number of available FD upon create metadata");
2764 goto error;
2765 }
2766
2767 /* Get the right consumer socket for the application. */
2768 socket = consumer_find_socket_by_bitness(app->bits_per_long, consumer);
2769 if (!socket) {
2770 ret = -EINVAL;
2771 goto error_consumer;
2772 }
2773
2774 /*
2775 * Keep metadata key so we can identify it on the consumer side. Assign it
2776 * to the registry *before* we ask the consumer so we avoid the race of the
2777 * consumer requesting the metadata and the ask_channel call on our side
2778 * did not returned yet.
2779 */
2780 registry->metadata_key = metadata->key;
2781
2782 /*
2783 * Ask the metadata channel creation to the consumer. The metadata object
2784 * will be created by the consumer and kept their. However, the stream is
2785 * never added or monitored until we do a first push metadata to the
2786 * consumer.
2787 */
2788 ret = ust_consumer_ask_channel(ua_sess, metadata, consumer, socket,
2789 registry);
2790 if (ret < 0) {
2791 /* Nullify the metadata key so we don't try to close it later on. */
2792 registry->metadata_key = 0;
2793 goto error_consumer;
2794 }
2795
2796 /*
2797 * The setup command will make the metadata stream be sent to the relayd,
2798 * if applicable, and the thread managing the metadatas. This is important
2799 * because after this point, if an error occurs, the only way the stream
2800 * can be deleted is to be monitored in the consumer.
2801 */
2802 ret = consumer_setup_metadata(socket, metadata->key);
2803 if (ret < 0) {
2804 /* Nullify the metadata key so we don't try to close it later on. */
2805 registry->metadata_key = 0;
2806 goto error_consumer;
2807 }
2808
2809 DBG2("UST metadata with key %" PRIu64 " created for app pid %d",
2810 metadata->key, app->pid);
2811
2812 error_consumer:
2813 lttng_fd_put(LTTNG_FD_APPS, 1);
2814 delete_ust_app_channel(-1, metadata, app);
2815 error:
2816 return ret;
2817 }
2818
2819 /*
2820 * Return pointer to traceable apps list.
2821 */
2822 struct lttng_ht *ust_app_get_ht(void)
2823 {
2824 return ust_app_ht;
2825 }
2826
2827 /*
2828 * Return ust app pointer or NULL if not found. RCU read side lock MUST be
2829 * acquired before calling this function.
2830 */
2831 struct ust_app *ust_app_find_by_pid(pid_t pid)
2832 {
2833 struct ust_app *app = NULL;
2834 struct lttng_ht_node_ulong *node;
2835 struct lttng_ht_iter iter;
2836
2837 lttng_ht_lookup(ust_app_ht, (void *)((unsigned long) pid), &iter);
2838 node = lttng_ht_iter_get_node_ulong(&iter);
2839 if (node == NULL) {
2840 DBG2("UST app no found with pid %d", pid);
2841 goto error;
2842 }
2843
2844 DBG2("Found UST app by pid %d", pid);
2845
2846 app = caa_container_of(node, struct ust_app, pid_n);
2847
2848 error:
2849 return app;
2850 }
2851
2852 /*
2853 * Allocate and init an UST app object using the registration information and
2854 * the command socket. This is called when the command socket connects to the
2855 * session daemon.
2856 *
2857 * The object is returned on success or else NULL.
2858 */
2859 struct ust_app *ust_app_create(struct ust_register_msg *msg, int sock)
2860 {
2861 struct ust_app *lta = NULL;
2862
2863 assert(msg);
2864 assert(sock >= 0);
2865
2866 DBG3("UST app creating application for socket %d", sock);
2867
2868 if ((msg->bits_per_long == 64 &&
2869 (uatomic_read(&ust_consumerd64_fd) == -EINVAL))
2870 || (msg->bits_per_long == 32 &&
2871 (uatomic_read(&ust_consumerd32_fd) == -EINVAL))) {
2872 ERR("Registration failed: application \"%s\" (pid: %d) has "
2873 "%d-bit long, but no consumerd for this size is available.\n",
2874 msg->name, msg->pid, msg->bits_per_long);
2875 goto error;
2876 }
2877
2878 lta = zmalloc(sizeof(struct ust_app));
2879 if (lta == NULL) {
2880 PERROR("malloc");
2881 goto error;
2882 }
2883
2884 lta->ppid = msg->ppid;
2885 lta->uid = msg->uid;
2886 lta->gid = msg->gid;
2887
2888 lta->bits_per_long = msg->bits_per_long;
2889 lta->uint8_t_alignment = msg->uint8_t_alignment;
2890 lta->uint16_t_alignment = msg->uint16_t_alignment;
2891 lta->uint32_t_alignment = msg->uint32_t_alignment;
2892 lta->uint64_t_alignment = msg->uint64_t_alignment;
2893 lta->long_alignment = msg->long_alignment;
2894 lta->byte_order = msg->byte_order;
2895
2896 lta->v_major = msg->major;
2897 lta->v_minor = msg->minor;
2898 lta->sessions = lttng_ht_new(0, LTTNG_HT_TYPE_U64);
2899 lta->ust_objd = lttng_ht_new(0, LTTNG_HT_TYPE_ULONG);
2900 lta->notify_sock = -1;
2901
2902 /* Copy name and make sure it's NULL terminated. */
2903 strncpy(lta->name, msg->name, sizeof(lta->name));
2904 lta->name[UST_APP_PROCNAME_LEN] = '\0';
2905
2906 /*
2907 * Before this can be called, when receiving the registration information,
2908 * the application compatibility is checked. So, at this point, the
2909 * application can work with this session daemon.
2910 */
2911 lta->compatible = 1;
2912
2913 lta->pid = msg->pid;
2914 lttng_ht_node_init_ulong(&lta->pid_n, (unsigned long) lta->pid);
2915 lta->sock = sock;
2916 lttng_ht_node_init_ulong(&lta->sock_n, (unsigned long) lta->sock);
2917
2918 CDS_INIT_LIST_HEAD(&lta->teardown_head);
2919
2920 error:
2921 return lta;
2922 }
2923
2924 /*
2925 * For a given application object, add it to every hash table.
2926 */
2927 void ust_app_add(struct ust_app *app)
2928 {
2929 assert(app);
2930 assert(app->notify_sock >= 0);
2931
2932 rcu_read_lock();
2933
2934 /*
2935 * On a re-registration, we want to kick out the previous registration of
2936 * that pid
2937 */
2938 lttng_ht_add_replace_ulong(ust_app_ht, &app->pid_n);
2939
2940 /*
2941 * The socket _should_ be unique until _we_ call close. So, a add_unique
2942 * for the ust_app_ht_by_sock is used which asserts fail if the entry was
2943 * already in the table.
2944 */
2945 lttng_ht_add_unique_ulong(ust_app_ht_by_sock, &app->sock_n);
2946
2947 /* Add application to the notify socket hash table. */
2948 lttng_ht_node_init_ulong(&app->notify_sock_n, app->notify_sock);
2949 lttng_ht_add_unique_ulong(ust_app_ht_by_notify_sock, &app->notify_sock_n);
2950
2951 DBG("App registered with pid:%d ppid:%d uid:%d gid:%d sock:%d name:%s "
2952 "notify_sock:%d (version %d.%d)", app->pid, app->ppid, app->uid,
2953 app->gid, app->sock, app->name, app->notify_sock, app->v_major,
2954 app->v_minor);
2955
2956 rcu_read_unlock();
2957 }
2958
2959 /*
2960 * Set the application version into the object.
2961 *
2962 * Return 0 on success else a negative value either an errno code or a
2963 * LTTng-UST error code.
2964 */
2965 int ust_app_version(struct ust_app *app)
2966 {
2967 int ret;
2968
2969 assert(app);
2970
2971 ret = ustctl_tracer_version(app->sock, &app->version);
2972 if (ret < 0) {
2973 if (ret != -LTTNG_UST_ERR_EXITING && ret != -EPIPE) {
2974 ERR("UST app %d verson failed with ret %d", app->sock, ret);
2975 } else {
2976 DBG3("UST app %d verion failed. Application is dead", app->sock);
2977 }
2978 }
2979
2980 return ret;
2981 }
2982
2983 /*
2984 * Unregister app by removing it from the global traceable app list and freeing
2985 * the data struct.
2986 *
2987 * The socket is already closed at this point so no close to sock.
2988 */
2989 void ust_app_unregister(int sock)
2990 {
2991 struct ust_app *lta;
2992 struct lttng_ht_node_ulong *node;
2993 struct lttng_ht_iter iter;
2994 struct ust_app_session *ua_sess;
2995 int ret;
2996
2997 rcu_read_lock();
2998
2999 /* Get the node reference for a call_rcu */
3000 lttng_ht_lookup(ust_app_ht_by_sock, (void *)((unsigned long) sock), &iter);
3001 node = lttng_ht_iter_get_node_ulong(&iter);
3002 assert(node);
3003
3004 lta = caa_container_of(node, struct ust_app, sock_n);
3005 DBG("PID %d unregistering with sock %d", lta->pid, sock);
3006
3007 /* Remove application from PID hash table */
3008 ret = lttng_ht_del(ust_app_ht_by_sock, &iter);
3009 assert(!ret);
3010
3011 /*
3012 * Remove application from notify hash table. The thread handling the
3013 * notify socket could have deleted the node so ignore on error because
3014 * either way it's valid. The close of that socket is handled by the other
3015 * thread.
3016 */
3017 iter.iter.node = &lta->notify_sock_n.node;
3018 (void) lttng_ht_del(ust_app_ht_by_notify_sock, &iter);
3019
3020 /*
3021 * Ignore return value since the node might have been removed before by an
3022 * add replace during app registration because the PID can be reassigned by
3023 * the OS.
3024 */
3025 iter.iter.node = &lta->pid_n.node;
3026 ret = lttng_ht_del(ust_app_ht, &iter);
3027 if (ret) {
3028 DBG3("Unregister app by PID %d failed. This can happen on pid reuse",
3029 lta->pid);
3030 }
3031
3032 /* Remove sessions so they are not visible during deletion.*/
3033 cds_lfht_for_each_entry(lta->sessions->ht, &iter.iter, ua_sess,
3034 node.node) {
3035 struct ust_registry_session *registry;
3036
3037 ret = lttng_ht_del(lta->sessions, &iter);
3038 if (ret) {
3039 /* The session was already removed so scheduled for teardown. */
3040 continue;
3041 }
3042
3043 /*
3044 * Add session to list for teardown. This is safe since at this point we
3045 * are the only one using this list.
3046 */
3047 pthread_mutex_lock(&ua_sess->lock);
3048
3049 /*
3050 * Normally, this is done in the delete session process which is
3051 * executed in the call rcu below. However, upon registration we can't
3052 * afford to wait for the grace period before pushing data or else the
3053 * data pending feature can race between the unregistration and stop
3054 * command where the data pending command is sent *before* the grace
3055 * period ended.
3056 *
3057 * The close metadata below nullifies the metadata pointer in the
3058 * session so the delete session will NOT push/close a second time.
3059 */
3060 registry = get_session_registry(ua_sess);
3061 if (registry && !registry->metadata_closed) {
3062 /* Push metadata for application before freeing the application. */
3063 (void) push_metadata(registry, ua_sess->consumer);
3064
3065 /*
3066 * Don't ask to close metadata for global per UID buffers. Close
3067 * metadata only on destroy trace session in this case. Also, the
3068 * previous push metadata could have flag the metadata registry to
3069 * close so don't send a close command if closed.
3070 */
3071 if (ua_sess->buffer_type != LTTNG_BUFFER_PER_UID &&
3072 !registry->metadata_closed) {
3073 /* And ask to close it for this session registry. */
3074 (void) close_metadata(registry, ua_sess->consumer);
3075 }
3076 }
3077
3078 cds_list_add(&ua_sess->teardown_node, &lta->teardown_head);
3079 pthread_mutex_unlock(&ua_sess->lock);
3080 }
3081
3082 /* Free memory */
3083 call_rcu(&lta->pid_n.head, delete_ust_app_rcu);
3084
3085 rcu_read_unlock();
3086 return;
3087 }
3088
3089 /*
3090 * Return traceable_app_count
3091 */
3092 unsigned long ust_app_list_count(void)
3093 {
3094 unsigned long count;
3095
3096 rcu_read_lock();
3097 count = lttng_ht_get_count(ust_app_ht);
3098 rcu_read_unlock();
3099
3100 return count;
3101 }
3102
3103 /*
3104 * Fill events array with all events name of all registered apps.
3105 */
3106 int ust_app_list_events(struct lttng_event **events)
3107 {
3108 int ret, handle;
3109 size_t nbmem, count = 0;
3110 struct lttng_ht_iter iter;
3111 struct ust_app *app;
3112 struct lttng_event *tmp_event;
3113
3114 nbmem = UST_APP_EVENT_LIST_SIZE;
3115 tmp_event = zmalloc(nbmem * sizeof(struct lttng_event));
3116 if (tmp_event == NULL) {
3117 PERROR("zmalloc ust app events");
3118 ret = -ENOMEM;
3119 goto error;
3120 }
3121
3122 rcu_read_lock();
3123
3124 cds_lfht_for_each_entry(ust_app_ht->ht, &iter.iter, app, pid_n.node) {
3125 struct lttng_ust_tracepoint_iter uiter;
3126
3127 health_code_update();
3128
3129 if (!app->compatible) {
3130 /*
3131 * TODO: In time, we should notice the caller of this error by
3132 * telling him that this is a version error.
3133 */
3134 continue;
3135 }
3136 handle = ustctl_tracepoint_list(app->sock);
3137 if (handle < 0) {
3138 if (handle != -EPIPE && handle != -LTTNG_UST_ERR_EXITING) {
3139 ERR("UST app list events getting handle failed for app pid %d",
3140 app->pid);
3141 }
3142 continue;
3143 }
3144
3145 while ((ret = ustctl_tracepoint_list_get(app->sock, handle,
3146 &uiter)) != -LTTNG_UST_ERR_NOENT) {
3147 /* Handle ustctl error. */
3148 if (ret < 0) {
3149 free(tmp_event);
3150 if (ret != -LTTNG_UST_ERR_EXITING || ret != -EPIPE) {
3151 ERR("UST app tp list get failed for app %d with ret %d",
3152 app->sock, ret);
3153 } else {
3154 DBG3("UST app tp list get failed. Application is dead");
3155 /*
3156 * This is normal behavior, an application can die during the
3157 * creation process. Don't report an error so the execution can
3158 * continue normally. Continue normal execution.
3159 */
3160 break;
3161 }
3162 goto rcu_error;
3163 }
3164
3165 health_code_update();
3166 if (count >= nbmem) {
3167 /* In case the realloc fails, we free the memory */
3168 void *ptr;
3169
3170 DBG2("Reallocating event list from %zu to %zu entries", nbmem,
3171 2 * nbmem);
3172 nbmem *= 2;
3173 ptr = realloc(tmp_event, nbmem * sizeof(struct lttng_event));
3174 if (ptr == NULL) {
3175 PERROR("realloc ust app events");
3176 free(tmp_event);
3177 ret = -ENOMEM;
3178 goto rcu_error;
3179 }
3180 tmp_event = ptr;
3181 }
3182 memcpy(tmp_event[count].name, uiter.name, LTTNG_UST_SYM_NAME_LEN);
3183 tmp_event[count].loglevel = uiter.loglevel;
3184 tmp_event[count].type = (enum lttng_event_type) LTTNG_UST_TRACEPOINT;
3185 tmp_event[count].pid = app->pid;
3186 tmp_event[count].enabled = -1;
3187 count++;
3188 }
3189 }
3190
3191 ret = count;
3192 *events = tmp_event;
3193
3194 DBG2("UST app list events done (%zu events)", count);
3195
3196 rcu_error:
3197 rcu_read_unlock();
3198 error:
3199 health_code_update();
3200 return ret;
3201 }
3202
3203 /*
3204 * Fill events array with all events name of all registered apps.
3205 */
3206 int ust_app_list_event_fields(struct lttng_event_field **fields)
3207 {
3208 int ret, handle;
3209 size_t nbmem, count = 0;
3210 struct lttng_ht_iter iter;
3211 struct ust_app *app;
3212 struct lttng_event_field *tmp_event;
3213
3214 nbmem = UST_APP_EVENT_LIST_SIZE;
3215 tmp_event = zmalloc(nbmem * sizeof(struct lttng_event_field));
3216 if (tmp_event == NULL) {
3217 PERROR("zmalloc ust app event fields");
3218 ret = -ENOMEM;
3219 goto error;
3220 }
3221
3222 rcu_read_lock();
3223
3224 cds_lfht_for_each_entry(ust_app_ht->ht, &iter.iter, app, pid_n.node) {
3225 struct lttng_ust_field_iter uiter;
3226
3227 health_code_update();
3228
3229 if (!app->compatible) {
3230 /*
3231 * TODO: In time, we should notice the caller of this error by
3232 * telling him that this is a version error.
3233 */
3234 continue;
3235 }
3236 handle = ustctl_tracepoint_field_list(app->sock);
3237 if (handle < 0) {
3238 if (handle != -EPIPE && handle != -LTTNG_UST_ERR_EXITING) {
3239 ERR("UST app list field getting handle failed for app pid %d",
3240 app->pid);
3241 }
3242 continue;
3243 }
3244
3245 while ((ret = ustctl_tracepoint_field_list_get(app->sock, handle,
3246 &uiter)) != -LTTNG_UST_ERR_NOENT) {
3247 /* Handle ustctl error. */
3248 if (ret < 0) {
3249 free(tmp_event);
3250 if (ret != -LTTNG_UST_ERR_EXITING || ret != -EPIPE) {
3251 ERR("UST app tp list field failed for app %d with ret %d",
3252 app->sock, ret);
3253 } else {
3254 DBG3("UST app tp list field failed. Application is dead");
3255 /*
3256 * This is normal behavior, an application can die during the
3257 * creation process. Don't report an error so the execution can
3258 * continue normally.
3259 */
3260 break;
3261 }
3262 goto rcu_error;
3263 }
3264
3265 health_code_update();
3266 if (count >= nbmem) {
3267 /* In case the realloc fails, we free the memory */
3268 void *ptr;
3269
3270 DBG2("Reallocating event field list from %zu to %zu entries", nbmem,
3271 2 * nbmem);
3272 nbmem *= 2;
3273 ptr = realloc(tmp_event, nbmem * sizeof(struct lttng_event_field));
3274 if (ptr == NULL) {
3275 PERROR("realloc ust app event fields");
3276 free(tmp_event);
3277 ret = -ENOMEM;
3278 goto rcu_error;
3279 }
3280 tmp_event = ptr;
3281 }
3282
3283 memcpy(tmp_event[count].field_name, uiter.field_name, LTTNG_UST_SYM_NAME_LEN);
3284 /* Mapping between these enums matches 1 to 1. */
3285 tmp_event[count].type = (enum lttng_event_field_type) uiter.type;
3286 tmp_event[count].nowrite = uiter.nowrite;
3287
3288 memcpy(tmp_event[count].event.name, uiter.event_name, LTTNG_UST_SYM_NAME_LEN);
3289 tmp_event[count].event.loglevel = uiter.loglevel;
3290 tmp_event[count].event.type = LTTNG_EVENT_TRACEPOINT;
3291 tmp_event[count].event.pid = app->pid;
3292 tmp_event[count].event.enabled = -1;
3293 count++;
3294 }
3295 }
3296
3297 ret = count;
3298 *fields = tmp_event;
3299
3300 DBG2("UST app list event fields done (%zu events)", count);
3301
3302 rcu_error:
3303 rcu_read_unlock();
3304 error:
3305 health_code_update();
3306 return ret;
3307 }
3308
3309 /*
3310 * Free and clean all traceable apps of the global list.
3311 *
3312 * Should _NOT_ be called with RCU read-side lock held.
3313 */
3314 void ust_app_clean_list(void)
3315 {
3316 int ret;
3317 struct ust_app *app;
3318 struct lttng_ht_iter iter;
3319
3320 DBG2("UST app cleaning registered apps hash table");
3321
3322 rcu_read_lock();
3323
3324 cds_lfht_for_each_entry(ust_app_ht->ht, &iter.iter, app, pid_n.node) {
3325 ret = lttng_ht_del(ust_app_ht, &iter);
3326 assert(!ret);
3327 call_rcu(&app->pid_n.head, delete_ust_app_rcu);
3328 }
3329
3330 /* Cleanup socket hash table */
3331 cds_lfht_for_each_entry(ust_app_ht_by_sock->ht, &iter.iter, app,
3332 sock_n.node) {
3333 ret = lttng_ht_del(ust_app_ht_by_sock, &iter);
3334 assert(!ret);
3335 }
3336
3337 /* Cleanup notify socket hash table */
3338 cds_lfht_for_each_entry(ust_app_ht_by_notify_sock->ht, &iter.iter, app,
3339 notify_sock_n.node) {
3340 ret = lttng_ht_del(ust_app_ht_by_notify_sock, &iter);
3341 assert(!ret);
3342 }
3343 rcu_read_unlock();
3344
3345 /* Destroy is done only when the ht is empty */
3346 ht_cleanup_push(ust_app_ht);
3347 ht_cleanup_push(ust_app_ht_by_sock);
3348 ht_cleanup_push(ust_app_ht_by_notify_sock);
3349 }
3350
3351 /*
3352 * Init UST app hash table.
3353 */
3354 void ust_app_ht_alloc(void)
3355 {
3356 ust_app_ht = lttng_ht_new(0, LTTNG_HT_TYPE_ULONG);
3357 ust_app_ht_by_sock = lttng_ht_new(0, LTTNG_HT_TYPE_ULONG);
3358 ust_app_ht_by_notify_sock = lttng_ht_new(0, LTTNG_HT_TYPE_ULONG);
3359 }
3360
3361 /*
3362 * For a specific UST session, disable the channel for all registered apps.
3363 */
3364 int ust_app_disable_channel_glb(struct ltt_ust_session *usess,
3365 struct ltt_ust_channel *uchan)
3366 {
3367 int ret = 0;
3368 struct lttng_ht_iter iter;
3369 struct lttng_ht_node_str *ua_chan_node;
3370 struct ust_app *app;
3371 struct ust_app_session *ua_sess;
3372 struct ust_app_channel *ua_chan;
3373
3374 if (usess == NULL || uchan == NULL) {
3375 ERR("Disabling UST global channel with NULL values");
3376 ret = -1;
3377 goto error;
3378 }
3379
3380 DBG2("UST app disabling channel %s from global domain for session id %" PRIu64,
3381 uchan->name, usess->id);
3382
3383 rcu_read_lock();
3384
3385 /* For every registered applications */
3386 cds_lfht_for_each_entry(ust_app_ht->ht, &iter.iter, app, pid_n.node) {
3387 struct lttng_ht_iter uiter;
3388 if (!app->compatible) {
3389 /*
3390 * TODO: In time, we should notice the caller of this error by
3391 * telling him that this is a version error.
3392 */
3393 continue;
3394 }
3395 ua_sess = lookup_session_by_app(usess, app);
3396 if (ua_sess == NULL) {
3397 continue;
3398 }
3399
3400 /* Get channel */
3401 lttng_ht_lookup(ua_sess->channels, (void *)uchan->name, &uiter);
3402 ua_chan_node = lttng_ht_iter_get_node_str(&uiter);
3403 /* If the session if found for the app, the channel must be there */
3404 assert(ua_chan_node);
3405
3406 ua_chan = caa_container_of(ua_chan_node, struct ust_app_channel, node);
3407 /* The channel must not be already disabled */
3408 assert(ua_chan->enabled == 1);
3409
3410 /* Disable channel onto application */
3411 ret = disable_ust_app_channel(ua_sess, ua_chan, app);
3412 if (ret < 0) {
3413 /* XXX: We might want to report this error at some point... */
3414 continue;
3415 }
3416 }
3417
3418 rcu_read_unlock();
3419
3420 error:
3421 return ret;
3422 }
3423
3424 /*
3425 * For a specific UST session, enable the channel for all registered apps.
3426 */
3427 int ust_app_enable_channel_glb(struct ltt_ust_session *usess,
3428 struct ltt_ust_channel *uchan)
3429 {
3430 int ret = 0;
3431 struct lttng_ht_iter iter;
3432 struct ust_app *app;
3433 struct ust_app_session *ua_sess;
3434
3435 if (usess == NULL || uchan == NULL) {
3436 ERR("Adding UST global channel to NULL values");
3437 ret = -1;
3438 goto error;
3439 }
3440
3441 DBG2("UST app enabling channel %s to global domain for session id %" PRIu64,
3442 uchan->name, usess->id);
3443
3444 rcu_read_lock();
3445
3446 /* For every registered applications */
3447 cds_lfht_for_each_entry(ust_app_ht->ht, &iter.iter, app, pid_n.node) {
3448 if (!app->compatible) {
3449 /*
3450 * TODO: In time, we should notice the caller of this error by
3451 * telling him that this is a version error.
3452 */
3453 continue;
3454 }
3455 ua_sess = lookup_session_by_app(usess, app);
3456 if (ua_sess == NULL) {
3457 continue;
3458 }
3459
3460 /* Enable channel onto application */
3461 ret = enable_ust_app_channel(ua_sess, uchan, app);
3462 if (ret < 0) {
3463 /* XXX: We might want to report this error at some point... */
3464 continue;
3465 }
3466 }
3467
3468 rcu_read_unlock();
3469
3470 error:
3471 return ret;
3472 }
3473
3474 /*
3475 * Disable an event in a channel and for a specific session.
3476 */
3477 int ust_app_disable_event_glb(struct ltt_ust_session *usess,
3478 struct ltt_ust_channel *uchan, struct ltt_ust_event *uevent)
3479 {
3480 int ret = 0;
3481 struct lttng_ht_iter iter, uiter;
3482 struct lttng_ht_node_str *ua_chan_node, *ua_event_node;
3483 struct ust_app *app;
3484 struct ust_app_session *ua_sess;
3485 struct ust_app_channel *ua_chan;
3486 struct ust_app_event *ua_event;
3487
3488 DBG("UST app disabling event %s for all apps in channel "
3489 "%s for session id %" PRIu64,
3490 uevent->attr.name, uchan->name, usess->id);
3491
3492 rcu_read_lock();
3493
3494 /* For all registered applications */
3495 cds_lfht_for_each_entry(ust_app_ht->ht, &iter.iter, app, pid_n.node) {
3496 if (!app->compatible) {
3497 /*
3498 * TODO: In time, we should notice the caller of this error by
3499 * telling him that this is a version error.
3500 */
3501 continue;
3502 }
3503 ua_sess = lookup_session_by_app(usess, app);
3504 if (ua_sess == NULL) {
3505 /* Next app */
3506 continue;
3507 }
3508
3509 /* Lookup channel in the ust app session */
3510 lttng_ht_lookup(ua_sess->channels, (void *)uchan->name, &uiter);
3511 ua_chan_node = lttng_ht_iter_get_node_str(&uiter);
3512 if (ua_chan_node == NULL) {
3513 DBG2("Channel %s not found in session id %" PRIu64 " for app pid %d."
3514 "Skipping", uchan->name, usess->id, app->pid);
3515 continue;
3516 }
3517 ua_chan = caa_container_of(ua_chan_node, struct ust_app_channel, node);
3518
3519 lttng_ht_lookup(ua_chan->events, (void *)uevent->attr.name, &uiter);
3520 ua_event_node = lttng_ht_iter_get_node_str(&uiter);
3521 if (ua_event_node == NULL) {
3522 DBG2("Event %s not found in channel %s for app pid %d."
3523 "Skipping", uevent->attr.name, uchan->name, app->pid);
3524 continue;
3525 }
3526 ua_event = caa_container_of(ua_event_node, struct ust_app_event, node);
3527
3528 ret = disable_ust_app_event(ua_sess, ua_event, app);
3529 if (ret < 0) {
3530 /* XXX: Report error someday... */
3531 continue;
3532 }
3533 }
3534
3535 rcu_read_unlock();
3536
3537 return ret;
3538 }
3539
3540 /*
3541 * For a specific UST session and UST channel, the event for all
3542 * registered apps.
3543 */
3544 int ust_app_disable_all_event_glb(struct ltt_ust_session *usess,
3545 struct ltt_ust_channel *uchan)
3546 {
3547 int ret = 0;
3548 struct lttng_ht_iter iter, uiter;
3549 struct lttng_ht_node_str *ua_chan_node;
3550 struct ust_app *app;
3551 struct ust_app_session *ua_sess;
3552 struct ust_app_channel *ua_chan;
3553 struct ust_app_event *ua_event;
3554
3555 DBG("UST app disabling all event for all apps in channel "
3556 "%s for session id %" PRIu64, uchan->name, usess->id);
3557
3558 rcu_read_lock();
3559
3560 /* For all registered applications */
3561 cds_lfht_for_each_entry(ust_app_ht->ht, &iter.iter, app, pid_n.node) {
3562 if (!app->compatible) {
3563 /*
3564 * TODO: In time, we should notice the caller of this error by
3565 * telling him that this is a version error.
3566 */
3567 continue;
3568 }
3569 ua_sess = lookup_session_by_app(usess, app);
3570 if (!ua_sess) {
3571 /* The application has problem or is probably dead. */
3572 continue;
3573 }
3574
3575 /* Lookup channel in the ust app session */
3576 lttng_ht_lookup(ua_sess->channels, (void *)uchan->name, &uiter);
3577 ua_chan_node = lttng_ht_iter_get_node_str(&uiter);
3578 /* If the channel is not found, there is a code flow error */
3579 assert(ua_chan_node);
3580
3581 ua_chan = caa_container_of(ua_chan_node, struct ust_app_channel, node);
3582
3583 /* Disable each events of channel */
3584 cds_lfht_for_each_entry(ua_chan->events->ht, &uiter.iter, ua_event,
3585 node.node) {
3586 ret = disable_ust_app_event(ua_sess, ua_event, app);
3587 if (ret < 0) {
3588 /* XXX: Report error someday... */
3589 continue;
3590 }
3591 }
3592 }
3593
3594 rcu_read_unlock();
3595
3596 return ret;
3597 }
3598
3599 /*
3600 * For a specific UST session, create the channel for all registered apps.
3601 */
3602 int ust_app_create_channel_glb(struct ltt_ust_session *usess,
3603 struct ltt_ust_channel *uchan)
3604 {
3605 int ret = 0, created;
3606 struct lttng_ht_iter iter;
3607 struct ust_app *app;
3608 struct ust_app_session *ua_sess = NULL;
3609
3610 /* Very wrong code flow */
3611 assert(usess);
3612 assert(uchan);
3613
3614 DBG2("UST app adding channel %s to UST domain for session id %" PRIu64,
3615 uchan->name, usess->id);
3616
3617 rcu_read_lock();
3618
3619 /* For every registered applications */
3620 cds_lfht_for_each_entry(ust_app_ht->ht, &iter.iter, app, pid_n.node) {
3621 if (!app->compatible) {
3622 /*
3623 * TODO: In time, we should notice the caller of this error by
3624 * telling him that this is a version error.
3625 */
3626 continue;
3627 }
3628 /*
3629 * Create session on the tracer side and add it to app session HT. Note
3630 * that if session exist, it will simply return a pointer to the ust
3631 * app session.
3632 */
3633 ret = create_ust_app_session(usess, app, &ua_sess, &created);
3634 if (ret < 0) {
3635 switch (ret) {
3636 case -ENOTCONN:
3637 /*
3638 * The application's socket is not valid. Either a bad socket
3639 * or a timeout on it. We can't inform the caller that for a
3640 * specific app, the session failed so lets continue here.
3641 */
3642 continue;
3643 case -ENOMEM:
3644 default:
3645 goto error_rcu_unlock;
3646 }
3647 }
3648 assert(ua_sess);
3649
3650 pthread_mutex_lock(&ua_sess->lock);
3651 if (!strncmp(uchan->name, DEFAULT_METADATA_NAME,
3652 sizeof(uchan->name))) {
3653 struct ustctl_consumer_channel_attr attr;
3654 copy_channel_attr_to_ustctl(&attr, &uchan->attr);
3655 ret = create_ust_app_metadata(ua_sess, app, usess->consumer,
3656 &attr);
3657 } else {
3658 /* Create channel onto application. We don't need the chan ref. */
3659 ret = create_ust_app_channel(ua_sess, uchan, app,
3660 LTTNG_UST_CHAN_PER_CPU, usess, NULL);
3661 }
3662 pthread_mutex_unlock(&ua_sess->lock);
3663 if (ret < 0) {
3664 if (ret == -ENOMEM) {
3665 /* No more memory is a fatal error. Stop right now. */
3666 goto error_rcu_unlock;
3667 }
3668 /* Cleanup the created session if it's the case. */
3669 if (created) {
3670 destroy_app_session(app, ua_sess);
3671 }
3672 }
3673 }
3674
3675 error_rcu_unlock:
3676 rcu_read_unlock();
3677 return ret;
3678 }
3679
3680 /*
3681 * Enable event for a specific session and channel on the tracer.
3682 */
3683 int ust_app_enable_event_glb(struct ltt_ust_session *usess,
3684 struct ltt_ust_channel *uchan, struct ltt_ust_event *uevent)
3685 {
3686 int ret = 0;
3687 struct lttng_ht_iter iter, uiter;
3688 struct lttng_ht_node_str *ua_chan_node;
3689 struct ust_app *app;
3690 struct ust_app_session *ua_sess;
3691 struct ust_app_channel *ua_chan;
3692 struct ust_app_event *ua_event;
3693
3694 DBG("UST app enabling event %s for all apps for session id %" PRIu64,
3695 uevent->attr.name, usess->id);
3696
3697 /*
3698 * NOTE: At this point, this function is called only if the session and
3699 * channel passed are already created for all apps. and enabled on the
3700 * tracer also.
3701 */
3702
3703 rcu_read_lock();
3704
3705 /* For all registered applications */
3706 cds_lfht_for_each_entry(ust_app_ht->ht, &iter.iter, app, pid_n.node) {
3707 if (!app->compatible) {
3708 /*
3709 * TODO: In time, we should notice the caller of this error by
3710 * telling him that this is a version error.
3711 */
3712 continue;
3713 }
3714 ua_sess = lookup_session_by_app(usess, app);
3715 if (!ua_sess) {
3716 /* The application has problem or is probably dead. */
3717 continue;
3718 }
3719
3720 pthread_mutex_lock(&ua_sess->lock);
3721
3722 /* Lookup channel in the ust app session */
3723 lttng_ht_lookup(ua_sess->channels, (void *)uchan->name, &uiter);
3724 ua_chan_node = lttng_ht_iter_get_node_str(&uiter);
3725 /* If the channel is not found, there is a code flow error */
3726 assert(ua_chan_node);
3727
3728 ua_chan = caa_container_of(ua_chan_node, struct ust_app_channel, node);
3729
3730 /* Get event node */
3731 ua_event = find_ust_app_event(ua_chan->events, uevent->attr.name,
3732 uevent->filter, uevent->attr.loglevel, uevent->exclusion);
3733 if (ua_event == NULL) {
3734 DBG3("UST app enable event %s not found for app PID %d."
3735 "Skipping app", uevent->attr.name, app->pid);
3736 goto next_app;
3737 }
3738
3739 ret = enable_ust_app_event(ua_sess, ua_event, app);
3740 if (ret < 0) {
3741 pthread_mutex_unlock(&ua_sess->lock);
3742 goto error;
3743 }
3744 next_app:
3745 pthread_mutex_unlock(&ua_sess->lock);
3746 }
3747
3748 error:
3749 rcu_read_unlock();
3750 return ret;
3751 }
3752
3753 /*
3754 * For a specific existing UST session and UST channel, creates the event for
3755 * all registered apps.
3756 */
3757 int ust_app_create_event_glb(struct ltt_ust_session *usess,
3758 struct ltt_ust_channel *uchan, struct ltt_ust_event *uevent)
3759 {
3760 int ret = 0;
3761 struct lttng_ht_iter iter, uiter;
3762 struct lttng_ht_node_str *ua_chan_node;
3763 struct ust_app *app;
3764 struct ust_app_session *ua_sess;
3765 struct ust_app_channel *ua_chan;
3766
3767 DBG("UST app creating event %s for all apps for session id %" PRIu64,
3768 uevent->attr.name, usess->id);
3769
3770 rcu_read_lock();
3771
3772 /* For all registered applications */
3773 cds_lfht_for_each_entry(ust_app_ht->ht, &iter.iter, app, pid_n.node) {
3774 if (!app->compatible) {
3775 /*
3776 * TODO: In time, we should notice the caller of this error by
3777 * telling him that this is a version error.
3778 */
3779 continue;
3780 }
3781 ua_sess = lookup_session_by_app(usess, app);
3782 if (!ua_sess) {
3783 /* The application has problem or is probably dead. */
3784 continue;
3785 }
3786
3787 pthread_mutex_lock(&ua_sess->lock);
3788 /* Lookup channel in the ust app session */
3789 lttng_ht_lookup(ua_sess->channels, (void *)uchan->name, &uiter);
3790 ua_chan_node = lttng_ht_iter_get_node_str(&uiter);
3791 /* If the channel is not found, there is a code flow error */
3792 assert(ua_chan_node);
3793
3794 ua_chan = caa_container_of(ua_chan_node, struct ust_app_channel, node);
3795
3796 ret = create_ust_app_event(ua_sess, ua_chan, uevent, app);
3797 pthread_mutex_unlock(&ua_sess->lock);
3798 if (ret < 0) {
3799 if (ret != -LTTNG_UST_ERR_EXIST) {
3800 /* Possible value at this point: -ENOMEM. If so, we stop! */
3801 break;
3802 }
3803 DBG2("UST app event %s already exist on app PID %d",
3804 uevent->attr.name, app->pid);
3805 continue;
3806 }
3807 }
3808
3809 rcu_read_unlock();
3810
3811 return ret;
3812 }
3813
3814 /*
3815 * Start tracing for a specific UST session and app.
3816 */
3817 static
3818 int ust_app_start_trace(struct ltt_ust_session *usess, struct ust_app *app)
3819 {
3820 int ret = 0;
3821 struct ust_app_session *ua_sess;
3822
3823 DBG("Starting tracing for ust app pid %d", app->pid);
3824
3825 rcu_read_lock();
3826
3827 if (!app->compatible) {
3828 goto end;
3829 }
3830
3831 ua_sess = lookup_session_by_app(usess, app);
3832 if (ua_sess == NULL) {
3833 /* The session is in teardown process. Ignore and continue. */
3834 goto end;
3835 }
3836
3837 pthread_mutex_lock(&ua_sess->lock);
3838
3839 /* Upon restart, we skip the setup, already done */
3840 if (ua_sess->started) {
3841 goto skip_setup;
3842 }
3843
3844 /* Create directories if consumer is LOCAL and has a path defined. */
3845 if (usess->consumer->type == CONSUMER_DST_LOCAL &&
3846 strlen(usess->consumer->dst.trace_path) > 0) {
3847 ret = run_as_mkdir_recursive(usess->consumer->dst.trace_path,
3848 S_IRWXU | S_IRWXG, ua_sess->euid, ua_sess->egid);
3849 if (ret < 0) {
3850 if (ret != -EEXIST) {
3851 ERR("Trace directory creation error");
3852 goto error_unlock;
3853 }
3854 }
3855 }
3856
3857 /*
3858 * Create the metadata for the application. This returns gracefully if a
3859 * metadata was already set for the session.
3860 */
3861 ret = create_ust_app_metadata(ua_sess, app, usess->consumer, NULL);
3862 if (ret < 0) {
3863 goto error_unlock;
3864 }
3865
3866 health_code_update();
3867
3868 skip_setup:
3869 /* This start the UST tracing */
3870 ret = ustctl_start_session(app->sock, ua_sess->handle);
3871 if (ret < 0) {
3872 if (ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
3873 ERR("Error starting tracing for app pid: %d (ret: %d)",
3874 app->pid, ret);
3875 } else {
3876 DBG("UST app start session failed. Application is dead.");
3877 /*
3878 * This is normal behavior, an application can die during the
3879 * creation process. Don't report an error so the execution can
3880 * continue normally.
3881 */
3882 pthread_mutex_unlock(&ua_sess->lock);
3883 goto end;
3884 }
3885 goto error_unlock;
3886 }
3887
3888 /* Indicate that the session has been started once */
3889 ua_sess->started = 1;
3890
3891 pthread_mutex_unlock(&ua_sess->lock);
3892
3893 health_code_update();
3894
3895 /* Quiescent wait after starting trace */
3896 ret = ustctl_wait_quiescent(app->sock);
3897 if (ret < 0 && ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
3898 ERR("UST app wait quiescent failed for app pid %d ret %d",
3899 app->pid, ret);
3900 }
3901
3902 end:
3903 rcu_read_unlock();
3904 health_code_update();
3905 return 0;
3906
3907 error_unlock:
3908 pthread_mutex_unlock(&ua_sess->lock);
3909 rcu_read_unlock();
3910 health_code_update();
3911 return -1;
3912 }
3913
3914 /*
3915 * Stop tracing for a specific UST session and app.
3916 */
3917 static
3918 int ust_app_stop_trace(struct ltt_ust_session *usess, struct ust_app *app)
3919 {
3920 int ret = 0;
3921 struct ust_app_session *ua_sess;
3922 struct ust_registry_session *registry;
3923
3924 DBG("Stopping tracing for ust app pid %d", app->pid);
3925
3926 rcu_read_lock();
3927
3928 if (!app->compatible) {
3929 goto end_no_session;
3930 }
3931
3932 ua_sess = lookup_session_by_app(usess, app);
3933 if (ua_sess == NULL) {
3934 goto end_no_session;
3935 }
3936
3937 pthread_mutex_lock(&ua_sess->lock);
3938
3939 /*
3940 * If started = 0, it means that stop trace has been called for a session
3941 * that was never started. It's possible since we can have a fail start
3942 * from either the application manager thread or the command thread. Simply
3943 * indicate that this is a stop error.
3944 */
3945 if (!ua_sess->started) {
3946 goto error_rcu_unlock;
3947 }
3948
3949 health_code_update();
3950
3951 /* This inhibits UST tracing */
3952 ret = ustctl_stop_session(app->sock, ua_sess->handle);
3953 if (ret < 0) {
3954 if (ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
3955 ERR("Error stopping tracing for app pid: %d (ret: %d)",
3956 app->pid, ret);
3957 } else {
3958 DBG("UST app stop session failed. Application is dead.");
3959 /*
3960 * This is normal behavior, an application can die during the
3961 * creation process. Don't report an error so the execution can
3962 * continue normally.
3963 */
3964 goto end_unlock;
3965 }
3966 goto error_rcu_unlock;
3967 }
3968
3969 health_code_update();
3970
3971 /* Quiescent wait after stopping trace */
3972 ret = ustctl_wait_quiescent(app->sock);
3973 if (ret < 0 && ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
3974 ERR("UST app wait quiescent failed for app pid %d ret %d",
3975 app->pid, ret);
3976 }
3977
3978 health_code_update();
3979
3980 registry = get_session_registry(ua_sess);
3981 assert(registry);
3982
3983 if (!registry->metadata_closed) {
3984 /* Push metadata for application before freeing the application. */
3985 (void) push_metadata(registry, ua_sess->consumer);
3986 }
3987
3988 end_unlock:
3989 pthread_mutex_unlock(&ua_sess->lock);
3990 end_no_session:
3991 rcu_read_unlock();
3992 health_code_update();
3993 return 0;
3994
3995 error_rcu_unlock:
3996 pthread_mutex_unlock(&ua_sess->lock);
3997 rcu_read_unlock();
3998 health_code_update();
3999 return -1;
4000 }
4001
4002 /*
4003 * Flush buffers for a specific UST session and app.
4004 */
4005 static
4006 int ust_app_flush_trace(struct ltt_ust_session *usess, struct ust_app *app)
4007 {
4008 int ret = 0;
4009 struct lttng_ht_iter iter;
4010 struct ust_app_session *ua_sess;
4011 struct ust_app_channel *ua_chan;
4012
4013 DBG("Flushing buffers for ust app pid %d", app->pid);
4014
4015 rcu_read_lock();
4016
4017 if (!app->compatible) {
4018 goto end_no_session;
4019 }
4020
4021 ua_sess = lookup_session_by_app(usess, app);
4022 if (ua_sess == NULL) {
4023 goto end_no_session;
4024 }
4025
4026 pthread_mutex_lock(&ua_sess->lock);
4027
4028 health_code_update();
4029
4030 /* Flushing buffers */
4031 cds_lfht_for_each_entry(ua_sess->channels->ht, &iter.iter, ua_chan,
4032 node.node) {
4033 health_code_update();
4034 assert(ua_chan->is_sent);
4035 ret = ustctl_sock_flush_buffer(app->sock, ua_chan->obj);
4036 if (ret < 0) {
4037 if (ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
4038 ERR("UST app PID %d channel %s flush failed with ret %d",
4039 app->pid, ua_chan->name, ret);
4040 } else {
4041 DBG3("UST app failed to flush %s. Application is dead.",
4042 ua_chan->name);
4043 /*
4044 * This is normal behavior, an application can die during the
4045 * creation process. Don't report an error so the execution can
4046 * continue normally.
4047 */
4048 }
4049 /* Continuing flushing all buffers */
4050 continue;
4051 }
4052 }
4053
4054 health_code_update();
4055
4056 pthread_mutex_unlock(&ua_sess->lock);
4057 end_no_session:
4058 rcu_read_unlock();
4059 health_code_update();
4060 return 0;
4061 }
4062
4063 /*
4064 * Destroy a specific UST session in apps.
4065 */
4066 static int destroy_trace(struct ltt_ust_session *usess, struct ust_app *app)
4067 {
4068 int ret;
4069 struct ust_app_session *ua_sess;
4070 struct lttng_ht_iter iter;
4071 struct lttng_ht_node_u64 *node;
4072
4073 DBG("Destroy tracing for ust app pid %d", app->pid);
4074
4075 rcu_read_lock();
4076
4077 if (!app->compatible) {
4078 goto end;
4079 }
4080
4081 __lookup_session_by_app(usess, app, &iter);
4082 node = lttng_ht_iter_get_node_u64(&iter);
4083 if (node == NULL) {
4084 /* Session is being or is deleted. */
4085 goto end;
4086 }
4087 ua_sess = caa_container_of(node, struct ust_app_session, node);
4088
4089 health_code_update();
4090 destroy_app_session(app, ua_sess);
4091
4092 health_code_update();
4093
4094 /* Quiescent wait after stopping trace */
4095 ret = ustctl_wait_quiescent(app->sock);
4096 if (ret < 0 && ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
4097 ERR("UST app wait quiescent failed for app pid %d ret %d",
4098 app->pid, ret);
4099 }
4100 end:
4101 rcu_read_unlock();
4102 health_code_update();
4103 return 0;
4104 }
4105
4106 /*
4107 * Start tracing for the UST session.
4108 */
4109 int ust_app_start_trace_all(struct ltt_ust_session *usess)
4110 {
4111 int ret = 0;
4112 struct lttng_ht_iter iter;
4113 struct ust_app *app;
4114
4115 DBG("Starting all UST traces");
4116
4117 rcu_read_lock();
4118
4119 cds_lfht_for_each_entry(ust_app_ht->ht, &iter.iter, app, pid_n.node) {
4120 ret = ust_app_start_trace(usess, app);
4121 if (ret < 0) {
4122 /* Continue to next apps even on error */
4123 continue;
4124 }
4125 }
4126
4127 rcu_read_unlock();
4128
4129 return 0;
4130 }
4131
4132 /*
4133 * Start tracing for the UST session.
4134 */
4135 int ust_app_stop_trace_all(struct ltt_ust_session *usess)
4136 {
4137 int ret = 0;
4138 struct lttng_ht_iter iter;
4139 struct ust_app *app;
4140
4141 DBG("Stopping all UST traces");
4142
4143 rcu_read_lock();
4144
4145 cds_lfht_for_each_entry(ust_app_ht->ht, &iter.iter, app, pid_n.node) {
4146 ret = ust_app_stop_trace(usess, app);
4147 if (ret < 0) {
4148 /* Continue to next apps even on error */
4149 continue;
4150 }
4151 }
4152
4153 /* Flush buffers and push metadata (for UID buffers). */
4154 switch (usess->buffer_type) {
4155 case LTTNG_BUFFER_PER_UID:
4156 {
4157 struct buffer_reg_uid *reg;
4158
4159 /* Flush all per UID buffers associated to that session. */
4160 cds_list_for_each_entry(reg, &usess->buffer_reg_uid_list, lnode) {
4161 struct ust_registry_session *ust_session_reg;
4162 struct buffer_reg_channel *reg_chan;
4163 struct consumer_socket *socket;
4164
4165 /* Get consumer socket to use to push the metadata.*/
4166 socket = consumer_find_socket_by_bitness(reg->bits_per_long,
4167 usess->consumer);
4168 if (!socket) {
4169 /* Ignore request if no consumer is found for the session. */
4170 continue;
4171 }
4172
4173 cds_lfht_for_each_entry(reg->registry->channels->ht, &iter.iter,
4174 reg_chan, node.node) {
4175 /*
4176 * The following call will print error values so the return
4177 * code is of little importance because whatever happens, we
4178 * have to try them all.
4179 */
4180 (void) consumer_flush_channel(socket, reg_chan->consumer_key);
4181 }
4182
4183 ust_session_reg = reg->registry->reg.ust;
4184 if (!ust_session_reg->metadata_closed) {
4185 /* Push metadata. */
4186 (void) push_metadata(ust_session_reg, usess->consumer);
4187 }
4188 }
4189
4190 break;
4191 }
4192 case LTTNG_BUFFER_PER_PID:
4193 cds_lfht_for_each_entry(ust_app_ht->ht, &iter.iter, app, pid_n.node) {
4194 ret = ust_app_flush_trace(usess, app);
4195 if (ret < 0) {
4196 /* Continue to next apps even on error */
4197 continue;
4198 }
4199 }
4200 break;
4201 default:
4202 assert(0);
4203 break;
4204 }
4205
4206 rcu_read_unlock();
4207
4208 return 0;
4209 }
4210
4211 /*
4212 * Destroy app UST session.
4213 */
4214 int ust_app_destroy_trace_all(struct ltt_ust_session *usess)
4215 {
4216 int ret = 0;
4217 struct lttng_ht_iter iter;
4218 struct ust_app *app;
4219
4220 DBG("Destroy all UST traces");
4221
4222 rcu_read_lock();
4223
4224 cds_lfht_for_each_entry(ust_app_ht->ht, &iter.iter, app, pid_n.node) {
4225 ret = destroy_trace(usess, app);
4226 if (ret < 0) {
4227 /* Continue to next apps even on error */
4228 continue;
4229 }
4230 }
4231
4232 rcu_read_unlock();
4233
4234 return 0;
4235 }
4236
4237 /*
4238 * Add channels/events from UST global domain to registered apps at sock.
4239 */
4240 void ust_app_global_update(struct ltt_ust_session *usess, int sock)
4241 {
4242 int ret = 0;
4243 struct lttng_ht_iter iter, uiter;
4244 struct ust_app *app;
4245 struct ust_app_session *ua_sess = NULL;
4246 struct ust_app_channel *ua_chan;
4247 struct ust_app_event *ua_event;
4248 struct ust_app_ctx *ua_ctx;
4249
4250 assert(usess);
4251 assert(sock >= 0);
4252
4253 DBG2("UST app global update for app sock %d for session id %" PRIu64, sock,
4254 usess->id);
4255
4256 rcu_read_lock();
4257
4258 app = ust_app_find_by_sock(sock);
4259 if (app == NULL) {
4260 /*
4261 * Application can be unregistered before so this is possible hence
4262 * simply stopping the update.
4263 */
4264 DBG3("UST app update failed to find app sock %d", sock);
4265 goto error;
4266 }
4267
4268 if (!app->compatible) {
4269 goto error;
4270 }
4271
4272 ret = create_ust_app_session(usess, app, &ua_sess, NULL);
4273 if (ret < 0) {
4274 /* Tracer is probably gone or ENOMEM. */
4275 goto error;
4276 }
4277 assert(ua_sess);
4278
4279 pthread_mutex_lock(&ua_sess->lock);
4280
4281 /*
4282 * We can iterate safely here over all UST app session since the create ust
4283 * app session above made a shadow copy of the UST global domain from the
4284 * ltt ust session.
4285 */
4286 cds_lfht_for_each_entry(ua_sess->channels->ht, &iter.iter, ua_chan,
4287 node.node) {
4288 /*
4289 * For a metadata channel, handle it differently.
4290 */
4291 if (!strncmp(ua_chan->name, DEFAULT_METADATA_NAME,
4292 sizeof(ua_chan->name))) {
4293 ret = create_ust_app_metadata(ua_sess, app, usess->consumer,
4294 &ua_chan->attr);
4295 if (ret < 0) {
4296 goto error_unlock;
4297 }
4298 /* Remove it from the hash table and continue!. */
4299 ret = lttng_ht_del(ua_sess->channels, &iter);
4300 assert(!ret);
4301 delete_ust_app_channel(-1, ua_chan, app);
4302 continue;
4303 } else {
4304 ret = do_create_channel(app, usess, ua_sess, ua_chan);
4305 if (ret < 0) {
4306 /*
4307 * Stop everything. On error, the application failed, no more
4308 * file descriptor are available or ENOMEM so stopping here is
4309 * the only thing we can do for now.
4310 */
4311 goto error_unlock;
4312 }
4313 }
4314
4315 /*
4316 * Add context using the list so they are enabled in the same order the
4317 * user added them.
4318 */
4319 cds_list_for_each_entry(ua_ctx, &ua_chan->ctx_list, list) {
4320 ret = create_ust_channel_context(ua_chan, ua_ctx, app);
4321 if (ret < 0) {
4322 goto error_unlock;
4323 }
4324 }
4325
4326
4327 /* For each events */
4328 cds_lfht_for_each_entry(ua_chan->events->ht, &uiter.iter, ua_event,
4329 node.node) {
4330 ret = create_ust_event(app, ua_sess, ua_chan, ua_event);
4331 if (ret < 0) {
4332 goto error_unlock;
4333 }
4334 }
4335 }
4336
4337 pthread_mutex_unlock(&ua_sess->lock);
4338
4339 if (usess->start_trace) {
4340 ret = ust_app_start_trace(usess, app);
4341 if (ret < 0) {
4342 goto error;
4343 }
4344
4345 DBG2("UST trace started for app pid %d", app->pid);
4346 }
4347
4348 /* Everything went well at this point. */
4349 rcu_read_unlock();
4350 return;
4351
4352 error_unlock:
4353 pthread_mutex_unlock(&ua_sess->lock);
4354 error:
4355 if (ua_sess) {
4356 destroy_app_session(app, ua_sess);
4357 }
4358 rcu_read_unlock();
4359 return;
4360 }
4361
4362 /*
4363 * Add context to a specific channel for global UST domain.
4364 */
4365 int ust_app_add_ctx_channel_glb(struct ltt_ust_session *usess,
4366 struct ltt_ust_channel *uchan, struct ltt_ust_context *uctx)
4367 {
4368 int ret = 0;
4369 struct lttng_ht_node_str *ua_chan_node;
4370 struct lttng_ht_iter iter, uiter;
4371 struct ust_app_channel *ua_chan = NULL;
4372 struct ust_app_session *ua_sess;
4373 struct ust_app *app;
4374
4375 rcu_read_lock();
4376
4377 cds_lfht_for_each_entry(ust_app_ht->ht, &iter.iter, app, pid_n.node) {
4378 if (!app->compatible) {
4379 /*
4380 * TODO: In time, we should notice the caller of this error by
4381 * telling him that this is a version error.
4382 */
4383 continue;
4384 }
4385 ua_sess = lookup_session_by_app(usess, app);
4386 if (ua_sess == NULL) {
4387 continue;
4388 }
4389
4390 pthread_mutex_lock(&ua_sess->lock);
4391 /* Lookup channel in the ust app session */
4392 lttng_ht_lookup(ua_sess->channels, (void *)uchan->name, &uiter);
4393 ua_chan_node = lttng_ht_iter_get_node_str(&uiter);
4394 if (ua_chan_node == NULL) {
4395 goto next_app;
4396 }
4397 ua_chan = caa_container_of(ua_chan_node, struct ust_app_channel,
4398 node);
4399 ret = create_ust_app_channel_context(ua_sess, ua_chan, &uctx->ctx, app);
4400 if (ret < 0) {
4401 goto next_app;
4402 }
4403 next_app:
4404 pthread_mutex_unlock(&ua_sess->lock);
4405 }
4406
4407 rcu_read_unlock();
4408 return ret;
4409 }
4410
4411 /*
4412 * Enable event for a channel from a UST session for a specific PID.
4413 */
4414 int ust_app_enable_event_pid(struct ltt_ust_session *usess,
4415 struct ltt_ust_channel *uchan, struct ltt_ust_event *uevent, pid_t pid)
4416 {
4417 int ret = 0;
4418 struct lttng_ht_iter iter;
4419 struct lttng_ht_node_str *ua_chan_node;
4420 struct ust_app *app;
4421 struct ust_app_session *ua_sess;
4422 struct ust_app_channel *ua_chan;
4423 struct ust_app_event *ua_event;
4424
4425 DBG("UST app enabling event %s for PID %d", uevent->attr.name, pid);
4426
4427 rcu_read_lock();
4428
4429 app = ust_app_find_by_pid(pid);
4430 if (app == NULL) {
4431 ERR("UST app enable event per PID %d not found", pid);
4432 ret = -1;
4433 goto end;
4434 }
4435
4436 if (!app->compatible) {
4437 ret = 0;
4438 goto end;
4439 }
4440
4441 ua_sess = lookup_session_by_app(usess, app);
4442 if (!ua_sess) {
4443 /* The application has problem or is probably dead. */
4444 ret = 0;
4445 goto end;
4446 }
4447
4448 pthread_mutex_lock(&ua_sess->lock);
4449 /* Lookup channel in the ust app session */
4450 lttng_ht_lookup(ua_sess->channels, (void *)uchan->name, &iter);
4451 ua_chan_node = lttng_ht_iter_get_node_str(&iter);
4452 /* If the channel is not found, there is a code flow error */
4453 assert(ua_chan_node);
4454
4455 ua_chan = caa_container_of(ua_chan_node, struct ust_app_channel, node);
4456
4457 ua_event = find_ust_app_event(ua_chan->events, uevent->attr.name,
4458 uevent->filter, uevent->attr.loglevel, uevent->exclusion);
4459 if (ua_event == NULL) {
4460 ret = create_ust_app_event(ua_sess, ua_chan, uevent, app);
4461 if (ret < 0) {
4462 goto end_unlock;
4463 }
4464 } else {
4465 ret = enable_ust_app_event(ua_sess, ua_event, app);
4466 if (ret < 0) {
4467 goto end_unlock;
4468 }
4469 }
4470
4471 end_unlock:
4472 pthread_mutex_unlock(&ua_sess->lock);
4473 end:
4474 rcu_read_unlock();
4475 return ret;
4476 }
4477
4478 /*
4479 * Disable event for a channel from a UST session for a specific PID.
4480 */
4481 int ust_app_disable_event_pid(struct ltt_ust_session *usess,
4482 struct ltt_ust_channel *uchan, struct ltt_ust_event *uevent, pid_t pid)
4483 {
4484 int ret = 0;
4485 struct lttng_ht_iter iter;
4486 struct lttng_ht_node_str *ua_chan_node, *ua_event_node;
4487 struct ust_app *app;
4488 struct ust_app_session *ua_sess;
4489 struct ust_app_channel *ua_chan;
4490 struct ust_app_event *ua_event;
4491
4492 DBG("UST app disabling event %s for PID %d", uevent->attr.name, pid);
4493
4494 rcu_read_lock();
4495
4496 app = ust_app_find_by_pid(pid);
4497 if (app == NULL) {
4498 ERR("UST app disable event per PID %d not found", pid);
4499 ret = -1;
4500 goto error;
4501 }
4502
4503 if (!app->compatible) {
4504 ret = 0;
4505 goto error;
4506 }
4507
4508 ua_sess = lookup_session_by_app(usess, app);
4509 if (!ua_sess) {
4510 /* The application has problem or is probably dead. */
4511 goto error;
4512 }
4513
4514 /* Lookup channel in the ust app session */
4515 lttng_ht_lookup(ua_sess->channels, (void *)uchan->name, &iter);
4516 ua_chan_node = lttng_ht_iter_get_node_str(&iter);
4517 if (ua_chan_node == NULL) {
4518 /* Channel does not exist, skip disabling */
4519 goto error;
4520 }
4521 ua_chan = caa_container_of(ua_chan_node, struct ust_app_channel, node);
4522
4523 lttng_ht_lookup(ua_chan->events, (void *)uevent->attr.name, &iter);
4524 ua_event_node = lttng_ht_iter_get_node_str(&iter);
4525 if (ua_event_node == NULL) {
4526 /* Event does not exist, skip disabling */
4527 goto error;
4528 }
4529 ua_event = caa_container_of(ua_event_node, struct ust_app_event, node);
4530
4531 ret = disable_ust_app_event(ua_sess, ua_event, app);
4532 if (ret < 0) {
4533 goto error;
4534 }
4535
4536 error:
4537 rcu_read_unlock();
4538 return ret;
4539 }
4540
4541 /*
4542 * Calibrate registered applications.
4543 */
4544 int ust_app_calibrate_glb(struct lttng_ust_calibrate *calibrate)
4545 {
4546 int ret = 0;
4547 struct lttng_ht_iter iter;
4548 struct ust_app *app;
4549
4550 rcu_read_lock();
4551
4552 cds_lfht_for_each_entry(ust_app_ht->ht, &iter.iter, app, pid_n.node) {
4553 if (!app->compatible) {
4554 /*
4555 * TODO: In time, we should notice the caller of this error by
4556 * telling him that this is a version error.
4557 */
4558 continue;
4559 }
4560
4561 health_code_update();
4562
4563 ret = ustctl_calibrate(app->sock, calibrate);
4564 if (ret < 0) {
4565 switch (ret) {
4566 case -ENOSYS:
4567 /* Means that it's not implemented on the tracer side. */
4568 ret = 0;
4569 break;
4570 default:
4571 DBG2("Calibrate app PID %d returned with error %d",
4572 app->pid, ret);
4573 break;
4574 }
4575 }
4576 }
4577
4578 DBG("UST app global domain calibration finished");
4579
4580 rcu_read_unlock();
4581
4582 health_code_update();
4583
4584 return ret;
4585 }
4586
4587 /*
4588 * Receive registration and populate the given msg structure.
4589 *
4590 * On success return 0 else a negative value returned by the ustctl call.
4591 */
4592 int ust_app_recv_registration(int sock, struct ust_register_msg *msg)
4593 {
4594 int ret;
4595 uint32_t pid, ppid, uid, gid;
4596
4597 assert(msg);
4598
4599 ret = ustctl_recv_reg_msg(sock, &msg->type, &msg->major, &msg->minor,
4600 &pid, &ppid, &uid, &gid,
4601 &msg->bits_per_long,
4602 &msg->uint8_t_alignment,
4603 &msg->uint16_t_alignment,
4604 &msg->uint32_t_alignment,
4605 &msg->uint64_t_alignment,
4606 &msg->long_alignment,
4607 &msg->byte_order,
4608 msg->name);
4609 if (ret < 0) {
4610 switch (-ret) {
4611 case EPIPE:
4612 case ECONNRESET:
4613 case LTTNG_UST_ERR_EXITING:
4614 DBG3("UST app recv reg message failed. Application died");
4615 break;
4616 case LTTNG_UST_ERR_UNSUP_MAJOR:
4617 ERR("UST app recv reg unsupported version %d.%d. Supporting %d.%d",
4618 msg->major, msg->minor, LTTNG_UST_ABI_MAJOR_VERSION,
4619 LTTNG_UST_ABI_MINOR_VERSION);
4620 break;
4621 default:
4622 ERR("UST app recv reg message failed with ret %d", ret);
4623 break;
4624 }
4625 goto error;
4626 }
4627 msg->pid = (pid_t) pid;
4628 msg->ppid = (pid_t) ppid;
4629 msg->uid = (uid_t) uid;
4630 msg->gid = (gid_t) gid;
4631
4632 error:
4633 return ret;
4634 }
4635
4636 /*
4637 * Return a ust app channel object using the application object and the channel
4638 * object descriptor has a key. If not found, NULL is returned. A RCU read side
4639 * lock MUST be acquired before calling this function.
4640 */
4641 static struct ust_app_channel *find_channel_by_objd(struct ust_app *app,
4642 int objd)
4643 {
4644 struct lttng_ht_node_ulong *node;
4645 struct lttng_ht_iter iter;
4646 struct ust_app_channel *ua_chan = NULL;
4647
4648 assert(app);
4649
4650 lttng_ht_lookup(app->ust_objd, (void *)((unsigned long) objd), &iter);
4651 node = lttng_ht_iter_get_node_ulong(&iter);
4652 if (node == NULL) {
4653 DBG2("UST app channel find by objd %d not found", objd);
4654 goto error;
4655 }
4656
4657 ua_chan = caa_container_of(node, struct ust_app_channel, ust_objd_node);
4658
4659 error:
4660 return ua_chan;
4661 }
4662
4663 /*
4664 * Reply to a register channel notification from an application on the notify
4665 * socket. The channel metadata is also created.
4666 *
4667 * The session UST registry lock is acquired in this function.
4668 *
4669 * On success 0 is returned else a negative value.
4670 */
4671 static int reply_ust_register_channel(int sock, int sobjd, int cobjd,
4672 size_t nr_fields, struct ustctl_field *fields)
4673 {
4674 int ret, ret_code = 0;
4675 uint32_t chan_id, reg_count;
4676 uint64_t chan_reg_key;
4677 enum ustctl_channel_header type;
4678 struct ust_app *app;
4679 struct ust_app_channel *ua_chan;
4680 struct ust_app_session *ua_sess;
4681 struct ust_registry_session *registry;
4682 struct ust_registry_channel *chan_reg;
4683
4684 rcu_read_lock();
4685
4686 /* Lookup application. If not found, there is a code flow error. */
4687 app = find_app_by_notify_sock(sock);
4688 if (!app) {
4689 DBG("Application socket %d is being teardown. Abort event notify",
4690 sock);
4691 ret = 0;
4692 free(fields);
4693 goto error_rcu_unlock;
4694 }
4695
4696 /* Lookup channel by UST object descriptor. */
4697 ua_chan = find_channel_by_objd(app, cobjd);
4698 if (!ua_chan) {
4699 DBG("Application channel is being teardown. Abort event notify");
4700 ret = 0;
4701 free(fields);
4702 goto error_rcu_unlock;
4703 }
4704
4705 assert(ua_chan->session);
4706 ua_sess = ua_chan->session;
4707
4708 /* Get right session registry depending on the session buffer type. */
4709 registry = get_session_registry(ua_sess);
4710 assert(registry);
4711
4712 /* Depending on the buffer type, a different channel key is used. */
4713 if (ua_sess->buffer_type == LTTNG_BUFFER_PER_UID) {
4714 chan_reg_key = ua_chan->tracing_channel_id;
4715 } else {
4716 chan_reg_key = ua_chan->key;
4717 }
4718
4719 pthread_mutex_lock(&registry->lock);
4720
4721 chan_reg = ust_registry_channel_find(registry, chan_reg_key);
4722 assert(chan_reg);
4723
4724 if (!chan_reg->register_done) {
4725 reg_count = ust_registry_get_event_count(chan_reg);
4726 if (reg_count < 31) {
4727 type = USTCTL_CHANNEL_HEADER_COMPACT;
4728 } else {
4729 type = USTCTL_CHANNEL_HEADER_LARGE;
4730 }
4731
4732 chan_reg->nr_ctx_fields = nr_fields;
4733 chan_reg->ctx_fields = fields;
4734 chan_reg->header_type = type;
4735 } else {
4736 /* Get current already assigned values. */
4737 type = chan_reg->header_type;
4738 free(fields);
4739 /* Set to NULL so the error path does not do a double free. */
4740 fields = NULL;
4741 }
4742 /* Channel id is set during the object creation. */
4743 chan_id = chan_reg->chan_id;
4744
4745 /* Append to metadata */
4746 if (!chan_reg->metadata_dumped) {
4747 ret_code = ust_metadata_channel_statedump(registry, chan_reg);
4748 if (ret_code) {
4749 ERR("Error appending channel metadata (errno = %d)", ret_code);
4750 goto reply;
4751 }
4752 }
4753
4754 reply:
4755 DBG3("UST app replying to register channel key %" PRIu64
4756 " with id %u, type: %d, ret: %d", chan_reg_key, chan_id, type,
4757 ret_code);
4758
4759 ret = ustctl_reply_register_channel(sock, chan_id, type, ret_code);
4760 if (ret < 0) {
4761 if (ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
4762 ERR("UST app reply channel failed with ret %d", ret);
4763 } else {
4764 DBG3("UST app reply channel failed. Application died");
4765 }
4766 goto error;
4767 }
4768
4769 /* This channel registry registration is completed. */
4770 chan_reg->register_done = 1;
4771
4772 error:
4773 pthread_mutex_unlock(&registry->lock);
4774 error_rcu_unlock:
4775 rcu_read_unlock();
4776 if (ret) {
4777 free(fields);
4778 }
4779 return ret;
4780 }
4781
4782 /*
4783 * Add event to the UST channel registry. When the event is added to the
4784 * registry, the metadata is also created. Once done, this replies to the
4785 * application with the appropriate error code.
4786 *
4787 * The session UST registry lock is acquired in the function.
4788 *
4789 * On success 0 is returned else a negative value.
4790 */
4791 static int add_event_ust_registry(int sock, int sobjd, int cobjd, char *name,
4792 char *sig, size_t nr_fields, struct ustctl_field *fields, int loglevel,
4793 char *model_emf_uri)
4794 {
4795 int ret, ret_code;
4796 uint32_t event_id = 0;
4797 uint64_t chan_reg_key;
4798 struct ust_app *app;
4799 struct ust_app_channel *ua_chan;
4800 struct ust_app_session *ua_sess;
4801 struct ust_registry_session *registry;
4802
4803 rcu_read_lock();
4804
4805 /* Lookup application. If not found, there is a code flow error. */
4806 app = find_app_by_notify_sock(sock);
4807 if (!app) {
4808 DBG("Application socket %d is being teardown. Abort event notify",
4809 sock);
4810 ret = 0;
4811 free(sig);
4812 free(fields);
4813 free(model_emf_uri);
4814 goto error_rcu_unlock;
4815 }
4816
4817 /* Lookup channel by UST object descriptor. */
4818 ua_chan = find_channel_by_objd(app, cobjd);
4819 if (!ua_chan) {
4820 DBG("Application channel is being teardown. Abort event notify");
4821 ret = 0;
4822 free(sig);
4823 free(fields);
4824 free(model_emf_uri);
4825 goto error_rcu_unlock;
4826 }
4827
4828 assert(ua_chan->session);
4829 ua_sess = ua_chan->session;
4830
4831 registry = get_session_registry(ua_sess);
4832 assert(registry);
4833
4834 if (ua_sess->buffer_type == LTTNG_BUFFER_PER_UID) {
4835 chan_reg_key = ua_chan->tracing_channel_id;
4836 } else {
4837 chan_reg_key = ua_chan->key;
4838 }
4839
4840 pthread_mutex_lock(&registry->lock);
4841
4842 /*
4843 * From this point on, this call acquires the ownership of the sig, fields
4844 * and model_emf_uri meaning any free are done inside it if needed. These
4845 * three variables MUST NOT be read/write after this.
4846 */
4847 ret_code = ust_registry_create_event(registry, chan_reg_key,
4848 sobjd, cobjd, name, sig, nr_fields, fields, loglevel,
4849 model_emf_uri, ua_sess->buffer_type, &event_id,
4850 app);
4851
4852 /*
4853 * The return value is returned to ustctl so in case of an error, the
4854 * application can be notified. In case of an error, it's important not to
4855 * return a negative error or else the application will get closed.
4856 */
4857 ret = ustctl_reply_register_event(sock, event_id, ret_code);
4858 if (ret < 0) {
4859 if (ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
4860 ERR("UST app reply event failed with ret %d", ret);
4861 } else {
4862 DBG3("UST app reply event failed. Application died");
4863 }
4864 /*
4865 * No need to wipe the create event since the application socket will
4866 * get close on error hence cleaning up everything by itself.
4867 */
4868 goto error;
4869 }
4870
4871 DBG3("UST registry event %s with id %" PRId32 " added successfully",
4872 name, event_id);
4873
4874 error:
4875 pthread_mutex_unlock(&registry->lock);
4876 error_rcu_unlock:
4877 rcu_read_unlock();
4878 return ret;
4879 }
4880
4881 /*
4882 * Handle application notification through the given notify socket.
4883 *
4884 * Return 0 on success or else a negative value.
4885 */
4886 int ust_app_recv_notify(int sock)
4887 {
4888 int ret;
4889 enum ustctl_notify_cmd cmd;
4890
4891 DBG3("UST app receiving notify from sock %d", sock);
4892
4893 ret = ustctl_recv_notify(sock, &cmd);
4894 if (ret < 0) {
4895 if (ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
4896 ERR("UST app recv notify failed with ret %d", ret);
4897 } else {
4898 DBG3("UST app recv notify failed. Application died");
4899 }
4900 goto error;
4901 }
4902
4903 switch (cmd) {
4904 case USTCTL_NOTIFY_CMD_EVENT:
4905 {
4906 int sobjd, cobjd, loglevel;
4907 char name[LTTNG_UST_SYM_NAME_LEN], *sig, *model_emf_uri;
4908 size_t nr_fields;
4909 struct ustctl_field *fields;
4910
4911 DBG2("UST app ustctl register event received");
4912
4913 ret = ustctl_recv_register_event(sock, &sobjd, &cobjd, name, &loglevel,
4914 &sig, &nr_fields, &fields, &model_emf_uri);
4915 if (ret < 0) {
4916 if (ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
4917 ERR("UST app recv event failed with ret %d", ret);
4918 } else {
4919 DBG3("UST app recv event failed. Application died");
4920 }
4921 goto error;
4922 }
4923
4924 /*
4925 * Add event to the UST registry coming from the notify socket. This
4926 * call will free if needed the sig, fields and model_emf_uri. This
4927 * code path loses the ownsership of these variables and transfer them
4928 * to the this function.
4929 */
4930 ret = add_event_ust_registry(sock, sobjd, cobjd, name, sig, nr_fields,
4931 fields, loglevel, model_emf_uri);
4932 if (ret < 0) {
4933 goto error;
4934 }
4935
4936 break;
4937 }
4938 case USTCTL_NOTIFY_CMD_CHANNEL:
4939 {
4940 int sobjd, cobjd;
4941 size_t nr_fields;
4942 struct ustctl_field *fields;
4943
4944 DBG2("UST app ustctl register channel received");
4945
4946 ret = ustctl_recv_register_channel(sock, &sobjd, &cobjd, &nr_fields,
4947 &fields);
4948 if (ret < 0) {
4949 if (ret != -EPIPE && ret != -LTTNG_UST_ERR_EXITING) {
4950 ERR("UST app recv channel failed with ret %d", ret);
4951 } else {
4952 DBG3("UST app recv channel failed. Application died");
4953 }
4954 goto error;
4955 }
4956
4957 /*
4958 * The fields ownership are transfered to this function call meaning
4959 * that if needed it will be freed. After this, it's invalid to access
4960 * fields or clean it up.
4961 */
4962 ret = reply_ust_register_channel(sock, sobjd, cobjd, nr_fields,
4963 fields);
4964 if (ret < 0) {
4965 goto error;
4966 }
4967
4968 break;
4969 }
4970 default:
4971 /* Should NEVER happen. */
4972 assert(0);
4973 }
4974
4975 error:
4976 return ret;
4977 }
4978
4979 /*
4980 * Once the notify socket hangs up, this is called. First, it tries to find the
4981 * corresponding application. On failure, the call_rcu to close the socket is
4982 * executed. If an application is found, it tries to delete it from the notify
4983 * socket hash table. Whathever the result, it proceeds to the call_rcu.
4984 *
4985 * Note that an object needs to be allocated here so on ENOMEM failure, the
4986 * call RCU is not done but the rest of the cleanup is.
4987 */
4988 void ust_app_notify_sock_unregister(int sock)
4989 {
4990 int err_enomem = 0;
4991 struct lttng_ht_iter iter;
4992 struct ust_app *app;
4993 struct ust_app_notify_sock_obj *obj;
4994
4995 assert(sock >= 0);
4996
4997 rcu_read_lock();
4998
4999 obj = zmalloc(sizeof(*obj));
5000 if (!obj) {
5001 /*
5002 * An ENOMEM is kind of uncool. If this strikes we continue the
5003 * procedure but the call_rcu will not be called. In this case, we
5004 * accept the fd leak rather than possibly creating an unsynchronized
5005 * state between threads.
5006 *
5007 * TODO: The notify object should be created once the notify socket is
5008 * registered and stored independantely from the ust app object. The
5009 * tricky part is to synchronize the teardown of the application and
5010 * this notify object. Let's keep that in mind so we can avoid this
5011 * kind of shenanigans with ENOMEM in the teardown path.
5012 */
5013 err_enomem = 1;
5014 } else {
5015 obj->fd = sock;
5016 }
5017
5018 DBG("UST app notify socket unregister %d", sock);
5019
5020 /*
5021 * Lookup application by notify socket. If this fails, this means that the
5022 * hash table delete has already been done by the application
5023 * unregistration process so we can safely close the notify socket in a
5024 * call RCU.
5025 */
5026 app = find_app_by_notify_sock(sock);
5027 if (!app) {
5028 goto close_socket;
5029 }
5030
5031 iter.iter.node = &app->notify_sock_n.node;
5032
5033 /*
5034 * Whatever happens here either we fail or succeed, in both cases we have
5035 * to close the socket after a grace period to continue to the call RCU
5036 * here. If the deletion is successful, the application is not visible
5037 * anymore by other threads and is it fails it means that it was already
5038 * deleted from the hash table so either way we just have to close the
5039 * socket.
5040 */
5041 (void) lttng_ht_del(ust_app_ht_by_notify_sock, &iter);
5042
5043 close_socket:
5044 rcu_read_unlock();
5045
5046 /*
5047 * Close socket after a grace period to avoid for the socket to be reused
5048 * before the application object is freed creating potential race between
5049 * threads trying to add unique in the global hash table.
5050 */
5051 if (!err_enomem) {
5052 call_rcu(&obj->head, close_notify_sock_rcu);
5053 }
5054 }
5055
5056 /*
5057 * Destroy a ust app data structure and free its memory.
5058 */
5059 void ust_app_destroy(struct ust_app *app)
5060 {
5061 if (!app) {
5062 return;
5063 }
5064
5065 call_rcu(&app->pid_n.head, delete_ust_app_rcu);
5066 }
5067
5068 /*
5069 * Take a snapshot for a given UST session. The snapshot is sent to the given
5070 * output.
5071 *
5072 * Return 0 on success or else a negative value.
5073 */
5074 int ust_app_snapshot_record(struct ltt_ust_session *usess,
5075 struct snapshot_output *output, int wait, unsigned int nb_streams)
5076 {
5077 int ret = 0;
5078 struct lttng_ht_iter iter;
5079 struct ust_app *app;
5080 char pathname[PATH_MAX];
5081 uint64_t max_stream_size = 0;
5082
5083 assert(usess);
5084 assert(output);
5085
5086 rcu_read_lock();
5087
5088 /*
5089 * Compute the maximum size of a single stream if a max size is asked by
5090 * the caller.
5091 */
5092 if (output->max_size > 0 && nb_streams > 0) {
5093 max_stream_size = output->max_size / nb_streams;
5094 }
5095
5096 switch (usess->buffer_type) {
5097 case LTTNG_BUFFER_PER_UID:
5098 {
5099 struct buffer_reg_uid *reg;
5100
5101 cds_list_for_each_entry(reg, &usess->buffer_reg_uid_list, lnode) {
5102 struct buffer_reg_channel *reg_chan;
5103 struct consumer_socket *socket;
5104
5105 /* Get consumer socket to use to push the metadata.*/
5106 socket = consumer_find_socket_by_bitness(reg->bits_per_long,
5107 usess->consumer);
5108 if (!socket) {
5109 ret = -EINVAL;
5110 goto error;
5111 }
5112
5113 memset(pathname, 0, sizeof(pathname));
5114 ret = snprintf(pathname, sizeof(pathname),
5115 DEFAULT_UST_TRACE_DIR "/" DEFAULT_UST_TRACE_UID_PATH,
5116 reg->uid, reg->bits_per_long);
5117 if (ret < 0) {
5118 PERROR("snprintf snapshot path");
5119 goto error;
5120 }
5121
5122 /* Add the UST default trace dir to path. */
5123 cds_lfht_for_each_entry(reg->registry->channels->ht, &iter.iter,
5124 reg_chan, node.node) {
5125
5126 /*
5127 * Make sure the maximum stream size is not lower than the
5128 * subbuffer size or else it's an error since we won't be able to
5129 * snapshot anything.
5130 */
5131 if (max_stream_size &&
5132 reg_chan->subbuf_size > max_stream_size) {
5133 ret = -EINVAL;
5134 DBG3("UST app snapshot record maximum stream size %" PRIu64
5135 " is smaller than subbuffer size of %zu",
5136 max_stream_size, reg_chan->subbuf_size);
5137 goto error;
5138 }
5139 ret = consumer_snapshot_channel(socket, reg_chan->consumer_key, output, 0,
5140 usess->uid, usess->gid, pathname, wait,
5141 max_stream_size);
5142 if (ret < 0) {
5143 goto error;
5144 }
5145 }
5146 ret = consumer_snapshot_channel(socket, reg->registry->reg.ust->metadata_key, output,
5147 1, usess->uid, usess->gid, pathname, wait,
5148 max_stream_size);
5149 if (ret < 0) {
5150 goto error;
5151 }
5152 }
5153 break;
5154 }
5155 case LTTNG_BUFFER_PER_PID:
5156 {
5157 cds_lfht_for_each_entry(ust_app_ht->ht, &iter.iter, app, pid_n.node) {
5158 struct consumer_socket *socket;
5159 struct lttng_ht_iter chan_iter;
5160 struct ust_app_channel *ua_chan;
5161 struct ust_app_session *ua_sess;
5162 struct ust_registry_session *registry;
5163
5164 ua_sess = lookup_session_by_app(usess, app);
5165 if (!ua_sess) {
5166 /* Session not associated with this app. */
5167 continue;
5168 }
5169
5170 /* Get the right consumer socket for the application. */
5171 socket = consumer_find_socket_by_bitness(app->bits_per_long,
5172 output->consumer);
5173 if (!socket) {
5174 ret = -EINVAL;
5175 goto error;
5176 }
5177
5178 /* Add the UST default trace dir to path. */
5179 memset(pathname, 0, sizeof(pathname));
5180 ret = snprintf(pathname, sizeof(pathname), DEFAULT_UST_TRACE_DIR "/%s",
5181 ua_sess->path);
5182 if (ret < 0) {
5183 PERROR("snprintf snapshot path");
5184 goto error;
5185 }
5186
5187 cds_lfht_for_each_entry(ua_sess->channels->ht, &chan_iter.iter,
5188 ua_chan, node.node) {
5189 /*
5190 * Make sure the maximum stream size is not lower than the
5191 * subbuffer size or else it's an error since we won't be able to
5192 * snapshot anything.
5193 */
5194 if (max_stream_size &&
5195 ua_chan->attr.subbuf_size > max_stream_size) {
5196 ret = -EINVAL;
5197 DBG3("UST app snapshot record maximum stream size %" PRIu64
5198 " is smaller than subbuffer size of %" PRIu64,
5199 max_stream_size, ua_chan->attr.subbuf_size);
5200 goto error;
5201 }
5202
5203 ret = consumer_snapshot_channel(socket, ua_chan->key, output, 0,
5204 ua_sess->euid, ua_sess->egid, pathname, wait,
5205 max_stream_size);
5206 if (ret < 0) {
5207 goto error;
5208 }
5209 }
5210
5211 registry = get_session_registry(ua_sess);
5212 assert(registry);
5213 ret = consumer_snapshot_channel(socket, registry->metadata_key, output,
5214 1, ua_sess->euid, ua_sess->egid, pathname, wait,
5215 max_stream_size);
5216 if (ret < 0) {
5217 goto error;
5218 }
5219 }
5220 break;
5221 }
5222 default:
5223 assert(0);
5224 break;
5225 }
5226
5227 error:
5228 rcu_read_unlock();
5229 return ret;
5230 }
5231
5232 /*
5233 * Return the number of streams for a UST session.
5234 */
5235 unsigned int ust_app_get_nb_stream(struct ltt_ust_session *usess)
5236 {
5237 unsigned int ret = 0;
5238 struct ust_app *app;
5239 struct lttng_ht_iter iter;
5240
5241 assert(usess);
5242
5243 switch (usess->buffer_type) {
5244 case LTTNG_BUFFER_PER_UID:
5245 {
5246 struct buffer_reg_uid *reg;
5247
5248 cds_list_for_each_entry(reg, &usess->buffer_reg_uid_list, lnode) {
5249 struct buffer_reg_channel *reg_chan;
5250
5251 cds_lfht_for_each_entry(reg->registry->channels->ht, &iter.iter,
5252 reg_chan, node.node) {
5253 ret += reg_chan->stream_count;
5254 }
5255 }
5256 break;
5257 }
5258 case LTTNG_BUFFER_PER_PID:
5259 {
5260 rcu_read_lock();
5261 cds_lfht_for_each_entry(ust_app_ht->ht, &iter.iter, app, pid_n.node) {
5262 struct ust_app_channel *ua_chan;
5263 struct ust_app_session *ua_sess;
5264 struct lttng_ht_iter chan_iter;
5265
5266 ua_sess = lookup_session_by_app(usess, app);
5267 if (!ua_sess) {
5268 /* Session not associated with this app. */
5269 continue;
5270 }
5271
5272 cds_lfht_for_each_entry(ua_sess->channels->ht, &chan_iter.iter,
5273 ua_chan, node.node) {
5274 ret += ua_chan->streams.count;
5275 }
5276 }
5277 rcu_read_unlock();
5278 break;
5279 }
5280 default:
5281 assert(0);
5282 break;
5283 }
5284
5285 return ret;
5286 }
This page took 0.171826 seconds and 5 git commands to generate.