2 * Copyright (C) 2012 David Goulet <dgoulet@efficios.com>
3 * Copyright (C) 2016 Jérémie Galarneau <jeremie.galarneau@efficios.com>
5 * SPDX-License-Identifier: GPL-2.0-only
15 #include <urcu/list.h>
16 #include <urcu/uatomic.h>
18 #include <common/buffer-view.h>
19 #include <common/common.h>
20 #include <common/compat/string.h>
21 #include <common/defaults.h>
22 #include <common/dynamic-buffer.h>
23 #include <common/kernel-ctl/kernel-ctl.h>
24 #include <common/relayd/relayd.h>
25 #include <common/sessiond-comm/sessiond-comm.h>
26 #include <common/string-utils/string-utils.h>
27 #include <common/trace-chunk.h>
28 #include <common/utils.h>
30 #include <lttng/action/action.h>
31 #include <lttng/action/action-internal.h>
32 #include <lttng/channel-internal.h>
33 #include <lttng/channel.h>
34 #include <lttng/condition/condition.h>
35 #include <lttng/error-query-internal.h>
36 #include <lttng/event-internal.h>
37 #include <lttng/location-internal.h>
38 #include <lttng/rotate-internal.h>
39 #include <lttng/session-descriptor-internal.h>
40 #include <lttng/session-internal.h>
41 #include <lttng/trigger/trigger-internal.h>
42 #include <lttng/userspace-probe-internal.h>
44 #include "agent-thread.h"
46 #include "buffer-registry.h"
50 #include "event-notifier-error-accounting.h"
52 #include "health-sessiond.h"
53 #include "kernel-consumer.h"
55 #include "lttng-sessiond.h"
56 #include "lttng-syscall.h"
57 #include "notification-thread-commands.h"
58 #include "notification-thread.h"
60 #include "rotation-thread.h"
66 /* Sleep for 100ms between each check for the shm path's deletion. */
67 #define SESSION_DESTROY_SHM_PATH_CHECK_DELAY_US 100000
69 struct cmd_destroy_session_reply_context
{
71 bool implicit_rotation_on_destroy
;
73 * Indicates whether or not an error occurred while launching the
74 * destruction of a session.
76 enum lttng_error_code destruction_status
;
79 static enum lttng_error_code
wait_on_path(void *path
);
82 * Command completion handler that is used by the destroy command
83 * when a session that has a non-default shm_path is being destroyed.
85 * See comment in cmd_destroy_session() for the rationale.
87 static struct destroy_completion_handler
{
88 struct cmd_completion_handler handler
;
89 char shm_path
[member_sizeof(struct ltt_session
, shm_path
)];
90 } destroy_completion_handler
= {
93 .data
= destroy_completion_handler
.shm_path
98 static struct cmd_completion_handler
*current_completion_handler
;
101 * Used to keep a unique index for each relayd socket created where this value
102 * is associated with streams on the consumer so it can match the right relayd
103 * to send to. It must be accessed with the relayd_net_seq_idx_lock
106 static pthread_mutex_t relayd_net_seq_idx_lock
= PTHREAD_MUTEX_INITIALIZER
;
107 static uint64_t relayd_net_seq_idx
;
109 static int validate_ust_event_name(const char *);
110 static int cmd_enable_event_internal(struct ltt_session
*session
,
111 const struct lttng_domain
*domain
,
112 char *channel_name
, struct lttng_event
*event
,
113 char *filter_expression
,
114 struct lttng_bytecode
*filter
,
115 struct lttng_event_exclusion
*exclusion
,
117 static int cmd_enable_channel_internal(struct ltt_session
*session
,
118 const struct lttng_domain
*domain
,
119 const struct lttng_channel
*_attr
,
123 * Create a session path used by list_lttng_sessions for the case that the
124 * session consumer is on the network.
126 static int build_network_session_path(char *dst
, size_t size
,
127 struct ltt_session
*session
)
129 int ret
, kdata_port
, udata_port
;
130 struct lttng_uri
*kuri
= NULL
, *uuri
= NULL
, *uri
= NULL
;
131 char tmp_uurl
[PATH_MAX
], tmp_urls
[PATH_MAX
];
136 memset(tmp_urls
, 0, sizeof(tmp_urls
));
137 memset(tmp_uurl
, 0, sizeof(tmp_uurl
));
139 kdata_port
= udata_port
= DEFAULT_NETWORK_DATA_PORT
;
141 if (session
->kernel_session
&& session
->kernel_session
->consumer
) {
142 kuri
= &session
->kernel_session
->consumer
->dst
.net
.control
;
143 kdata_port
= session
->kernel_session
->consumer
->dst
.net
.data
.port
;
146 if (session
->ust_session
&& session
->ust_session
->consumer
) {
147 uuri
= &session
->ust_session
->consumer
->dst
.net
.control
;
148 udata_port
= session
->ust_session
->consumer
->dst
.net
.data
.port
;
151 if (uuri
== NULL
&& kuri
== NULL
) {
152 uri
= &session
->consumer
->dst
.net
.control
;
153 kdata_port
= session
->consumer
->dst
.net
.data
.port
;
154 } else if (kuri
&& uuri
) {
155 ret
= uri_compare(kuri
, uuri
);
159 /* Build uuri URL string */
160 ret
= uri_to_str_url(uuri
, tmp_uurl
, sizeof(tmp_uurl
));
167 } else if (kuri
&& uuri
== NULL
) {
169 } else if (uuri
&& kuri
== NULL
) {
173 ret
= uri_to_str_url(uri
, tmp_urls
, sizeof(tmp_urls
));
179 * Do we have a UST url set. If yes, this means we have both kernel and UST
182 if (*tmp_uurl
!= '\0') {
183 ret
= snprintf(dst
, size
, "[K]: %s [data: %d] -- [U]: %s [data: %d]",
184 tmp_urls
, kdata_port
, tmp_uurl
, udata_port
);
187 if (kuri
|| (!kuri
&& !uuri
)) {
190 /* No kernel URI, use the UST port. */
193 ret
= snprintf(dst
, size
, "%s [data: %d]", tmp_urls
, dport
);
201 * Get run-time attributes if the session has been started (discarded events,
204 static int get_kernel_runtime_stats(struct ltt_session
*session
,
205 struct ltt_kernel_channel
*kchan
, uint64_t *discarded_events
,
206 uint64_t *lost_packets
)
210 if (!session
->has_been_started
) {
212 *discarded_events
= 0;
217 ret
= consumer_get_discarded_events(session
->id
, kchan
->key
,
218 session
->kernel_session
->consumer
,
224 ret
= consumer_get_lost_packets(session
->id
, kchan
->key
,
225 session
->kernel_session
->consumer
,
236 * Get run-time attributes if the session has been started (discarded events,
239 static int get_ust_runtime_stats(struct ltt_session
*session
,
240 struct ltt_ust_channel
*uchan
, uint64_t *discarded_events
,
241 uint64_t *lost_packets
)
244 struct ltt_ust_session
*usess
;
246 if (!discarded_events
|| !lost_packets
) {
251 usess
= session
->ust_session
;
252 assert(discarded_events
);
253 assert(lost_packets
);
255 if (!usess
|| !session
->has_been_started
) {
256 *discarded_events
= 0;
262 if (usess
->buffer_type
== LTTNG_BUFFER_PER_UID
) {
263 ret
= ust_app_uid_get_channel_runtime_stats(usess
->id
,
264 &usess
->buffer_reg_uid_list
,
265 usess
->consumer
, uchan
->id
,
266 uchan
->attr
.overwrite
,
269 } else if (usess
->buffer_type
== LTTNG_BUFFER_PER_PID
) {
270 ret
= ust_app_pid_get_channel_runtime_stats(usess
,
271 uchan
, usess
->consumer
,
272 uchan
->attr
.overwrite
,
278 *discarded_events
+= uchan
->per_pid_closed_app_discarded
;
279 *lost_packets
+= uchan
->per_pid_closed_app_lost
;
281 ERR("Unsupported buffer type");
292 * Create a list of agent domain events.
294 * Return number of events in list on success or else a negative value.
296 static enum lttng_error_code
list_lttng_agent_events(
297 struct agent
*agt
, struct lttng_payload
*reply_payload
,
298 unsigned int *nb_events
)
300 enum lttng_error_code ret_code
;
302 unsigned int local_nb_events
= 0;
303 struct agent_event
*event
;
304 struct lttng_ht_iter iter
;
305 unsigned long agent_event_count
;
308 assert(reply_payload
);
310 DBG3("Listing agent events");
314 agent_event_count
= lttng_ht_get_count(agt
->events
);
315 if (agent_event_count
== 0) {
320 if (agent_event_count
> UINT_MAX
) {
321 ret_code
= LTTNG_ERR_OVERFLOW
;
325 local_nb_events
= (unsigned int) agent_event_count
;
327 cds_lfht_for_each_entry(agt
->events
->ht
, &iter
.iter
, event
, node
.node
) {
328 struct lttng_event
*tmp_event
= lttng_event_create();
331 ret_code
= LTTNG_ERR_NOMEM
;
335 if (lttng_strncpy(tmp_event
->name
, event
->name
, sizeof(tmp_event
->name
))) {
336 lttng_event_destroy(tmp_event
);
337 ret_code
= LTTNG_ERR_FATAL
;
341 tmp_event
->name
[sizeof(tmp_event
->name
) - 1] = '\0';
342 tmp_event
->enabled
= !!event
->enabled_count
;
343 tmp_event
->loglevel
= event
->loglevel_value
;
344 tmp_event
->loglevel_type
= event
->loglevel_type
;
346 ret
= lttng_event_serialize(tmp_event
, 0, NULL
,
347 event
->filter_expression
, 0, NULL
, reply_payload
);
348 lttng_event_destroy(tmp_event
);
350 ret_code
= LTTNG_ERR_FATAL
;
357 *nb_events
= local_nb_events
;
364 * Create a list of ust global domain events.
366 static enum lttng_error_code
list_lttng_ust_global_events(char *channel_name
,
367 struct ltt_ust_domain_global
*ust_global
,
368 struct lttng_payload
*reply_payload
,
369 unsigned int *nb_events
)
371 enum lttng_error_code ret_code
;
373 struct lttng_ht_iter iter
;
374 struct lttng_ht_node_str
*node
;
375 struct ltt_ust_channel
*uchan
;
376 struct ltt_ust_event
*uevent
;
377 unsigned long channel_event_count
;
378 unsigned int local_nb_events
= 0;
380 assert(reply_payload
);
383 DBG("Listing UST global events for channel %s", channel_name
);
387 lttng_ht_lookup(ust_global
->channels
, (void *) channel_name
, &iter
);
388 node
= lttng_ht_iter_get_node_str(&iter
);
390 ret_code
= LTTNG_ERR_UST_CHAN_NOT_FOUND
;
394 uchan
= caa_container_of(&node
->node
, struct ltt_ust_channel
, node
.node
);
396 channel_event_count
= lttng_ht_get_count(uchan
->events
);
397 if (channel_event_count
== 0) {
403 if (channel_event_count
> UINT_MAX
) {
404 ret_code
= LTTNG_ERR_OVERFLOW
;
408 local_nb_events
= (unsigned int) channel_event_count
;
410 DBG3("Listing UST global %d events", *nb_events
);
412 cds_lfht_for_each_entry(uchan
->events
->ht
, &iter
.iter
, uevent
, node
.node
) {
413 struct lttng_event
*tmp_event
= NULL
;
415 if (uevent
->internal
) {
416 /* This event should remain hidden from clients */
421 tmp_event
= lttng_event_create();
423 ret_code
= LTTNG_ERR_NOMEM
;
427 if (lttng_strncpy(tmp_event
->name
, uevent
->attr
.name
,
428 LTTNG_SYMBOL_NAME_LEN
)) {
429 ret_code
= LTTNG_ERR_FATAL
;
430 lttng_event_destroy(tmp_event
);
434 tmp_event
->name
[LTTNG_SYMBOL_NAME_LEN
- 1] = '\0';
435 tmp_event
->enabled
= uevent
->enabled
;
437 switch (uevent
->attr
.instrumentation
) {
438 case LTTNG_UST_ABI_TRACEPOINT
:
439 tmp_event
->type
= LTTNG_EVENT_TRACEPOINT
;
441 case LTTNG_UST_ABI_PROBE
:
442 tmp_event
->type
= LTTNG_EVENT_PROBE
;
444 case LTTNG_UST_ABI_FUNCTION
:
445 tmp_event
->type
= LTTNG_EVENT_FUNCTION
;
449 tmp_event
->loglevel
= uevent
->attr
.loglevel
;
450 switch (uevent
->attr
.loglevel_type
) {
451 case LTTNG_UST_ABI_LOGLEVEL_ALL
:
452 tmp_event
->loglevel_type
= LTTNG_EVENT_LOGLEVEL_ALL
;
454 case LTTNG_UST_ABI_LOGLEVEL_RANGE
:
455 tmp_event
->loglevel_type
= LTTNG_EVENT_LOGLEVEL_RANGE
;
457 case LTTNG_UST_ABI_LOGLEVEL_SINGLE
:
458 tmp_event
->loglevel_type
= LTTNG_EVENT_LOGLEVEL_SINGLE
;
461 if (uevent
->filter
) {
462 tmp_event
->filter
= 1;
464 if (uevent
->exclusion
) {
465 tmp_event
->exclusion
= 1;
469 * We do not care about the filter bytecode and the fd from the
470 * userspace_probe_location.
472 ret
= lttng_event_serialize(tmp_event
, uevent
->exclusion
? uevent
->exclusion
->count
: 0,
473 uevent
->exclusion
? (char **) uevent
->exclusion
->names
: NULL
,
474 uevent
->filter_expression
, 0, NULL
, reply_payload
);
475 lttng_event_destroy(tmp_event
);
477 ret_code
= LTTNG_ERR_FATAL
;
483 /* nb_events is already set at this point. */
485 *nb_events
= local_nb_events
;
492 * Fill lttng_event array of all kernel events in the channel.
494 static enum lttng_error_code
list_lttng_kernel_events(char *channel_name
,
495 struct ltt_kernel_session
*kernel_session
,
496 struct lttng_payload
*reply_payload
,
497 unsigned int *nb_events
)
499 enum lttng_error_code ret_code
;
501 struct ltt_kernel_event
*event
;
502 struct ltt_kernel_channel
*kchan
;
504 assert(reply_payload
);
506 kchan
= trace_kernel_get_channel_by_name(channel_name
, kernel_session
);
508 ret_code
= LTTNG_ERR_KERN_CHAN_NOT_FOUND
;
512 *nb_events
= kchan
->event_count
;
514 DBG("Listing events for channel %s", kchan
->channel
->name
);
516 if (*nb_events
== 0) {
521 /* Kernel channels */
522 cds_list_for_each_entry(event
, &kchan
->events_list
.head
, list
) {
523 struct lttng_event
*tmp_event
= lttng_event_create();
526 ret_code
= LTTNG_ERR_NOMEM
;
530 if (lttng_strncpy(tmp_event
->name
, event
->event
->name
, LTTNG_SYMBOL_NAME_LEN
)) {
531 lttng_event_destroy(tmp_event
);
532 ret_code
= LTTNG_ERR_FATAL
;
537 tmp_event
->name
[LTTNG_SYMBOL_NAME_LEN
- 1] = '\0';
538 tmp_event
->enabled
= event
->enabled
;
539 tmp_event
->filter
= (unsigned char) !!event
->filter_expression
;
541 switch (event
->event
->instrumentation
) {
542 case LTTNG_KERNEL_ABI_TRACEPOINT
:
543 tmp_event
->type
= LTTNG_EVENT_TRACEPOINT
;
545 case LTTNG_KERNEL_ABI_KRETPROBE
:
546 tmp_event
->type
= LTTNG_EVENT_FUNCTION
;
547 memcpy(&tmp_event
->attr
.probe
, &event
->event
->u
.kprobe
,
548 sizeof(struct lttng_kernel_abi_kprobe
));
550 case LTTNG_KERNEL_ABI_KPROBE
:
551 tmp_event
->type
= LTTNG_EVENT_PROBE
;
552 memcpy(&tmp_event
->attr
.probe
, &event
->event
->u
.kprobe
,
553 sizeof(struct lttng_kernel_abi_kprobe
));
555 case LTTNG_KERNEL_ABI_UPROBE
:
556 tmp_event
->type
= LTTNG_EVENT_USERSPACE_PROBE
;
558 case LTTNG_KERNEL_ABI_FUNCTION
:
559 tmp_event
->type
= LTTNG_EVENT_FUNCTION
;
560 memcpy(&(tmp_event
->attr
.ftrace
), &event
->event
->u
.ftrace
,
561 sizeof(struct lttng_kernel_abi_function
));
563 case LTTNG_KERNEL_ABI_NOOP
:
564 tmp_event
->type
= LTTNG_EVENT_NOOP
;
566 case LTTNG_KERNEL_ABI_SYSCALL
:
567 tmp_event
->type
= LTTNG_EVENT_SYSCALL
;
569 case LTTNG_KERNEL_ABI_ALL
:
576 if (event
->userspace_probe_location
) {
577 struct lttng_userspace_probe_location
*location_copy
=
578 lttng_userspace_probe_location_copy(
579 event
->userspace_probe_location
);
581 if (!location_copy
) {
582 lttng_event_destroy(tmp_event
);
583 ret_code
= LTTNG_ERR_NOMEM
;
587 ret
= lttng_event_set_userspace_probe_location(
588 tmp_event
, location_copy
);
590 lttng_event_destroy(tmp_event
);
591 lttng_userspace_probe_location_destroy(
593 ret_code
= LTTNG_ERR_INVALID
;
598 ret
= lttng_event_serialize(tmp_event
, 0, NULL
,
599 event
->filter_expression
, 0, NULL
, reply_payload
);
600 lttng_event_destroy(tmp_event
);
602 ret_code
= LTTNG_ERR_FATAL
;
613 * Add URI so the consumer output object. Set the correct path depending on the
614 * domain adding the default trace directory.
616 static enum lttng_error_code
add_uri_to_consumer(
617 const struct ltt_session
*session
,
618 struct consumer_output
*consumer
,
619 struct lttng_uri
*uri
, enum lttng_domain_type domain
)
622 enum lttng_error_code ret_code
= LTTNG_OK
;
626 if (consumer
== NULL
) {
627 DBG("No consumer detected. Don't add URI. Stopping.");
628 ret_code
= LTTNG_ERR_NO_CONSUMER
;
633 case LTTNG_DOMAIN_KERNEL
:
634 ret
= lttng_strncpy(consumer
->domain_subdir
,
635 DEFAULT_KERNEL_TRACE_DIR
,
636 sizeof(consumer
->domain_subdir
));
638 case LTTNG_DOMAIN_UST
:
639 ret
= lttng_strncpy(consumer
->domain_subdir
,
640 DEFAULT_UST_TRACE_DIR
,
641 sizeof(consumer
->domain_subdir
));
645 * This case is possible is we try to add the URI to the global
646 * tracing session consumer object which in this case there is
649 memset(consumer
->domain_subdir
, 0,
650 sizeof(consumer
->domain_subdir
));
654 ERR("Failed to initialize consumer output domain subdirectory");
655 ret_code
= LTTNG_ERR_FATAL
;
659 switch (uri
->dtype
) {
662 DBG2("Setting network URI to consumer");
664 if (consumer
->type
== CONSUMER_DST_NET
) {
665 if ((uri
->stype
== LTTNG_STREAM_CONTROL
&&
666 consumer
->dst
.net
.control_isset
) ||
667 (uri
->stype
== LTTNG_STREAM_DATA
&&
668 consumer
->dst
.net
.data_isset
)) {
669 ret_code
= LTTNG_ERR_URL_EXIST
;
673 memset(&consumer
->dst
, 0, sizeof(consumer
->dst
));
676 /* Set URI into consumer output object */
677 ret
= consumer_set_network_uri(session
, consumer
, uri
);
681 } else if (ret
== 1) {
683 * URI was the same in the consumer so we do not append the subdir
684 * again so to not duplicate output dir.
691 if (*uri
->dst
.path
!= '/' || strstr(uri
->dst
.path
, "../")) {
692 ret_code
= LTTNG_ERR_INVALID
;
695 DBG2("Setting trace directory path from URI to %s",
697 memset(&consumer
->dst
, 0, sizeof(consumer
->dst
));
699 ret
= lttng_strncpy(consumer
->dst
.session_root_path
,
701 sizeof(consumer
->dst
.session_root_path
));
703 ret_code
= LTTNG_ERR_FATAL
;
706 consumer
->type
= CONSUMER_DST_LOCAL
;
716 * Init tracing by creating trace directory and sending fds kernel consumer.
718 static int init_kernel_tracing(struct ltt_kernel_session
*session
)
721 struct lttng_ht_iter iter
;
722 struct consumer_socket
*socket
;
728 if (session
->consumer_fds_sent
== 0 && session
->consumer
!= NULL
) {
729 cds_lfht_for_each_entry(session
->consumer
->socks
->ht
, &iter
.iter
,
731 pthread_mutex_lock(socket
->lock
);
732 ret
= kernel_consumer_send_session(socket
, session
);
733 pthread_mutex_unlock(socket
->lock
);
735 ret
= LTTNG_ERR_KERN_CONSUMER_FAIL
;
747 * Create a socket to the relayd using the URI.
749 * On success, the relayd_sock pointer is set to the created socket.
750 * Else, it remains untouched and an LTTng error code is returned.
752 static enum lttng_error_code
create_connect_relayd(struct lttng_uri
*uri
,
753 struct lttcomm_relayd_sock
**relayd_sock
,
754 struct consumer_output
*consumer
)
757 enum lttng_error_code status
= LTTNG_OK
;
758 struct lttcomm_relayd_sock
*rsock
;
760 rsock
= lttcomm_alloc_relayd_sock(uri
, RELAYD_VERSION_COMM_MAJOR
,
761 RELAYD_VERSION_COMM_MINOR
);
763 status
= LTTNG_ERR_FATAL
;
768 * Connect to relayd so we can proceed with a session creation. This call
769 * can possibly block for an arbitrary amount of time to set the health
770 * state to be in poll execution.
773 ret
= relayd_connect(rsock
);
776 ERR("Unable to reach lttng-relayd");
777 status
= LTTNG_ERR_RELAYD_CONNECT_FAIL
;
781 /* Create socket for control stream. */
782 if (uri
->stype
== LTTNG_STREAM_CONTROL
) {
783 uint64_t result_flags
;
785 DBG3("Creating relayd stream socket from URI");
787 /* Check relayd version */
788 ret
= relayd_version_check(rsock
);
789 if (ret
== LTTNG_ERR_RELAYD_VERSION_FAIL
) {
790 status
= LTTNG_ERR_RELAYD_VERSION_FAIL
;
792 } else if (ret
< 0) {
793 ERR("Unable to reach lttng-relayd");
794 status
= LTTNG_ERR_RELAYD_CONNECT_FAIL
;
797 consumer
->relay_major_version
= rsock
->major
;
798 consumer
->relay_minor_version
= rsock
->minor
;
799 ret
= relayd_get_configuration(rsock
, 0,
802 ERR("Unable to get relayd configuration");
803 status
= LTTNG_ERR_RELAYD_CONNECT_FAIL
;
806 if (result_flags
& LTTCOMM_RELAYD_CONFIGURATION_FLAG_CLEAR_ALLOWED
) {
807 consumer
->relay_allows_clear
= true;
809 } else if (uri
->stype
== LTTNG_STREAM_DATA
) {
810 DBG3("Creating relayd data socket from URI");
812 /* Command is not valid */
813 ERR("Relayd invalid stream type: %d", uri
->stype
);
814 status
= LTTNG_ERR_INVALID
;
818 *relayd_sock
= rsock
;
823 /* The returned value is not useful since we are on an error path. */
824 (void) relayd_close(rsock
);
832 * Connect to the relayd using URI and send the socket to the right consumer.
834 * The consumer socket lock must be held by the caller.
836 * Returns LTTNG_OK on success or an LTTng error code on failure.
838 static enum lttng_error_code
send_consumer_relayd_socket(
839 unsigned int session_id
,
840 struct lttng_uri
*relayd_uri
,
841 struct consumer_output
*consumer
,
842 struct consumer_socket
*consumer_sock
,
843 const char *session_name
, const char *hostname
,
844 const char *base_path
, int session_live_timer
,
845 const uint64_t *current_chunk_id
,
846 time_t session_creation_time
,
847 bool session_name_contains_creation_time
)
850 struct lttcomm_relayd_sock
*rsock
= NULL
;
851 enum lttng_error_code status
;
853 /* Connect to relayd and make version check if uri is the control. */
854 status
= create_connect_relayd(relayd_uri
, &rsock
, consumer
);
855 if (status
!= LTTNG_OK
) {
856 goto relayd_comm_error
;
860 /* Set the network sequence index if not set. */
861 if (consumer
->net_seq_index
== (uint64_t) -1ULL) {
862 pthread_mutex_lock(&relayd_net_seq_idx_lock
);
864 * Increment net_seq_idx because we are about to transfer the
865 * new relayd socket to the consumer.
866 * Assign unique key so the consumer can match streams.
868 consumer
->net_seq_index
= ++relayd_net_seq_idx
;
869 pthread_mutex_unlock(&relayd_net_seq_idx_lock
);
872 /* Send relayd socket to consumer. */
873 ret
= consumer_send_relayd_socket(consumer_sock
, rsock
, consumer
,
874 relayd_uri
->stype
, session_id
,
875 session_name
, hostname
, base_path
,
876 session_live_timer
, current_chunk_id
,
877 session_creation_time
, session_name_contains_creation_time
);
879 status
= LTTNG_ERR_ENABLE_CONSUMER_FAIL
;
883 /* Flag that the corresponding socket was sent. */
884 if (relayd_uri
->stype
== LTTNG_STREAM_CONTROL
) {
885 consumer_sock
->control_sock_sent
= 1;
886 } else if (relayd_uri
->stype
== LTTNG_STREAM_DATA
) {
887 consumer_sock
->data_sock_sent
= 1;
891 * Close socket which was dup on the consumer side. The session daemon does
892 * NOT keep track of the relayd socket(s) once transfer to the consumer.
896 if (status
!= LTTNG_OK
) {
898 * The consumer output for this session should not be used anymore
899 * since the relayd connection failed thus making any tracing or/and
900 * streaming not usable.
902 consumer
->enabled
= 0;
904 (void) relayd_close(rsock
);
912 * Send both relayd sockets to a specific consumer and domain. This is a
913 * helper function to facilitate sending the information to the consumer for a
916 * The consumer socket lock must be held by the caller.
918 * Returns LTTNG_OK, or an LTTng error code on failure.
920 static enum lttng_error_code
send_consumer_relayd_sockets(
921 enum lttng_domain_type domain
,
922 unsigned int session_id
, struct consumer_output
*consumer
,
923 struct consumer_socket
*sock
, const char *session_name
,
924 const char *hostname
, const char *base_path
, int session_live_timer
,
925 const uint64_t *current_chunk_id
, time_t session_creation_time
,
926 bool session_name_contains_creation_time
)
928 enum lttng_error_code status
= LTTNG_OK
;
933 /* Sending control relayd socket. */
934 if (!sock
->control_sock_sent
) {
935 status
= send_consumer_relayd_socket(session_id
,
936 &consumer
->dst
.net
.control
, consumer
, sock
,
937 session_name
, hostname
, base_path
, session_live_timer
,
938 current_chunk_id
, session_creation_time
,
939 session_name_contains_creation_time
);
940 if (status
!= LTTNG_OK
) {
945 /* Sending data relayd socket. */
946 if (!sock
->data_sock_sent
) {
947 status
= send_consumer_relayd_socket(session_id
,
948 &consumer
->dst
.net
.data
, consumer
, sock
,
949 session_name
, hostname
, base_path
, session_live_timer
,
950 current_chunk_id
, session_creation_time
,
951 session_name_contains_creation_time
);
952 if (status
!= LTTNG_OK
) {
962 * Setup relayd connections for a tracing session. First creates the socket to
963 * the relayd and send them to the right domain consumer. Consumer type MUST be
966 int cmd_setup_relayd(struct ltt_session
*session
)
969 struct ltt_ust_session
*usess
;
970 struct ltt_kernel_session
*ksess
;
971 struct consumer_socket
*socket
;
972 struct lttng_ht_iter iter
;
973 LTTNG_OPTIONAL(uint64_t) current_chunk_id
= {};
977 usess
= session
->ust_session
;
978 ksess
= session
->kernel_session
;
980 DBG("Setting relayd for session %s", session
->name
);
983 if (session
->current_trace_chunk
) {
984 enum lttng_trace_chunk_status status
= lttng_trace_chunk_get_id(
985 session
->current_trace_chunk
, ¤t_chunk_id
.value
);
987 if (status
== LTTNG_TRACE_CHUNK_STATUS_OK
) {
988 current_chunk_id
.is_set
= true;
990 ERR("Failed to get current trace chunk id");
996 if (usess
&& usess
->consumer
&& usess
->consumer
->type
== CONSUMER_DST_NET
997 && usess
->consumer
->enabled
) {
998 /* For each consumer socket, send relayd sockets */
999 cds_lfht_for_each_entry(usess
->consumer
->socks
->ht
, &iter
.iter
,
1000 socket
, node
.node
) {
1001 pthread_mutex_lock(socket
->lock
);
1002 ret
= send_consumer_relayd_sockets(LTTNG_DOMAIN_UST
, session
->id
,
1003 usess
->consumer
, socket
,
1004 session
->name
, session
->hostname
,
1006 session
->live_timer
,
1007 current_chunk_id
.is_set
? ¤t_chunk_id
.value
: NULL
,
1008 session
->creation_time
,
1009 session
->name_contains_creation_time
);
1010 pthread_mutex_unlock(socket
->lock
);
1011 if (ret
!= LTTNG_OK
) {
1014 /* Session is now ready for network streaming. */
1015 session
->net_handle
= 1;
1017 session
->consumer
->relay_major_version
=
1018 usess
->consumer
->relay_major_version
;
1019 session
->consumer
->relay_minor_version
=
1020 usess
->consumer
->relay_minor_version
;
1021 session
->consumer
->relay_allows_clear
=
1022 usess
->consumer
->relay_allows_clear
;
1025 if (ksess
&& ksess
->consumer
&& ksess
->consumer
->type
== CONSUMER_DST_NET
1026 && ksess
->consumer
->enabled
) {
1027 cds_lfht_for_each_entry(ksess
->consumer
->socks
->ht
, &iter
.iter
,
1028 socket
, node
.node
) {
1029 pthread_mutex_lock(socket
->lock
);
1030 ret
= send_consumer_relayd_sockets(LTTNG_DOMAIN_KERNEL
, session
->id
,
1031 ksess
->consumer
, socket
,
1032 session
->name
, session
->hostname
,
1034 session
->live_timer
,
1035 current_chunk_id
.is_set
? ¤t_chunk_id
.value
: NULL
,
1036 session
->creation_time
,
1037 session
->name_contains_creation_time
);
1038 pthread_mutex_unlock(socket
->lock
);
1039 if (ret
!= LTTNG_OK
) {
1042 /* Session is now ready for network streaming. */
1043 session
->net_handle
= 1;
1045 session
->consumer
->relay_major_version
=
1046 ksess
->consumer
->relay_major_version
;
1047 session
->consumer
->relay_minor_version
=
1048 ksess
->consumer
->relay_minor_version
;
1049 session
->consumer
->relay_allows_clear
=
1050 ksess
->consumer
->relay_allows_clear
;
1059 * Start a kernel session by opening all necessary streams.
1061 int start_kernel_session(struct ltt_kernel_session
*ksess
)
1064 struct ltt_kernel_channel
*kchan
;
1066 /* Open kernel metadata */
1067 if (ksess
->metadata
== NULL
&& ksess
->output_traces
) {
1068 ret
= kernel_open_metadata(ksess
);
1070 ret
= LTTNG_ERR_KERN_META_FAIL
;
1075 /* Open kernel metadata stream */
1076 if (ksess
->metadata
&& ksess
->metadata_stream_fd
< 0) {
1077 ret
= kernel_open_metadata_stream(ksess
);
1079 ERR("Kernel create metadata stream failed");
1080 ret
= LTTNG_ERR_KERN_STREAM_FAIL
;
1085 /* For each channel */
1086 cds_list_for_each_entry(kchan
, &ksess
->channel_list
.head
, list
) {
1087 if (kchan
->stream_count
== 0) {
1088 ret
= kernel_open_channel_stream(kchan
);
1090 ret
= LTTNG_ERR_KERN_STREAM_FAIL
;
1093 /* Update the stream global counter */
1094 ksess
->stream_count_global
+= ret
;
1098 /* Setup kernel consumer socket and send fds to it */
1099 ret
= init_kernel_tracing(ksess
);
1101 ret
= LTTNG_ERR_KERN_START_FAIL
;
1105 /* This start the kernel tracing */
1106 ret
= kernel_start_session(ksess
);
1108 ret
= LTTNG_ERR_KERN_START_FAIL
;
1112 /* Quiescent wait after starting trace */
1113 kernel_wait_quiescent();
1123 int stop_kernel_session(struct ltt_kernel_session
*ksess
)
1125 struct ltt_kernel_channel
*kchan
;
1126 bool error_occurred
= false;
1129 if (!ksess
|| !ksess
->active
) {
1132 DBG("Stopping kernel tracing");
1134 ret
= kernel_stop_session(ksess
);
1136 ret
= LTTNG_ERR_KERN_STOP_FAIL
;
1140 kernel_wait_quiescent();
1142 /* Flush metadata after stopping (if exists) */
1143 if (ksess
->metadata_stream_fd
>= 0) {
1144 ret
= kernel_metadata_flush_buffer(ksess
->metadata_stream_fd
);
1146 ERR("Kernel metadata flush failed");
1147 error_occurred
= true;
1151 /* Flush all buffers after stopping */
1152 cds_list_for_each_entry(kchan
, &ksess
->channel_list
.head
, list
) {
1153 ret
= kernel_flush_buffer(kchan
);
1155 ERR("Kernel flush buffer error");
1156 error_occurred
= true;
1161 if (error_occurred
) {
1162 ret
= LTTNG_ERR_UNK
;
1171 * Command LTTNG_DISABLE_CHANNEL processed by the client thread.
1173 int cmd_disable_channel(struct ltt_session
*session
,
1174 enum lttng_domain_type domain
, char *channel_name
)
1177 struct ltt_ust_session
*usess
;
1179 usess
= session
->ust_session
;
1184 case LTTNG_DOMAIN_KERNEL
:
1186 ret
= channel_kernel_disable(session
->kernel_session
,
1188 if (ret
!= LTTNG_OK
) {
1192 kernel_wait_quiescent();
1195 case LTTNG_DOMAIN_UST
:
1197 struct ltt_ust_channel
*uchan
;
1198 struct lttng_ht
*chan_ht
;
1200 chan_ht
= usess
->domain_global
.channels
;
1202 uchan
= trace_ust_find_channel_by_name(chan_ht
, channel_name
);
1203 if (uchan
== NULL
) {
1204 ret
= LTTNG_ERR_UST_CHAN_NOT_FOUND
;
1208 ret
= channel_ust_disable(usess
, uchan
);
1209 if (ret
!= LTTNG_OK
) {
1215 ret
= LTTNG_ERR_UNKNOWN_DOMAIN
;
1227 * Command LTTNG_ENABLE_CHANNEL processed by the client thread.
1229 * The wpipe arguments is used as a notifier for the kernel thread.
1231 int cmd_enable_channel(struct command_ctx
*cmd_ctx
, int sock
, int wpipe
)
1235 ssize_t sock_recv_len
;
1236 struct lttng_channel
*channel
= NULL
;
1237 struct lttng_buffer_view view
;
1238 struct lttng_dynamic_buffer channel_buffer
;
1239 const struct lttng_domain command_domain
= cmd_ctx
->lsm
.domain
;
1241 lttng_dynamic_buffer_init(&channel_buffer
);
1242 channel_len
= (size_t) cmd_ctx
->lsm
.u
.channel
.length
;
1243 ret
= lttng_dynamic_buffer_set_size(&channel_buffer
, channel_len
);
1245 ret
= LTTNG_ERR_NOMEM
;
1249 sock_recv_len
= lttcomm_recv_unix_sock(sock
, channel_buffer
.data
,
1251 if (sock_recv_len
< 0 || sock_recv_len
!= channel_len
) {
1252 ERR("Failed to receive \"enable channel\" command payload");
1253 ret
= LTTNG_ERR_INVALID
;
1257 view
= lttng_buffer_view_from_dynamic_buffer(&channel_buffer
, 0, channel_len
);
1258 if (!lttng_buffer_view_is_valid(&view
)) {
1259 ret
= LTTNG_ERR_INVALID
;
1263 if (lttng_channel_create_from_buffer(&view
, &channel
) != channel_len
) {
1264 ERR("Invalid channel payload received in \"enable channel\" command");
1265 ret
= LTTNG_ERR_INVALID
;
1269 ret
= cmd_enable_channel_internal(
1270 cmd_ctx
->session
, &command_domain
, channel
, wpipe
);
1273 lttng_dynamic_buffer_reset(&channel_buffer
);
1274 lttng_channel_destroy(channel
);
1278 static int cmd_enable_channel_internal(struct ltt_session
*session
,
1279 const struct lttng_domain
*domain
,
1280 const struct lttng_channel
*_attr
,
1284 struct ltt_ust_session
*usess
= session
->ust_session
;
1285 struct lttng_ht
*chan_ht
;
1287 struct lttng_channel
*attr
= NULL
;
1293 attr
= lttng_channel_copy(_attr
);
1295 ret
= -LTTNG_ERR_NOMEM
;
1299 len
= lttng_strnlen(attr
->name
, sizeof(attr
->name
));
1301 /* Validate channel name */
1302 if (attr
->name
[0] == '.' ||
1303 memchr(attr
->name
, '/', len
) != NULL
) {
1304 ret
= LTTNG_ERR_INVALID_CHANNEL_NAME
;
1308 DBG("Enabling channel %s for session %s", attr
->name
, session
->name
);
1313 * If the session is a live session, remove the switch timer, the
1314 * live timer does the same thing but sends also synchronisation
1315 * beacons for inactive streams.
1317 if (session
->live_timer
> 0) {
1318 attr
->attr
.live_timer_interval
= session
->live_timer
;
1319 attr
->attr
.switch_timer_interval
= 0;
1322 /* Check for feature support */
1323 switch (domain
->type
) {
1324 case LTTNG_DOMAIN_KERNEL
:
1326 if (kernel_supports_ring_buffer_snapshot_sample_positions() != 1) {
1327 /* Sampling position of buffer is not supported */
1328 WARN("Kernel tracer does not support buffer monitoring. "
1329 "Setting the monitor interval timer to 0 "
1330 "(disabled) for channel '%s' of session '%s'",
1331 attr
->name
, session
->name
);
1332 lttng_channel_set_monitor_timer_interval(attr
, 0);
1336 case LTTNG_DOMAIN_UST
:
1338 case LTTNG_DOMAIN_JUL
:
1339 case LTTNG_DOMAIN_LOG4J
:
1340 case LTTNG_DOMAIN_PYTHON
:
1341 if (!agent_tracing_is_enabled()) {
1342 DBG("Attempted to enable a channel in an agent domain but the agent thread is not running");
1343 ret
= LTTNG_ERR_AGENT_TRACING_DISABLED
;
1348 ret
= LTTNG_ERR_UNKNOWN_DOMAIN
;
1352 switch (domain
->type
) {
1353 case LTTNG_DOMAIN_KERNEL
:
1355 struct ltt_kernel_channel
*kchan
;
1357 kchan
= trace_kernel_get_channel_by_name(
1358 attr
->name
, session
->kernel_session
);
1359 if (kchan
== NULL
) {
1361 * Don't try to create a channel if the session has been started at
1362 * some point in time before. The tracer does not allow it.
1364 if (session
->has_been_started
) {
1365 ret
= LTTNG_ERR_TRACE_ALREADY_STARTED
;
1369 if (session
->snapshot
.nb_output
> 0 ||
1370 session
->snapshot_mode
) {
1371 /* Enforce mmap output for snapshot sessions. */
1372 attr
->attr
.output
= LTTNG_EVENT_MMAP
;
1374 ret
= channel_kernel_create(
1375 session
->kernel_session
, attr
, wpipe
);
1376 if (attr
->name
[0] != '\0') {
1377 session
->kernel_session
->has_non_default_channel
= 1;
1380 ret
= channel_kernel_enable(session
->kernel_session
, kchan
);
1383 if (ret
!= LTTNG_OK
) {
1387 kernel_wait_quiescent();
1390 case LTTNG_DOMAIN_UST
:
1391 case LTTNG_DOMAIN_JUL
:
1392 case LTTNG_DOMAIN_LOG4J
:
1393 case LTTNG_DOMAIN_PYTHON
:
1395 struct ltt_ust_channel
*uchan
;
1400 * Current agent implementation limitations force us to allow
1401 * only one channel at once in "agent" subdomains. Each
1402 * subdomain has a default channel name which must be strictly
1405 if (domain
->type
== LTTNG_DOMAIN_JUL
) {
1406 if (strncmp(attr
->name
, DEFAULT_JUL_CHANNEL_NAME
,
1407 LTTNG_SYMBOL_NAME_LEN
)) {
1408 ret
= LTTNG_ERR_INVALID_CHANNEL_NAME
;
1411 } else if (domain
->type
== LTTNG_DOMAIN_LOG4J
) {
1412 if (strncmp(attr
->name
, DEFAULT_LOG4J_CHANNEL_NAME
,
1413 LTTNG_SYMBOL_NAME_LEN
)) {
1414 ret
= LTTNG_ERR_INVALID_CHANNEL_NAME
;
1417 } else if (domain
->type
== LTTNG_DOMAIN_PYTHON
) {
1418 if (strncmp(attr
->name
, DEFAULT_PYTHON_CHANNEL_NAME
,
1419 LTTNG_SYMBOL_NAME_LEN
)) {
1420 ret
= LTTNG_ERR_INVALID_CHANNEL_NAME
;
1425 chan_ht
= usess
->domain_global
.channels
;
1427 uchan
= trace_ust_find_channel_by_name(chan_ht
, attr
->name
);
1428 if (uchan
== NULL
) {
1430 * Don't try to create a channel if the session has been started at
1431 * some point in time before. The tracer does not allow it.
1433 if (session
->has_been_started
) {
1434 ret
= LTTNG_ERR_TRACE_ALREADY_STARTED
;
1438 ret
= channel_ust_create(usess
, attr
, domain
->buf_type
);
1439 if (attr
->name
[0] != '\0') {
1440 usess
->has_non_default_channel
= 1;
1443 ret
= channel_ust_enable(usess
, uchan
);
1448 ret
= LTTNG_ERR_UNKNOWN_DOMAIN
;
1452 if (ret
== LTTNG_OK
&& attr
->attr
.output
!= LTTNG_EVENT_MMAP
) {
1453 session
->has_non_mmap_channel
= true;
1458 lttng_channel_destroy(attr
);
1462 enum lttng_error_code
cmd_process_attr_tracker_get_tracking_policy(
1463 struct ltt_session
*session
,
1464 enum lttng_domain_type domain
,
1465 enum lttng_process_attr process_attr
,
1466 enum lttng_tracking_policy
*policy
)
1468 enum lttng_error_code ret_code
= LTTNG_OK
;
1469 const struct process_attr_tracker
*tracker
;
1472 case LTTNG_DOMAIN_KERNEL
:
1473 if (!session
->kernel_session
) {
1474 ret_code
= LTTNG_ERR_INVALID
;
1477 tracker
= kernel_get_process_attr_tracker(
1478 session
->kernel_session
, process_attr
);
1480 case LTTNG_DOMAIN_UST
:
1481 if (!session
->ust_session
) {
1482 ret_code
= LTTNG_ERR_INVALID
;
1485 tracker
= trace_ust_get_process_attr_tracker(
1486 session
->ust_session
, process_attr
);
1489 ret_code
= LTTNG_ERR_UNSUPPORTED_DOMAIN
;
1493 *policy
= process_attr_tracker_get_tracking_policy(tracker
);
1495 ret_code
= LTTNG_ERR_INVALID
;
1501 enum lttng_error_code
cmd_process_attr_tracker_set_tracking_policy(
1502 struct ltt_session
*session
,
1503 enum lttng_domain_type domain
,
1504 enum lttng_process_attr process_attr
,
1505 enum lttng_tracking_policy policy
)
1507 enum lttng_error_code ret_code
= LTTNG_OK
;
1510 case LTTNG_TRACKING_POLICY_INCLUDE_SET
:
1511 case LTTNG_TRACKING_POLICY_EXCLUDE_ALL
:
1512 case LTTNG_TRACKING_POLICY_INCLUDE_ALL
:
1515 ret_code
= LTTNG_ERR_INVALID
;
1520 case LTTNG_DOMAIN_KERNEL
:
1521 if (!session
->kernel_session
) {
1522 ret_code
= LTTNG_ERR_INVALID
;
1525 ret_code
= kernel_process_attr_tracker_set_tracking_policy(
1526 session
->kernel_session
, process_attr
, policy
);
1528 case LTTNG_DOMAIN_UST
:
1529 if (!session
->ust_session
) {
1530 ret_code
= LTTNG_ERR_INVALID
;
1533 ret_code
= trace_ust_process_attr_tracker_set_tracking_policy(
1534 session
->ust_session
, process_attr
, policy
);
1537 ret_code
= LTTNG_ERR_UNSUPPORTED_DOMAIN
;
1544 enum lttng_error_code
cmd_process_attr_tracker_inclusion_set_add_value(
1545 struct ltt_session
*session
,
1546 enum lttng_domain_type domain
,
1547 enum lttng_process_attr process_attr
,
1548 const struct process_attr_value
*value
)
1550 enum lttng_error_code ret_code
= LTTNG_OK
;
1553 case LTTNG_DOMAIN_KERNEL
:
1554 if (!session
->kernel_session
) {
1555 ret_code
= LTTNG_ERR_INVALID
;
1558 ret_code
= kernel_process_attr_tracker_inclusion_set_add_value(
1559 session
->kernel_session
, process_attr
, value
);
1561 case LTTNG_DOMAIN_UST
:
1562 if (!session
->ust_session
) {
1563 ret_code
= LTTNG_ERR_INVALID
;
1566 ret_code
= trace_ust_process_attr_tracker_inclusion_set_add_value(
1567 session
->ust_session
, process_attr
, value
);
1570 ret_code
= LTTNG_ERR_UNSUPPORTED_DOMAIN
;
1577 enum lttng_error_code
cmd_process_attr_tracker_inclusion_set_remove_value(
1578 struct ltt_session
*session
,
1579 enum lttng_domain_type domain
,
1580 enum lttng_process_attr process_attr
,
1581 const struct process_attr_value
*value
)
1583 enum lttng_error_code ret_code
= LTTNG_OK
;
1586 case LTTNG_DOMAIN_KERNEL
:
1587 if (!session
->kernel_session
) {
1588 ret_code
= LTTNG_ERR_INVALID
;
1591 ret_code
= kernel_process_attr_tracker_inclusion_set_remove_value(
1592 session
->kernel_session
, process_attr
, value
);
1594 case LTTNG_DOMAIN_UST
:
1595 if (!session
->ust_session
) {
1596 ret_code
= LTTNG_ERR_INVALID
;
1599 ret_code
= trace_ust_process_attr_tracker_inclusion_set_remove_value(
1600 session
->ust_session
, process_attr
, value
);
1603 ret_code
= LTTNG_ERR_UNSUPPORTED_DOMAIN
;
1610 enum lttng_error_code
cmd_process_attr_tracker_get_inclusion_set(
1611 struct ltt_session
*session
,
1612 enum lttng_domain_type domain
,
1613 enum lttng_process_attr process_attr
,
1614 struct lttng_process_attr_values
**values
)
1616 enum lttng_error_code ret_code
= LTTNG_OK
;
1617 const struct process_attr_tracker
*tracker
;
1618 enum process_attr_tracker_status status
;
1621 case LTTNG_DOMAIN_KERNEL
:
1622 if (!session
->kernel_session
) {
1623 ret_code
= LTTNG_ERR_INVALID
;
1626 tracker
= kernel_get_process_attr_tracker(
1627 session
->kernel_session
, process_attr
);
1629 case LTTNG_DOMAIN_UST
:
1630 if (!session
->ust_session
) {
1631 ret_code
= LTTNG_ERR_INVALID
;
1634 tracker
= trace_ust_get_process_attr_tracker(
1635 session
->ust_session
, process_attr
);
1638 ret_code
= LTTNG_ERR_UNSUPPORTED_DOMAIN
;
1643 ret_code
= LTTNG_ERR_INVALID
;
1647 status
= process_attr_tracker_get_inclusion_set(tracker
, values
);
1649 case PROCESS_ATTR_TRACKER_STATUS_OK
:
1650 ret_code
= LTTNG_OK
;
1652 case PROCESS_ATTR_TRACKER_STATUS_INVALID_TRACKING_POLICY
:
1653 ret_code
= LTTNG_ERR_PROCESS_ATTR_TRACKER_INVALID_TRACKING_POLICY
;
1655 case PROCESS_ATTR_TRACKER_STATUS_ERROR
:
1656 ret_code
= LTTNG_ERR_NOMEM
;
1659 ret_code
= LTTNG_ERR_UNK
;
1668 * Command LTTNG_DISABLE_EVENT processed by the client thread.
1670 int cmd_disable_event(struct command_ctx
*cmd_ctx
,
1671 struct lttng_event
*event
,
1672 char *filter_expression
,
1673 struct lttng_bytecode
*bytecode
,
1674 struct lttng_event_exclusion
*exclusion
)
1677 const char *event_name
;
1678 const struct ltt_session
*session
= cmd_ctx
->session
;
1679 const char *channel_name
= cmd_ctx
->lsm
.u
.disable
.channel_name
;
1680 const enum lttng_domain_type domain
= cmd_ctx
->lsm
.domain
.type
;
1682 DBG("Disable event command for event \'%s\'", event
->name
);
1685 * Filter and exclusions are simply not handled by the
1686 * disable event command at this time.
1690 (void) filter_expression
;
1693 /* Ignore the presence of filter or exclusion for the event */
1695 event
->exclusion
= 0;
1697 event_name
= event
->name
;
1699 /* Error out on unhandled search criteria */
1700 if (event
->loglevel_type
|| event
->loglevel
!= -1 || event
->enabled
1701 || event
->pid
|| event
->filter
|| event
->exclusion
) {
1702 ret
= LTTNG_ERR_UNK
;
1709 case LTTNG_DOMAIN_KERNEL
:
1711 struct ltt_kernel_channel
*kchan
;
1712 struct ltt_kernel_session
*ksess
;
1714 ksess
= session
->kernel_session
;
1717 * If a non-default channel has been created in the
1718 * session, explicitely require that -c chan_name needs
1721 if (ksess
->has_non_default_channel
&& channel_name
[0] == '\0') {
1722 ret
= LTTNG_ERR_NEED_CHANNEL_NAME
;
1726 kchan
= trace_kernel_get_channel_by_name(channel_name
, ksess
);
1727 if (kchan
== NULL
) {
1728 ret
= LTTNG_ERR_KERN_CHAN_NOT_FOUND
;
1732 switch (event
->type
) {
1733 case LTTNG_EVENT_ALL
:
1734 case LTTNG_EVENT_TRACEPOINT
:
1735 case LTTNG_EVENT_SYSCALL
:
1736 case LTTNG_EVENT_PROBE
:
1737 case LTTNG_EVENT_FUNCTION
:
1738 case LTTNG_EVENT_FUNCTION_ENTRY
:/* fall-through */
1739 if (event_name
[0] == '\0') {
1740 ret
= event_kernel_disable_event(kchan
,
1743 ret
= event_kernel_disable_event(kchan
,
1744 event_name
, event
->type
);
1746 if (ret
!= LTTNG_OK
) {
1751 ret
= LTTNG_ERR_UNK
;
1755 kernel_wait_quiescent();
1758 case LTTNG_DOMAIN_UST
:
1760 struct ltt_ust_channel
*uchan
;
1761 struct ltt_ust_session
*usess
;
1763 usess
= session
->ust_session
;
1765 if (validate_ust_event_name(event_name
)) {
1766 ret
= LTTNG_ERR_INVALID_EVENT_NAME
;
1771 * If a non-default channel has been created in the
1772 * session, explicitly require that -c chan_name needs
1775 if (usess
->has_non_default_channel
&& channel_name
[0] == '\0') {
1776 ret
= LTTNG_ERR_NEED_CHANNEL_NAME
;
1780 uchan
= trace_ust_find_channel_by_name(usess
->domain_global
.channels
,
1782 if (uchan
== NULL
) {
1783 ret
= LTTNG_ERR_UST_CHAN_NOT_FOUND
;
1787 switch (event
->type
) {
1788 case LTTNG_EVENT_ALL
:
1790 * An empty event name means that everything
1791 * should be disabled.
1793 if (event
->name
[0] == '\0') {
1794 ret
= event_ust_disable_all_tracepoints(usess
, uchan
);
1796 ret
= event_ust_disable_tracepoint(usess
, uchan
,
1799 if (ret
!= LTTNG_OK
) {
1804 ret
= LTTNG_ERR_UNK
;
1808 DBG3("Disable UST event %s in channel %s completed", event_name
,
1812 case LTTNG_DOMAIN_LOG4J
:
1813 case LTTNG_DOMAIN_JUL
:
1814 case LTTNG_DOMAIN_PYTHON
:
1817 struct ltt_ust_session
*usess
= session
->ust_session
;
1821 switch (event
->type
) {
1822 case LTTNG_EVENT_ALL
:
1825 ret
= LTTNG_ERR_UNK
;
1829 agt
= trace_ust_find_agent(usess
, domain
);
1831 ret
= -LTTNG_ERR_UST_EVENT_NOT_FOUND
;
1835 * An empty event name means that everything
1836 * should be disabled.
1838 if (event
->name
[0] == '\0') {
1839 ret
= event_agent_disable_all(usess
, agt
);
1841 ret
= event_agent_disable(usess
, agt
, event_name
);
1843 if (ret
!= LTTNG_OK
) {
1850 ret
= LTTNG_ERR_UND
;
1861 free(filter_expression
);
1866 * Command LTTNG_ADD_CONTEXT processed by the client thread.
1868 int cmd_add_context(struct command_ctx
*cmd_ctx
,
1869 const struct lttng_event_context
*event_context
, int kwpipe
)
1871 int ret
, chan_kern_created
= 0, chan_ust_created
= 0;
1872 const enum lttng_domain_type domain
= cmd_ctx
->lsm
.domain
.type
;
1873 const struct ltt_session
*session
= cmd_ctx
->session
;
1874 const char *channel_name
= cmd_ctx
->lsm
.u
.context
.channel_name
;
1877 * Don't try to add a context if the session has been started at
1878 * some point in time before. The tracer does not allow it and would
1879 * result in a corrupted trace.
1881 if (cmd_ctx
->session
->has_been_started
) {
1882 ret
= LTTNG_ERR_TRACE_ALREADY_STARTED
;
1887 case LTTNG_DOMAIN_KERNEL
:
1888 assert(session
->kernel_session
);
1890 if (session
->kernel_session
->channel_count
== 0) {
1891 /* Create default channel */
1892 ret
= channel_kernel_create(session
->kernel_session
, NULL
, kwpipe
);
1893 if (ret
!= LTTNG_OK
) {
1896 chan_kern_created
= 1;
1898 /* Add kernel context to kernel tracer */
1899 ret
= context_kernel_add(session
->kernel_session
,
1900 event_context
, channel_name
);
1901 if (ret
!= LTTNG_OK
) {
1905 case LTTNG_DOMAIN_JUL
:
1906 case LTTNG_DOMAIN_LOG4J
:
1909 * Validate channel name.
1910 * If no channel name is given and the domain is JUL or LOG4J,
1911 * set it to the appropriate domain-specific channel name. If
1912 * a name is provided but does not match the expexted channel
1913 * name, return an error.
1915 if (domain
== LTTNG_DOMAIN_JUL
&& *channel_name
&&
1916 strcmp(channel_name
,
1917 DEFAULT_JUL_CHANNEL_NAME
)) {
1918 ret
= LTTNG_ERR_UST_CHAN_NOT_FOUND
;
1920 } else if (domain
== LTTNG_DOMAIN_LOG4J
&& *channel_name
&&
1921 strcmp(channel_name
,
1922 DEFAULT_LOG4J_CHANNEL_NAME
)) {
1923 ret
= LTTNG_ERR_UST_CHAN_NOT_FOUND
;
1926 /* break is _not_ missing here. */
1928 case LTTNG_DOMAIN_UST
:
1930 struct ltt_ust_session
*usess
= session
->ust_session
;
1931 unsigned int chan_count
;
1935 chan_count
= lttng_ht_get_count(usess
->domain_global
.channels
);
1936 if (chan_count
== 0) {
1937 struct lttng_channel
*attr
;
1938 /* Create default channel */
1939 attr
= channel_new_default_attr(domain
, usess
->buffer_type
);
1941 ret
= LTTNG_ERR_FATAL
;
1945 ret
= channel_ust_create(usess
, attr
, usess
->buffer_type
);
1946 if (ret
!= LTTNG_OK
) {
1950 channel_attr_destroy(attr
);
1951 chan_ust_created
= 1;
1954 ret
= context_ust_add(usess
, domain
, event_context
,
1956 if (ret
!= LTTNG_OK
) {
1962 ret
= LTTNG_ERR_UND
;
1970 if (chan_kern_created
) {
1971 struct ltt_kernel_channel
*kchan
=
1972 trace_kernel_get_channel_by_name(DEFAULT_CHANNEL_NAME
,
1973 session
->kernel_session
);
1974 /* Created previously, this should NOT fail. */
1976 kernel_destroy_channel(kchan
);
1979 if (chan_ust_created
) {
1980 struct ltt_ust_channel
*uchan
=
1981 trace_ust_find_channel_by_name(
1982 session
->ust_session
->domain_global
.channels
,
1983 DEFAULT_CHANNEL_NAME
);
1984 /* Created previously, this should NOT fail. */
1986 /* Remove from the channel list of the session. */
1987 trace_ust_delete_channel(session
->ust_session
->domain_global
.channels
,
1989 trace_ust_destroy_channel(uchan
);
1995 static inline bool name_starts_with(const char *name
, const char *prefix
)
1997 const size_t max_cmp_len
= min(strlen(prefix
), LTTNG_SYMBOL_NAME_LEN
);
1999 return !strncmp(name
, prefix
, max_cmp_len
);
2002 /* Perform userspace-specific event name validation */
2003 static int validate_ust_event_name(const char *name
)
2013 * Check name against all internal UST event component namespaces used
2016 if (name_starts_with(name
, DEFAULT_JUL_EVENT_COMPONENT
) ||
2017 name_starts_with(name
, DEFAULT_LOG4J_EVENT_COMPONENT
) ||
2018 name_starts_with(name
, DEFAULT_PYTHON_EVENT_COMPONENT
)) {
2027 * Internal version of cmd_enable_event() with a supplemental
2028 * "internal_event" flag which is used to enable internal events which should
2029 * be hidden from clients. Such events are used in the agent implementation to
2030 * enable the events through which all "agent" events are funeled.
2032 static int _cmd_enable_event(struct ltt_session
*session
,
2033 const struct lttng_domain
*domain
,
2034 char *channel_name
, struct lttng_event
*event
,
2035 char *filter_expression
,
2036 struct lttng_bytecode
*filter
,
2037 struct lttng_event_exclusion
*exclusion
,
2038 int wpipe
, bool internal_event
)
2040 int ret
= 0, channel_created
= 0;
2041 struct lttng_channel
*attr
= NULL
;
2045 assert(channel_name
);
2047 /* If we have a filter, we must have its filter expression */
2048 assert(!(!!filter_expression
^ !!filter
));
2050 /* Normalize event name as a globbing pattern */
2051 strutils_normalize_star_glob_pattern(event
->name
);
2053 /* Normalize exclusion names as globbing patterns */
2057 for (i
= 0; i
< exclusion
->count
; i
++) {
2058 char *name
= LTTNG_EVENT_EXCLUSION_NAME_AT(exclusion
, i
);
2060 strutils_normalize_star_glob_pattern(name
);
2064 DBG("Enable event command for event \'%s\'", event
->name
);
2068 switch (domain
->type
) {
2069 case LTTNG_DOMAIN_KERNEL
:
2071 struct ltt_kernel_channel
*kchan
;
2074 * If a non-default channel has been created in the
2075 * session, explicitely require that -c chan_name needs
2078 if (session
->kernel_session
->has_non_default_channel
2079 && channel_name
[0] == '\0') {
2080 ret
= LTTNG_ERR_NEED_CHANNEL_NAME
;
2084 kchan
= trace_kernel_get_channel_by_name(channel_name
,
2085 session
->kernel_session
);
2086 if (kchan
== NULL
) {
2087 attr
= channel_new_default_attr(LTTNG_DOMAIN_KERNEL
,
2088 LTTNG_BUFFER_GLOBAL
);
2090 ret
= LTTNG_ERR_FATAL
;
2093 if (lttng_strncpy(attr
->name
, channel_name
,
2094 sizeof(attr
->name
))) {
2095 ret
= LTTNG_ERR_INVALID
;
2099 ret
= cmd_enable_channel_internal(
2100 session
, domain
, attr
, wpipe
);
2101 if (ret
!= LTTNG_OK
) {
2104 channel_created
= 1;
2107 /* Get the newly created kernel channel pointer */
2108 kchan
= trace_kernel_get_channel_by_name(channel_name
,
2109 session
->kernel_session
);
2110 if (kchan
== NULL
) {
2111 /* This sould not happen... */
2112 ret
= LTTNG_ERR_FATAL
;
2116 switch (event
->type
) {
2117 case LTTNG_EVENT_ALL
:
2119 char *filter_expression_a
= NULL
;
2120 struct lttng_bytecode
*filter_a
= NULL
;
2123 * We need to duplicate filter_expression and filter,
2124 * because ownership is passed to first enable
2127 if (filter_expression
) {
2128 filter_expression_a
= strdup(filter_expression
);
2129 if (!filter_expression_a
) {
2130 ret
= LTTNG_ERR_FATAL
;
2135 filter_a
= zmalloc(sizeof(*filter_a
) + filter
->len
);
2137 free(filter_expression_a
);
2138 ret
= LTTNG_ERR_FATAL
;
2141 memcpy(filter_a
, filter
, sizeof(*filter_a
) + filter
->len
);
2143 event
->type
= LTTNG_EVENT_TRACEPOINT
; /* Hack */
2144 ret
= event_kernel_enable_event(kchan
, event
,
2145 filter_expression
, filter
);
2146 /* We have passed ownership */
2147 filter_expression
= NULL
;
2149 if (ret
!= LTTNG_OK
) {
2150 if (channel_created
) {
2151 /* Let's not leak a useless channel. */
2152 kernel_destroy_channel(kchan
);
2154 free(filter_expression_a
);
2158 event
->type
= LTTNG_EVENT_SYSCALL
; /* Hack */
2159 ret
= event_kernel_enable_event(kchan
, event
,
2160 filter_expression_a
, filter_a
);
2161 /* We have passed ownership */
2162 filter_expression_a
= NULL
;
2164 if (ret
!= LTTNG_OK
) {
2169 case LTTNG_EVENT_PROBE
:
2170 case LTTNG_EVENT_USERSPACE_PROBE
:
2171 case LTTNG_EVENT_FUNCTION
:
2172 case LTTNG_EVENT_FUNCTION_ENTRY
:
2173 case LTTNG_EVENT_TRACEPOINT
:
2174 ret
= event_kernel_enable_event(kchan
, event
,
2175 filter_expression
, filter
);
2176 /* We have passed ownership */
2177 filter_expression
= NULL
;
2179 if (ret
!= LTTNG_OK
) {
2180 if (channel_created
) {
2181 /* Let's not leak a useless channel. */
2182 kernel_destroy_channel(kchan
);
2187 case LTTNG_EVENT_SYSCALL
:
2188 ret
= event_kernel_enable_event(kchan
, event
,
2189 filter_expression
, filter
);
2190 /* We have passed ownership */
2191 filter_expression
= NULL
;
2193 if (ret
!= LTTNG_OK
) {
2198 ret
= LTTNG_ERR_UNK
;
2202 kernel_wait_quiescent();
2205 case LTTNG_DOMAIN_UST
:
2207 struct ltt_ust_channel
*uchan
;
2208 struct ltt_ust_session
*usess
= session
->ust_session
;
2213 * If a non-default channel has been created in the
2214 * session, explicitely require that -c chan_name needs
2217 if (usess
->has_non_default_channel
&& channel_name
[0] == '\0') {
2218 ret
= LTTNG_ERR_NEED_CHANNEL_NAME
;
2222 /* Get channel from global UST domain */
2223 uchan
= trace_ust_find_channel_by_name(usess
->domain_global
.channels
,
2225 if (uchan
== NULL
) {
2226 /* Create default channel */
2227 attr
= channel_new_default_attr(LTTNG_DOMAIN_UST
,
2228 usess
->buffer_type
);
2230 ret
= LTTNG_ERR_FATAL
;
2233 if (lttng_strncpy(attr
->name
, channel_name
,
2234 sizeof(attr
->name
))) {
2235 ret
= LTTNG_ERR_INVALID
;
2239 ret
= cmd_enable_channel_internal(
2240 session
, domain
, attr
, wpipe
);
2241 if (ret
!= LTTNG_OK
) {
2245 /* Get the newly created channel reference back */
2246 uchan
= trace_ust_find_channel_by_name(
2247 usess
->domain_global
.channels
, channel_name
);
2251 if (uchan
->domain
!= LTTNG_DOMAIN_UST
&& !internal_event
) {
2253 * Don't allow users to add UST events to channels which
2254 * are assigned to a userspace subdomain (JUL, Log4J,
2257 ret
= LTTNG_ERR_INVALID_CHANNEL_DOMAIN
;
2261 if (!internal_event
) {
2263 * Ensure the event name is not reserved for internal
2266 ret
= validate_ust_event_name(event
->name
);
2268 WARN("Userspace event name %s failed validation.",
2270 ret
= LTTNG_ERR_INVALID_EVENT_NAME
;
2275 /* At this point, the session and channel exist on the tracer */
2276 ret
= event_ust_enable_tracepoint(usess
, uchan
, event
,
2277 filter_expression
, filter
, exclusion
,
2279 /* We have passed ownership */
2280 filter_expression
= NULL
;
2283 if (ret
== LTTNG_ERR_UST_EVENT_ENABLED
) {
2284 goto already_enabled
;
2285 } else if (ret
!= LTTNG_OK
) {
2290 case LTTNG_DOMAIN_LOG4J
:
2291 case LTTNG_DOMAIN_JUL
:
2292 case LTTNG_DOMAIN_PYTHON
:
2294 const char *default_event_name
, *default_chan_name
;
2296 struct lttng_event uevent
;
2297 struct lttng_domain tmp_dom
;
2298 struct ltt_ust_session
*usess
= session
->ust_session
;
2302 if (!agent_tracing_is_enabled()) {
2303 DBG("Attempted to enable an event in an agent domain but the agent thread is not running");
2304 ret
= LTTNG_ERR_AGENT_TRACING_DISABLED
;
2308 agt
= trace_ust_find_agent(usess
, domain
->type
);
2310 agt
= agent_create(domain
->type
);
2312 ret
= LTTNG_ERR_NOMEM
;
2315 agent_add(agt
, usess
->agents
);
2318 /* Create the default tracepoint. */
2319 memset(&uevent
, 0, sizeof(uevent
));
2320 uevent
.type
= LTTNG_EVENT_TRACEPOINT
;
2321 uevent
.loglevel_type
= LTTNG_EVENT_LOGLEVEL_ALL
;
2322 default_event_name
= event_get_default_agent_ust_name(
2324 if (!default_event_name
) {
2325 ret
= LTTNG_ERR_FATAL
;
2328 strncpy(uevent
.name
, default_event_name
, sizeof(uevent
.name
));
2329 uevent
.name
[sizeof(uevent
.name
) - 1] = '\0';
2332 * The domain type is changed because we are about to enable the
2333 * default channel and event for the JUL domain that are hardcoded.
2334 * This happens in the UST domain.
2336 memcpy(&tmp_dom
, domain
, sizeof(tmp_dom
));
2337 tmp_dom
.type
= LTTNG_DOMAIN_UST
;
2339 switch (domain
->type
) {
2340 case LTTNG_DOMAIN_LOG4J
:
2341 default_chan_name
= DEFAULT_LOG4J_CHANNEL_NAME
;
2343 case LTTNG_DOMAIN_JUL
:
2344 default_chan_name
= DEFAULT_JUL_CHANNEL_NAME
;
2346 case LTTNG_DOMAIN_PYTHON
:
2347 default_chan_name
= DEFAULT_PYTHON_CHANNEL_NAME
;
2350 /* The switch/case we are in makes this impossible */
2355 char *filter_expression_copy
= NULL
;
2356 struct lttng_bytecode
*filter_copy
= NULL
;
2359 const size_t filter_size
= sizeof(
2360 struct lttng_bytecode
)
2363 filter_copy
= zmalloc(filter_size
);
2365 ret
= LTTNG_ERR_NOMEM
;
2368 memcpy(filter_copy
, filter
, filter_size
);
2370 filter_expression_copy
=
2371 strdup(filter_expression
);
2372 if (!filter_expression
) {
2373 ret
= LTTNG_ERR_NOMEM
;
2376 if (!filter_expression_copy
|| !filter_copy
) {
2377 free(filter_expression_copy
);
2383 ret
= cmd_enable_event_internal(session
, &tmp_dom
,
2384 (char *) default_chan_name
,
2385 &uevent
, filter_expression_copy
,
2386 filter_copy
, NULL
, wpipe
);
2389 if (ret
== LTTNG_ERR_UST_EVENT_ENABLED
) {
2390 goto already_enabled
;
2391 } else if (ret
!= LTTNG_OK
) {
2395 /* The wild card * means that everything should be enabled. */
2396 if (strncmp(event
->name
, "*", 1) == 0 && strlen(event
->name
) == 1) {
2397 ret
= event_agent_enable_all(usess
, agt
, event
, filter
,
2400 ret
= event_agent_enable(usess
, agt
, event
, filter
,
2404 filter_expression
= NULL
;
2405 if (ret
!= LTTNG_OK
) {
2412 ret
= LTTNG_ERR_UND
;
2420 free(filter_expression
);
2423 channel_attr_destroy(attr
);
2429 * Command LTTNG_ENABLE_EVENT processed by the client thread.
2430 * We own filter, exclusion, and filter_expression.
2432 int cmd_enable_event(struct command_ctx
*cmd_ctx
,
2433 struct lttng_event
*event
,
2434 char *filter_expression
,
2435 struct lttng_event_exclusion
*exclusion
,
2436 struct lttng_bytecode
*bytecode
,
2441 * Copied to ensure proper alignment since 'lsm' is a packed structure.
2443 const struct lttng_domain command_domain
= cmd_ctx
->lsm
.domain
;
2446 * The ownership of the following parameters is transferred to
2447 * _cmd_enable_event:
2449 * - filter_expression,
2453 ret
= _cmd_enable_event(cmd_ctx
->session
,
2455 cmd_ctx
->lsm
.u
.enable
.channel_name
, event
,
2456 filter_expression
, bytecode
, exclusion
, wpipe
, false);
2457 filter_expression
= NULL
;
2464 * Enable an event which is internal to LTTng. An internal should
2465 * never be made visible to clients and are immune to checks such as
2468 static int cmd_enable_event_internal(struct ltt_session
*session
,
2469 const struct lttng_domain
*domain
,
2470 char *channel_name
, struct lttng_event
*event
,
2471 char *filter_expression
,
2472 struct lttng_bytecode
*filter
,
2473 struct lttng_event_exclusion
*exclusion
,
2476 return _cmd_enable_event(session
, domain
, channel_name
, event
,
2477 filter_expression
, filter
, exclusion
, wpipe
, true);
2481 * Command LTTNG_LIST_TRACEPOINTS processed by the client thread.
2483 enum lttng_error_code
cmd_list_tracepoints(enum lttng_domain_type domain
,
2484 struct lttng_payload
*reply_payload
)
2486 enum lttng_error_code ret_code
;
2488 ssize_t i
, nb_events
= 0;
2489 struct lttng_event
*events
= NULL
;
2490 struct lttcomm_list_command_header reply_command_header
= {};
2491 size_t reply_command_header_offset
;
2493 assert(reply_payload
);
2495 /* Reserve space for command reply header. */
2496 reply_command_header_offset
= reply_payload
->buffer
.size
;
2497 ret
= lttng_dynamic_buffer_set_size(&reply_payload
->buffer
,
2498 reply_command_header_offset
+
2499 sizeof(struct lttcomm_list_command_header
));
2501 ret_code
= LTTNG_ERR_NOMEM
;
2506 case LTTNG_DOMAIN_KERNEL
:
2507 nb_events
= kernel_list_events(&events
);
2508 if (nb_events
< 0) {
2509 ret_code
= LTTNG_ERR_KERN_LIST_FAIL
;
2513 case LTTNG_DOMAIN_UST
:
2514 nb_events
= ust_app_list_events(&events
);
2515 if (nb_events
< 0) {
2516 ret_code
= LTTNG_ERR_UST_LIST_FAIL
;
2520 case LTTNG_DOMAIN_LOG4J
:
2521 case LTTNG_DOMAIN_JUL
:
2522 case LTTNG_DOMAIN_PYTHON
:
2523 nb_events
= agent_list_events(&events
, domain
);
2524 if (nb_events
< 0) {
2525 ret_code
= LTTNG_ERR_UST_LIST_FAIL
;
2530 ret_code
= LTTNG_ERR_UND
;
2534 for (i
= 0; i
< nb_events
; i
++) {
2535 ret
= lttng_event_serialize(&events
[i
], 0, NULL
, NULL
, 0, NULL
,
2538 ret_code
= LTTNG_ERR_NOMEM
;
2543 if (nb_events
> UINT32_MAX
) {
2544 ERR("Tracepoint count would overflow the tracepoint listing command's reply");
2545 ret_code
= LTTNG_ERR_OVERFLOW
;
2549 /* Update command reply header. */
2550 reply_command_header
.count
= (uint32_t) nb_events
;
2551 memcpy(reply_payload
->buffer
.data
+ reply_command_header_offset
, &reply_command_header
,
2552 sizeof(reply_command_header
));
2554 ret_code
= LTTNG_OK
;
2561 * Command LTTNG_LIST_TRACEPOINT_FIELDS processed by the client thread.
2563 ssize_t
cmd_list_tracepoint_fields(enum lttng_domain_type domain
,
2564 struct lttng_event_field
**fields
)
2567 ssize_t nb_fields
= 0;
2570 case LTTNG_DOMAIN_UST
:
2571 nb_fields
= ust_app_list_event_fields(fields
);
2572 if (nb_fields
< 0) {
2573 ret
= LTTNG_ERR_UST_LIST_FAIL
;
2577 case LTTNG_DOMAIN_KERNEL
:
2578 default: /* fall-through */
2579 ret
= LTTNG_ERR_UND
;
2586 /* Return negative value to differentiate return code */
2590 enum lttng_error_code
cmd_list_syscalls(
2591 struct lttng_payload
*reply_payload
)
2593 enum lttng_error_code ret_code
;
2594 ssize_t nb_events
, i
;
2596 struct lttng_event
*events
= NULL
;
2597 struct lttcomm_list_command_header reply_command_header
= {};
2598 size_t reply_command_header_offset
;
2600 assert(reply_payload
);
2602 /* Reserve space for command reply header. */
2603 reply_command_header_offset
= reply_payload
->buffer
.size
;
2604 ret
= lttng_dynamic_buffer_set_size(&reply_payload
->buffer
,
2605 reply_command_header_offset
+
2606 sizeof(struct lttcomm_list_command_header
));
2608 ret_code
= LTTNG_ERR_NOMEM
;
2612 nb_events
= syscall_table_list(&events
);
2613 if (nb_events
< 0) {
2614 ret_code
= (enum lttng_error_code
) -nb_events
;
2618 for (i
= 0; i
< nb_events
; i
++) {
2619 ret
= lttng_event_serialize(&events
[i
], 0, NULL
, NULL
, 0, NULL
,
2622 ret_code
= LTTNG_ERR_NOMEM
;
2627 if (nb_events
> UINT32_MAX
) {
2628 ERR("Syscall count would overflow the syscall listing command's reply");
2629 ret_code
= LTTNG_ERR_OVERFLOW
;
2633 /* Update command reply header. */
2634 reply_command_header
.count
= (uint32_t) nb_events
;
2635 memcpy(reply_payload
->buffer
.data
+ reply_command_header_offset
, &reply_command_header
,
2636 sizeof(reply_command_header
));
2638 ret_code
= LTTNG_OK
;
2645 * Command LTTNG_START_TRACE processed by the client thread.
2647 * Called with session mutex held.
2649 int cmd_start_trace(struct ltt_session
*session
)
2651 enum lttng_error_code ret
;
2652 unsigned long nb_chan
= 0;
2653 struct ltt_kernel_session
*ksession
;
2654 struct ltt_ust_session
*usess
;
2655 const bool session_rotated_after_last_stop
=
2656 session
->rotated_after_last_stop
;
2657 const bool session_cleared_after_last_stop
=
2658 session
->cleared_after_last_stop
;
2662 /* Ease our life a bit ;) */
2663 ksession
= session
->kernel_session
;
2664 usess
= session
->ust_session
;
2666 /* Is the session already started? */
2667 if (session
->active
) {
2668 ret
= LTTNG_ERR_TRACE_ALREADY_STARTED
;
2669 /* Perform nothing */
2673 if (session
->rotation_state
== LTTNG_ROTATION_STATE_ONGOING
&&
2674 !session
->current_trace_chunk
) {
2676 * A rotation was launched while the session was stopped and
2677 * it has not been completed yet. It is not possible to start
2678 * the session since starting the session here would require a
2679 * rotation from "NULL" to a new trace chunk. That rotation
2680 * would overlap with the ongoing rotation, which is not
2683 WARN("Refusing to start session \"%s\" as a rotation launched after the last \"stop\" is still ongoing",
2685 ret
= LTTNG_ERR_ROTATION_PENDING
;
2690 * Starting a session without channel is useless since after that it's not
2691 * possible to enable channel thus inform the client.
2693 if (usess
&& usess
->domain_global
.channels
) {
2694 nb_chan
+= lttng_ht_get_count(usess
->domain_global
.channels
);
2697 nb_chan
+= ksession
->channel_count
;
2700 ret
= LTTNG_ERR_NO_CHANNEL
;
2704 session
->active
= 1;
2705 session
->rotated_after_last_stop
= false;
2706 session
->cleared_after_last_stop
= false;
2707 if (session
->output_traces
&& !session
->current_trace_chunk
) {
2708 if (!session
->has_been_started
) {
2709 struct lttng_trace_chunk
*trace_chunk
;
2711 DBG("Creating initial trace chunk of session \"%s\"",
2713 trace_chunk
= session_create_new_trace_chunk(
2714 session
, NULL
, NULL
, NULL
);
2716 ret
= LTTNG_ERR_CREATE_DIR_FAIL
;
2719 assert(!session
->current_trace_chunk
);
2720 ret
= session_set_trace_chunk(session
, trace_chunk
,
2722 lttng_trace_chunk_put(trace_chunk
);
2724 ret
= LTTNG_ERR_CREATE_TRACE_CHUNK_FAIL_CONSUMER
;
2728 DBG("Rotating session \"%s\" from its current \"NULL\" trace chunk to a new chunk",
2731 * Rotate existing streams into the new chunk.
2732 * This is a "quiet" rotation has no client has
2733 * explicitly requested this operation.
2735 * There is also no need to wait for the rotation
2736 * to complete as it will happen immediately. No data
2737 * was produced as the session was stopped, so the
2738 * rotation should happen on reception of the command.
2740 ret
= cmd_rotate_session(session
, NULL
, true,
2741 LTTNG_TRACE_CHUNK_COMMAND_TYPE_NO_OPERATION
);
2742 if (ret
!= LTTNG_OK
) {
2748 /* Kernel tracing */
2749 if (ksession
!= NULL
) {
2750 DBG("Start kernel tracing session %s", session
->name
);
2751 ret
= start_kernel_session(ksession
);
2752 if (ret
!= LTTNG_OK
) {
2757 /* Flag session that trace should start automatically */
2759 int int_ret
= ust_app_start_trace_all(usess
);
2762 ret
= LTTNG_ERR_UST_START_FAIL
;
2768 * Open a packet in every stream of the session to ensure that viewers
2769 * can correctly identify the boundaries of the periods during which
2770 * tracing was active for this session.
2772 ret
= session_open_packets(session
);
2773 if (ret
!= LTTNG_OK
) {
2778 * Clear the flag that indicates that a rotation was done while the
2779 * session was stopped.
2781 session
->rotated_after_last_stop
= false;
2783 if (session
->rotate_timer_period
) {
2784 int int_ret
= timer_session_rotation_schedule_timer_start(
2785 session
, session
->rotate_timer_period
);
2788 ERR("Failed to enable rotate timer");
2789 ret
= LTTNG_ERR_UNK
;
2797 if (ret
== LTTNG_OK
) {
2798 /* Flag this after a successful start. */
2799 session
->has_been_started
|= 1;
2801 session
->active
= 0;
2802 /* Restore initial state on error. */
2803 session
->rotated_after_last_stop
=
2804 session_rotated_after_last_stop
;
2805 session
->cleared_after_last_stop
=
2806 session_cleared_after_last_stop
;
2813 * Command LTTNG_STOP_TRACE processed by the client thread.
2815 int cmd_stop_trace(struct ltt_session
*session
)
2818 struct ltt_kernel_session
*ksession
;
2819 struct ltt_ust_session
*usess
;
2823 DBG("Begin stop session \"%s\" (id %" PRIu64
")", session
->name
, session
->id
);
2825 ksession
= session
->kernel_session
;
2826 usess
= session
->ust_session
;
2828 /* Session is not active. Skip everythong and inform the client. */
2829 if (!session
->active
) {
2830 ret
= LTTNG_ERR_TRACE_ALREADY_STOPPED
;
2834 ret
= stop_kernel_session(ksession
);
2835 if (ret
!= LTTNG_OK
) {
2839 if (usess
&& usess
->active
) {
2840 ret
= ust_app_stop_trace_all(usess
);
2842 ret
= LTTNG_ERR_UST_STOP_FAIL
;
2847 DBG("Completed stop session \"%s\" (id %" PRIu64
")", session
->name
,
2849 /* Flag inactive after a successful stop. */
2850 session
->active
= 0;
2858 * Set the base_path of the session only if subdir of a control uris is set.
2859 * Return LTTNG_OK on success, otherwise LTTNG_ERR_*.
2861 static int set_session_base_path_from_uris(struct ltt_session
*session
,
2863 struct lttng_uri
*uris
)
2868 for (i
= 0; i
< nb_uri
; i
++) {
2869 if (uris
[i
].stype
!= LTTNG_STREAM_CONTROL
||
2870 uris
[i
].subdir
[0] == '\0') {
2871 /* Not interested in these URIs */
2875 if (session
->base_path
!= NULL
) {
2876 free(session
->base_path
);
2877 session
->base_path
= NULL
;
2880 /* Set session base_path */
2881 session
->base_path
= strdup(uris
[i
].subdir
);
2882 if (!session
->base_path
) {
2883 PERROR("Failed to copy base path \"%s\" to session \"%s\"",
2884 uris
[i
].subdir
, session
->name
);
2885 ret
= LTTNG_ERR_NOMEM
;
2888 DBG2("Setting base path \"%s\" for session \"%s\"",
2889 session
->base_path
, session
->name
);
2897 * Command LTTNG_SET_CONSUMER_URI processed by the client thread.
2899 int cmd_set_consumer_uri(struct ltt_session
*session
, size_t nb_uri
,
2900 struct lttng_uri
*uris
)
2903 struct ltt_kernel_session
*ksess
= session
->kernel_session
;
2904 struct ltt_ust_session
*usess
= session
->ust_session
;
2910 /* Can't set consumer URI if the session is active. */
2911 if (session
->active
) {
2912 ret
= LTTNG_ERR_TRACE_ALREADY_STARTED
;
2917 * Set the session base path if any. This is done inside
2918 * cmd_set_consumer_uri to preserve backward compatibility of the
2919 * previous session creation api vs the session descriptor api.
2921 ret
= set_session_base_path_from_uris(session
, nb_uri
, uris
);
2922 if (ret
!= LTTNG_OK
) {
2926 /* Set the "global" consumer URIs */
2927 for (i
= 0; i
< nb_uri
; i
++) {
2928 ret
= add_uri_to_consumer(session
, session
->consumer
, &uris
[i
],
2930 if (ret
!= LTTNG_OK
) {
2935 /* Set UST session URIs */
2936 if (session
->ust_session
) {
2937 for (i
= 0; i
< nb_uri
; i
++) {
2938 ret
= add_uri_to_consumer(session
,
2939 session
->ust_session
->consumer
,
2940 &uris
[i
], LTTNG_DOMAIN_UST
);
2941 if (ret
!= LTTNG_OK
) {
2947 /* Set kernel session URIs */
2948 if (session
->kernel_session
) {
2949 for (i
= 0; i
< nb_uri
; i
++) {
2950 ret
= add_uri_to_consumer(session
,
2951 session
->kernel_session
->consumer
,
2952 &uris
[i
], LTTNG_DOMAIN_KERNEL
);
2953 if (ret
!= LTTNG_OK
) {
2960 * Make sure to set the session in output mode after we set URI since a
2961 * session can be created without URL (thus flagged in no output mode).
2963 session
->output_traces
= 1;
2965 ksess
->output_traces
= 1;
2969 usess
->output_traces
= 1;
2980 enum lttng_error_code
set_session_output_from_descriptor(
2981 struct ltt_session
*session
,
2982 const struct lttng_session_descriptor
*descriptor
)
2985 enum lttng_error_code ret_code
= LTTNG_OK
;
2986 enum lttng_session_descriptor_type session_type
=
2987 lttng_session_descriptor_get_type(descriptor
);
2988 enum lttng_session_descriptor_output_type output_type
=
2989 lttng_session_descriptor_get_output_type(descriptor
);
2990 struct lttng_uri uris
[2] = {};
2991 size_t uri_count
= 0;
2993 switch (output_type
) {
2994 case LTTNG_SESSION_DESCRIPTOR_OUTPUT_TYPE_NONE
:
2996 case LTTNG_SESSION_DESCRIPTOR_OUTPUT_TYPE_LOCAL
:
2997 lttng_session_descriptor_get_local_output_uri(descriptor
,
3001 case LTTNG_SESSION_DESCRIPTOR_OUTPUT_TYPE_NETWORK
:
3002 lttng_session_descriptor_get_network_output_uris(descriptor
,
3003 &uris
[0], &uris
[1]);
3007 ret_code
= LTTNG_ERR_INVALID
;
3011 switch (session_type
) {
3012 case LTTNG_SESSION_DESCRIPTOR_TYPE_SNAPSHOT
:
3014 struct snapshot_output
*new_output
= NULL
;
3016 new_output
= snapshot_output_alloc();
3018 ret_code
= LTTNG_ERR_NOMEM
;
3022 ret
= snapshot_output_init_with_uri(session
,
3023 DEFAULT_SNAPSHOT_MAX_SIZE
,
3024 NULL
, uris
, uri_count
, session
->consumer
,
3025 new_output
, &session
->snapshot
);
3027 ret_code
= (ret
== -ENOMEM
) ?
3028 LTTNG_ERR_NOMEM
: LTTNG_ERR_INVALID
;
3029 snapshot_output_destroy(new_output
);
3032 snapshot_add_output(&session
->snapshot
, new_output
);
3035 case LTTNG_SESSION_DESCRIPTOR_TYPE_REGULAR
:
3036 case LTTNG_SESSION_DESCRIPTOR_TYPE_LIVE
:
3038 ret_code
= cmd_set_consumer_uri(session
, uri_count
, uris
);
3042 ret_code
= LTTNG_ERR_INVALID
;
3050 enum lttng_error_code
cmd_create_session_from_descriptor(
3051 struct lttng_session_descriptor
*descriptor
,
3052 const lttng_sock_cred
*creds
,
3053 const char *home_path
)
3056 enum lttng_error_code ret_code
;
3057 const char *session_name
;
3058 struct ltt_session
*new_session
= NULL
;
3059 enum lttng_session_descriptor_status descriptor_status
;
3061 session_lock_list();
3063 if (*home_path
!= '/') {
3064 ERR("Home path provided by client is not absolute");
3065 ret_code
= LTTNG_ERR_INVALID
;
3070 descriptor_status
= lttng_session_descriptor_get_session_name(
3071 descriptor
, &session_name
);
3072 switch (descriptor_status
) {
3073 case LTTNG_SESSION_DESCRIPTOR_STATUS_OK
:
3075 case LTTNG_SESSION_DESCRIPTOR_STATUS_UNSET
:
3076 session_name
= NULL
;
3079 ret_code
= LTTNG_ERR_INVALID
;
3083 ret_code
= session_create(session_name
, creds
->uid
, creds
->gid
,
3085 if (ret_code
!= LTTNG_OK
) {
3089 if (!session_name
) {
3090 ret
= lttng_session_descriptor_set_session_name(descriptor
,
3093 ret_code
= LTTNG_ERR_SESSION_FAIL
;
3098 if (!lttng_session_descriptor_is_output_destination_initialized(
3101 * Only include the session's creation time in the output
3102 * destination if the name of the session itself was
3103 * not auto-generated.
3105 ret_code
= lttng_session_descriptor_set_default_output(
3107 session_name
? &new_session
->creation_time
: NULL
,
3109 if (ret_code
!= LTTNG_OK
) {
3113 new_session
->has_user_specified_directory
=
3114 lttng_session_descriptor_has_output_directory(
3118 switch (lttng_session_descriptor_get_type(descriptor
)) {
3119 case LTTNG_SESSION_DESCRIPTOR_TYPE_SNAPSHOT
:
3120 new_session
->snapshot_mode
= 1;
3122 case LTTNG_SESSION_DESCRIPTOR_TYPE_LIVE
:
3123 new_session
->live_timer
=
3124 lttng_session_descriptor_live_get_timer_interval(
3131 ret_code
= set_session_output_from_descriptor(new_session
, descriptor
);
3132 if (ret_code
!= LTTNG_OK
) {
3135 new_session
->consumer
->enabled
= 1;
3136 ret_code
= LTTNG_OK
;
3138 /* Release reference provided by the session_create function. */
3139 session_put(new_session
);
3140 if (ret_code
!= LTTNG_OK
&& new_session
) {
3141 /* Release the global reference on error. */
3142 session_destroy(new_session
);
3144 session_unlock_list();
3148 enum lttng_error_code
cmd_create_session(struct command_ctx
*cmd_ctx
, int sock
,
3149 struct lttng_session_descriptor
**return_descriptor
)
3152 size_t payload_size
;
3153 struct lttng_dynamic_buffer payload
;
3154 struct lttng_buffer_view home_dir_view
;
3155 struct lttng_buffer_view session_descriptor_view
;
3156 struct lttng_session_descriptor
*session_descriptor
= NULL
;
3157 enum lttng_error_code ret_code
;
3159 lttng_dynamic_buffer_init(&payload
);
3160 if (cmd_ctx
->lsm
.u
.create_session
.home_dir_size
>=
3162 ret_code
= LTTNG_ERR_INVALID
;
3165 if (cmd_ctx
->lsm
.u
.create_session
.session_descriptor_size
>
3166 LTTNG_SESSION_DESCRIPTOR_MAX_LEN
) {
3167 ret_code
= LTTNG_ERR_INVALID
;
3171 payload_size
= cmd_ctx
->lsm
.u
.create_session
.home_dir_size
+
3172 cmd_ctx
->lsm
.u
.create_session
.session_descriptor_size
;
3173 ret
= lttng_dynamic_buffer_set_size(&payload
, payload_size
);
3175 ret_code
= LTTNG_ERR_NOMEM
;
3179 ret
= lttcomm_recv_unix_sock(sock
, payload
.data
, payload
.size
);
3181 ERR("Reception of session descriptor failed, aborting.");
3182 ret_code
= LTTNG_ERR_SESSION_FAIL
;
3186 home_dir_view
= lttng_buffer_view_from_dynamic_buffer(
3189 cmd_ctx
->lsm
.u
.create_session
.home_dir_size
);
3190 if (cmd_ctx
->lsm
.u
.create_session
.home_dir_size
> 0 &&
3191 !lttng_buffer_view_is_valid(&home_dir_view
)) {
3192 ERR("Invalid payload in \"create session\" command: buffer too short to contain home directory");
3193 ret_code
= LTTNG_ERR_INVALID_PROTOCOL
;
3197 session_descriptor_view
= lttng_buffer_view_from_dynamic_buffer(
3199 cmd_ctx
->lsm
.u
.create_session
.home_dir_size
,
3200 cmd_ctx
->lsm
.u
.create_session
.session_descriptor_size
);
3201 if (!lttng_buffer_view_is_valid(&session_descriptor_view
)) {
3202 ERR("Invalid payload in \"create session\" command: buffer too short to contain session descriptor");
3203 ret_code
= LTTNG_ERR_INVALID_PROTOCOL
;
3207 ret
= lttng_session_descriptor_create_from_buffer(
3208 &session_descriptor_view
, &session_descriptor
);
3210 ERR("Failed to create session descriptor from payload of \"create session\" command");
3211 ret_code
= LTTNG_ERR_INVALID
;
3216 * Sets the descriptor's auto-generated properties (name, output) if
3219 ret_code
= cmd_create_session_from_descriptor(session_descriptor
,
3221 home_dir_view
.size
? home_dir_view
.data
: NULL
);
3222 if (ret_code
!= LTTNG_OK
) {
3226 ret_code
= LTTNG_OK
;
3227 *return_descriptor
= session_descriptor
;
3228 session_descriptor
= NULL
;
3230 lttng_dynamic_buffer_reset(&payload
);
3231 lttng_session_descriptor_destroy(session_descriptor
);
3236 void cmd_destroy_session_reply(const struct ltt_session
*session
,
3237 void *_reply_context
)
3241 const struct cmd_destroy_session_reply_context
*reply_context
=
3243 struct lttng_dynamic_buffer payload
;
3244 struct lttcomm_session_destroy_command_header cmd_header
;
3245 struct lttng_trace_archive_location
*location
= NULL
;
3246 struct lttcomm_lttng_msg llm
= {
3247 .cmd_type
= LTTNG_DESTROY_SESSION
,
3248 .ret_code
= reply_context
->destruction_status
,
3251 sizeof(struct lttcomm_session_destroy_command_header
),
3254 size_t payload_size_before_location
;
3256 lttng_dynamic_buffer_init(&payload
);
3258 ret
= lttng_dynamic_buffer_append(&payload
, &llm
, sizeof(llm
));
3260 ERR("Failed to append session destruction message");
3264 cmd_header
.rotation_state
=
3265 (int32_t) (reply_context
->implicit_rotation_on_destroy
?
3266 session
->rotation_state
:
3267 LTTNG_ROTATION_STATE_NO_ROTATION
);
3268 ret
= lttng_dynamic_buffer_append(&payload
, &cmd_header
,
3269 sizeof(cmd_header
));
3271 ERR("Failed to append session destruction command header");
3275 if (!reply_context
->implicit_rotation_on_destroy
) {
3276 DBG("No implicit rotation performed during the destruction of session \"%s\", sending reply",
3280 if (session
->rotation_state
!= LTTNG_ROTATION_STATE_COMPLETED
) {
3281 DBG("Rotation state of session \"%s\" is not \"completed\", sending session destruction reply",
3286 location
= session_get_trace_archive_location(session
);
3288 ERR("Failed to get the location of the trace archive produced during the destruction of session \"%s\"",
3293 payload_size_before_location
= payload
.size
;
3294 comm_ret
= lttng_trace_archive_location_serialize(location
,
3296 lttng_trace_archive_location_put(location
);
3298 ERR("Failed to serialize the location of the trace archive produced during the destruction of session \"%s\"",
3302 /* Update the message to indicate the location's length. */
3303 ((struct lttcomm_lttng_msg
*) payload
.data
)->data_size
=
3304 payload
.size
- payload_size_before_location
;
3306 comm_ret
= lttcomm_send_unix_sock(reply_context
->reply_sock_fd
,
3307 payload
.data
, payload
.size
);
3308 if (comm_ret
!= (ssize_t
) payload
.size
) {
3309 ERR("Failed to send result of the destruction of session \"%s\" to client",
3313 ret
= close(reply_context
->reply_sock_fd
);
3315 PERROR("Failed to close client socket in deferred session destroy reply");
3317 lttng_dynamic_buffer_reset(&payload
);
3318 free(_reply_context
);
3322 * Command LTTNG_DESTROY_SESSION processed by the client thread.
3324 * Called with session lock held.
3326 int cmd_destroy_session(struct ltt_session
*session
,
3327 struct notification_thread_handle
*notification_thread_handle
,
3331 enum lttng_error_code destruction_last_error
= LTTNG_OK
;
3332 struct cmd_destroy_session_reply_context
*reply_context
= NULL
;
3335 reply_context
= zmalloc(sizeof(*reply_context
));
3336 if (!reply_context
) {
3337 ret
= LTTNG_ERR_NOMEM
;
3340 reply_context
->reply_sock_fd
= *sock_fd
;
3346 DBG("Begin destroy session %s (id %" PRIu64
")", session
->name
,
3348 if (session
->active
) {
3349 DBG("Session \"%s\" is active, attempting to stop it before destroying it",
3351 ret
= cmd_stop_trace(session
);
3352 if (ret
!= LTTNG_OK
&& ret
!= LTTNG_ERR_TRACE_ALREADY_STOPPED
) {
3353 /* Carry on with the destruction of the session. */
3354 ERR("Failed to stop session \"%s\" as part of its destruction: %s",
3355 session
->name
, lttng_strerror(-ret
));
3356 destruction_last_error
= ret
;
3360 if (session
->rotation_schedule_timer_enabled
) {
3361 if (timer_session_rotation_schedule_timer_stop(
3363 ERR("Failed to stop the \"rotation schedule\" timer of session %s",
3365 destruction_last_error
= LTTNG_ERR_TIMER_STOP_ERROR
;
3369 if (session
->rotate_size
) {
3370 unsubscribe_session_consumed_size_rotation(session
, notification_thread_handle
);
3371 session
->rotate_size
= 0;
3374 if (session
->rotated
&& session
->current_trace_chunk
&& session
->output_traces
) {
3376 * Perform a last rotation on destruction if rotations have
3377 * occurred during the session's lifetime.
3379 ret
= cmd_rotate_session(session
, NULL
, false,
3380 LTTNG_TRACE_CHUNK_COMMAND_TYPE_MOVE_TO_COMPLETED
);
3381 if (ret
!= LTTNG_OK
) {
3382 ERR("Failed to perform an implicit rotation as part of the destruction of session \"%s\": %s",
3383 session
->name
, lttng_strerror(-ret
));
3384 destruction_last_error
= -ret
;
3386 if (reply_context
) {
3387 reply_context
->implicit_rotation_on_destroy
= true;
3389 } else if (session
->has_been_started
&& session
->current_trace_chunk
) {
3391 * The user has not triggered a session rotation. However, to
3392 * ensure all data has been consumed, the session is rotated
3393 * to a 'null' trace chunk before it is destroyed.
3395 * This is a "quiet" rotation meaning that no notification is
3396 * emitted and no renaming of the current trace chunk takes
3399 ret
= cmd_rotate_session(session
, NULL
, true,
3400 LTTNG_TRACE_CHUNK_COMMAND_TYPE_NO_OPERATION
);
3402 * Rotation operations may not be supported by the kernel
3403 * tracer. Hence, do not consider this implicit rotation as
3404 * a session destruction error. The library has already stopped
3405 * the session and waited for pending data; there is nothing
3406 * left to do but complete the destruction of the session.
3408 if (ret
!= LTTNG_OK
&&
3409 ret
!= -LTTNG_ERR_ROTATION_NOT_AVAILABLE_KERNEL
) {
3410 ERR("Failed to perform a quiet rotation as part of the destruction of session \"%s\": %s",
3411 session
->name
, lttng_strerror(ret
));
3412 destruction_last_error
= -ret
;
3416 if (session
->shm_path
[0]) {
3418 * When a session is created with an explicit shm_path,
3419 * the consumer daemon will create its shared memory files
3420 * at that location and will *not* unlink them. This is normal
3421 * as the intention of that feature is to make it possible
3422 * to retrieve the content of those files should a crash occur.
3424 * To ensure the content of those files can be used, the
3425 * sessiond daemon will replicate the content of the metadata
3426 * cache in a metadata file.
3428 * On clean-up, it is expected that the consumer daemon will
3429 * unlink the shared memory files and that the session daemon
3430 * will unlink the metadata file. Then, the session's directory
3431 * in the shm path can be removed.
3433 * Unfortunately, a flaw in the design of the sessiond's and
3434 * consumerd's tear down of channels makes it impossible to
3435 * determine when the sessiond _and_ the consumerd have both
3436 * destroyed their representation of a channel. For one, the
3437 * unlinking, close, and rmdir happen in deferred 'call_rcu'
3438 * callbacks in both daemons.
3440 * However, it is also impossible for the sessiond to know when
3441 * the consumer daemon is done destroying its channel(s) since
3442 * it occurs as a reaction to the closing of the channel's file
3443 * descriptor. There is no resulting communication initiated
3444 * from the consumerd to the sessiond to confirm that the
3445 * operation is completed (and was successful).
3447 * Until this is all fixed, the session daemon checks for the
3448 * removal of the session's shm path which makes it possible
3449 * to safely advertise a session as having been destroyed.
3451 * Prior to this fix, it was not possible to reliably save
3452 * a session making use of the --shm-path option, destroy it,
3453 * and load it again. This is because the creation of the
3454 * session would fail upon seeing the session's shm path
3455 * already in existence.
3457 * Note that none of the error paths in the check for the
3458 * directory's existence return an error. This is normal
3459 * as there isn't much that can be done. The session will
3460 * be destroyed properly, except that we can't offer the
3461 * guarantee that the same session can be re-created.
3463 current_completion_handler
= &destroy_completion_handler
.handler
;
3464 ret
= lttng_strncpy(destroy_completion_handler
.shm_path
,
3466 sizeof(destroy_completion_handler
.shm_path
));
3471 * The session is destroyed. However, note that the command context
3472 * still holds a reference to the session, thus delaying its destruction
3473 * _at least_ up to the point when that reference is released.
3475 session_destroy(session
);
3476 if (reply_context
) {
3477 reply_context
->destruction_status
= destruction_last_error
;
3478 ret
= session_add_destroy_notifier(session
,
3479 cmd_destroy_session_reply
,
3480 (void *) reply_context
);
3482 ret
= LTTNG_ERR_FATAL
;
3494 * Command LTTNG_REGISTER_CONSUMER processed by the client thread.
3496 int cmd_register_consumer(struct ltt_session
*session
,
3497 enum lttng_domain_type domain
, const char *sock_path
,
3498 struct consumer_data
*cdata
)
3501 struct consumer_socket
*socket
= NULL
;
3508 case LTTNG_DOMAIN_KERNEL
:
3510 struct ltt_kernel_session
*ksess
= session
->kernel_session
;
3514 /* Can't register a consumer if there is already one */
3515 if (ksess
->consumer_fds_sent
!= 0) {
3516 ret
= LTTNG_ERR_KERN_CONSUMER_FAIL
;
3520 sock
= lttcomm_connect_unix_sock(sock_path
);
3522 ret
= LTTNG_ERR_CONNECT_FAIL
;
3525 cdata
->cmd_sock
= sock
;
3527 socket
= consumer_allocate_socket(&cdata
->cmd_sock
);
3528 if (socket
== NULL
) {
3531 PERROR("close register consumer");
3533 cdata
->cmd_sock
= -1;
3534 ret
= LTTNG_ERR_FATAL
;
3538 socket
->lock
= zmalloc(sizeof(pthread_mutex_t
));
3539 if (socket
->lock
== NULL
) {
3540 PERROR("zmalloc pthread mutex");
3541 ret
= LTTNG_ERR_FATAL
;
3544 pthread_mutex_init(socket
->lock
, NULL
);
3545 socket
->registered
= 1;
3548 consumer_add_socket(socket
, ksess
->consumer
);
3551 pthread_mutex_lock(&cdata
->pid_mutex
);
3553 pthread_mutex_unlock(&cdata
->pid_mutex
);
3558 /* TODO: Userspace tracing */
3559 ret
= LTTNG_ERR_UND
;
3567 consumer_destroy_socket(socket
);
3573 * Command LTTNG_LIST_DOMAINS processed by the client thread.
3575 ssize_t
cmd_list_domains(struct ltt_session
*session
,
3576 struct lttng_domain
**domains
)
3581 struct lttng_ht_iter iter
;
3583 if (session
->kernel_session
!= NULL
) {
3584 DBG3("Listing domains found kernel domain");
3588 if (session
->ust_session
!= NULL
) {
3589 DBG3("Listing domains found UST global domain");
3593 cds_lfht_for_each_entry(session
->ust_session
->agents
->ht
, &iter
.iter
,
3595 if (agt
->being_used
) {
3606 *domains
= zmalloc(nb_dom
* sizeof(struct lttng_domain
));
3607 if (*domains
== NULL
) {
3608 ret
= LTTNG_ERR_FATAL
;
3612 if (session
->kernel_session
!= NULL
) {
3613 (*domains
)[index
].type
= LTTNG_DOMAIN_KERNEL
;
3615 /* Kernel session buffer type is always GLOBAL */
3616 (*domains
)[index
].buf_type
= LTTNG_BUFFER_GLOBAL
;
3621 if (session
->ust_session
!= NULL
) {
3622 (*domains
)[index
].type
= LTTNG_DOMAIN_UST
;
3623 (*domains
)[index
].buf_type
= session
->ust_session
->buffer_type
;
3627 cds_lfht_for_each_entry(session
->ust_session
->agents
->ht
, &iter
.iter
,
3629 if (agt
->being_used
) {
3630 (*domains
)[index
].type
= agt
->domain
;
3631 (*domains
)[index
].buf_type
= session
->ust_session
->buffer_type
;
3641 /* Return negative value to differentiate return code */
3647 * Command LTTNG_LIST_CHANNELS processed by the client thread.
3649 enum lttng_error_code
cmd_list_channels(enum lttng_domain_type domain
,
3650 struct ltt_session
*session
,
3651 struct lttng_payload
*payload
)
3655 struct lttcomm_list_command_header cmd_header
= {};
3656 size_t cmd_header_offset
;
3657 enum lttng_error_code ret_code
;
3662 DBG("Listing channels for session %s", session
->name
);
3664 cmd_header_offset
= payload
->buffer
.size
;
3666 /* Reserve space for command reply header. */
3667 ret
= lttng_dynamic_buffer_set_size(&payload
->buffer
,
3668 cmd_header_offset
+ sizeof(cmd_header
));
3670 ret_code
= LTTNG_ERR_NOMEM
;
3675 case LTTNG_DOMAIN_KERNEL
:
3677 /* Kernel channels */
3678 struct ltt_kernel_channel
*kchan
;
3679 if (session
->kernel_session
!= NULL
) {
3680 cds_list_for_each_entry(kchan
,
3681 &session
->kernel_session
->channel_list
.head
, list
) {
3682 uint64_t discarded_events
, lost_packets
;
3683 struct lttng_channel_extended
*extended
;
3685 extended
= (struct lttng_channel_extended
*)
3686 kchan
->channel
->attr
.extended
.ptr
;
3688 ret
= get_kernel_runtime_stats(session
, kchan
,
3689 &discarded_events
, &lost_packets
);
3691 ret_code
= LTTNG_ERR_UNK
;
3696 * Update the discarded_events and lost_packets
3697 * count for the channel
3699 extended
->discarded_events
= discarded_events
;
3700 extended
->lost_packets
= lost_packets
;
3702 ret
= lttng_channel_serialize(
3703 kchan
->channel
, &payload
->buffer
);
3705 ERR("Failed to serialize lttng_channel: channel name = '%s'",
3706 kchan
->channel
->name
);
3707 ret_code
= LTTNG_ERR_UNK
;
3716 case LTTNG_DOMAIN_UST
:
3718 struct lttng_ht_iter iter
;
3719 struct ltt_ust_channel
*uchan
;
3722 cds_lfht_for_each_entry(session
->ust_session
->domain_global
.channels
->ht
,
3723 &iter
.iter
, uchan
, node
.node
) {
3724 uint64_t discarded_events
= 0, lost_packets
= 0;
3725 struct lttng_channel
*channel
= NULL
;
3726 struct lttng_channel_extended
*extended
;
3728 channel
= trace_ust_channel_to_lttng_channel(uchan
);
3730 ret
= LTTNG_ERR_NOMEM
;
3734 extended
= (struct lttng_channel_extended
*)
3735 channel
->attr
.extended
.ptr
;
3737 ret
= get_ust_runtime_stats(session
, uchan
,
3738 &discarded_events
, &lost_packets
);
3740 lttng_channel_destroy(channel
);
3741 ret_code
= LTTNG_ERR_UNK
;
3745 extended
->discarded_events
= discarded_events
;
3746 extended
->lost_packets
= lost_packets
;
3748 ret
= lttng_channel_serialize(
3749 channel
, &payload
->buffer
);
3751 ERR("Failed to serialize lttng_channel: channel name = '%s'",
3753 ret_code
= LTTNG_ERR_UNK
;
3767 if (i
> UINT32_MAX
) {
3768 ERR("Channel count would overflow the channel listing command's reply");
3769 ret_code
= LTTNG_ERR_OVERFLOW
;
3773 /* Update command reply header. */
3774 cmd_header
.count
= (uint32_t) i
;
3775 memcpy(payload
->buffer
.data
+ cmd_header_offset
, &cmd_header
,
3776 sizeof(cmd_header
));
3777 ret_code
= LTTNG_OK
;
3784 * Command LTTNG_LIST_EVENTS processed by the client thread.
3786 enum lttng_error_code
cmd_list_events(enum lttng_domain_type domain
,
3787 struct ltt_session
*session
,
3789 struct lttng_payload
*reply_payload
)
3791 int buffer_resize_ret
;
3792 enum lttng_error_code ret_code
= LTTNG_OK
;
3793 struct lttcomm_list_command_header reply_command_header
= {};
3794 size_t reply_command_header_offset
;
3795 unsigned int nb_events
= 0;
3797 assert(reply_payload
);
3799 /* Reserve space for command reply header. */
3800 reply_command_header_offset
= reply_payload
->buffer
.size
;
3801 buffer_resize_ret
= lttng_dynamic_buffer_set_size(&reply_payload
->buffer
,
3802 reply_command_header_offset
+
3803 sizeof(struct lttcomm_list_command_header
));
3804 if (buffer_resize_ret
) {
3805 ret_code
= LTTNG_ERR_NOMEM
;
3810 case LTTNG_DOMAIN_KERNEL
:
3811 if (session
->kernel_session
!= NULL
) {
3812 ret_code
= list_lttng_kernel_events(channel_name
,
3813 session
->kernel_session
, reply_payload
, &nb_events
);
3817 case LTTNG_DOMAIN_UST
:
3819 if (session
->ust_session
!= NULL
) {
3820 ret_code
= list_lttng_ust_global_events(channel_name
,
3821 &session
->ust_session
->domain_global
,
3822 reply_payload
, &nb_events
);
3827 case LTTNG_DOMAIN_LOG4J
:
3828 case LTTNG_DOMAIN_JUL
:
3829 case LTTNG_DOMAIN_PYTHON
:
3830 if (session
->ust_session
) {
3831 struct lttng_ht_iter iter
;
3835 cds_lfht_for_each_entry(session
->ust_session
->agents
->ht
,
3836 &iter
.iter
, agt
, node
.node
) {
3837 if (agt
->domain
== domain
) {
3838 ret_code
= list_lttng_agent_events(
3839 agt
, reply_payload
, &nb_events
);
3848 ret_code
= LTTNG_ERR_UND
;
3852 if (nb_events
> UINT32_MAX
) {
3853 ret_code
= LTTNG_ERR_OVERFLOW
;
3857 /* Update command reply header. */
3858 reply_command_header
.count
= (uint32_t) nb_events
;
3859 memcpy(reply_payload
->buffer
.data
+ reply_command_header_offset
, &reply_command_header
,
3860 sizeof(reply_command_header
));
3867 * Using the session list, filled a lttng_session array to send back to the
3868 * client for session listing.
3870 * The session list lock MUST be acquired before calling this function. Use
3871 * session_lock_list() and session_unlock_list().
3873 void cmd_list_lttng_sessions(struct lttng_session
*sessions
,
3874 size_t session_count
, uid_t uid
, gid_t gid
)
3878 struct ltt_session
*session
;
3879 struct ltt_session_list
*list
= session_get_list();
3880 struct lttng_session_extended
*extended
=
3881 (typeof(extended
)) (&sessions
[session_count
]);
3883 DBG("Getting all available session for UID %d GID %d",
3886 * Iterate over session list and append data after the control struct in
3889 cds_list_for_each_entry(session
, &list
->head
, list
) {
3890 if (!session_get(session
)) {
3894 * Only list the sessions the user can control.
3896 if (!session_access_ok(session
, uid
) ||
3897 session
->destroyed
) {
3898 session_put(session
);
3902 struct ltt_kernel_session
*ksess
= session
->kernel_session
;
3903 struct ltt_ust_session
*usess
= session
->ust_session
;
3905 if (session
->consumer
->type
== CONSUMER_DST_NET
||
3906 (ksess
&& ksess
->consumer
->type
== CONSUMER_DST_NET
) ||
3907 (usess
&& usess
->consumer
->type
== CONSUMER_DST_NET
)) {
3908 ret
= build_network_session_path(sessions
[i
].path
,
3909 sizeof(sessions
[i
].path
), session
);
3911 ret
= snprintf(sessions
[i
].path
, sizeof(sessions
[i
].path
), "%s",
3912 session
->consumer
->dst
.session_root_path
);
3915 PERROR("snprintf session path");
3916 session_put(session
);
3920 strncpy(sessions
[i
].name
, session
->name
, NAME_MAX
);
3921 sessions
[i
].name
[NAME_MAX
- 1] = '\0';
3922 sessions
[i
].enabled
= session
->active
;
3923 sessions
[i
].snapshot_mode
= session
->snapshot_mode
;
3924 sessions
[i
].live_timer_interval
= session
->live_timer
;
3925 extended
[i
].creation_time
.value
= (uint64_t) session
->creation_time
;
3926 extended
[i
].creation_time
.is_set
= 1;
3928 session_put(session
);
3933 * Command LTTNG_DATA_PENDING returning 0 if the data is NOT pending meaning
3934 * ready for trace analysis (or any kind of reader) or else 1 for pending data.
3936 int cmd_data_pending(struct ltt_session
*session
)
3939 struct ltt_kernel_session
*ksess
= session
->kernel_session
;
3940 struct ltt_ust_session
*usess
= session
->ust_session
;
3944 DBG("Data pending for session %s", session
->name
);
3946 /* Session MUST be stopped to ask for data availability. */
3947 if (session
->active
) {
3948 ret
= LTTNG_ERR_SESSION_STARTED
;
3952 * If stopped, just make sure we've started before else the above call
3953 * will always send that there is data pending.
3955 * The consumer assumes that when the data pending command is received,
3956 * the trace has been started before or else no output data is written
3957 * by the streams which is a condition for data pending. So, this is
3958 * *VERY* important that we don't ask the consumer before a start
3961 if (!session
->has_been_started
) {
3967 /* A rotation is still pending, we have to wait. */
3968 if (session
->rotation_state
== LTTNG_ROTATION_STATE_ONGOING
) {
3969 DBG("Rotate still pending for session %s", session
->name
);
3974 if (ksess
&& ksess
->consumer
) {
3975 ret
= consumer_is_data_pending(ksess
->id
, ksess
->consumer
);
3977 /* Data is still being extracted for the kernel. */
3982 if (usess
&& usess
->consumer
) {
3983 ret
= consumer_is_data_pending(usess
->id
, usess
->consumer
);
3985 /* Data is still being extracted for the kernel. */
3990 /* Data is ready to be read by a viewer */
3998 * Command LTTNG_SNAPSHOT_ADD_OUTPUT from the lttng ctl library.
4000 * Return LTTNG_OK on success or else a LTTNG_ERR code.
4002 int cmd_snapshot_add_output(struct ltt_session
*session
,
4003 const struct lttng_snapshot_output
*output
, uint32_t *id
)
4006 struct snapshot_output
*new_output
;
4011 DBG("Cmd snapshot add output for session %s", session
->name
);
4014 * Can't create an output if the session is not set in no-output mode.
4016 if (session
->output_traces
) {
4017 ret
= LTTNG_ERR_NOT_SNAPSHOT_SESSION
;
4021 if (session
->has_non_mmap_channel
) {
4022 ret
= LTTNG_ERR_SNAPSHOT_UNSUPPORTED
;
4026 /* Only one output is allowed until we have the "tee" feature. */
4027 if (session
->snapshot
.nb_output
== 1) {
4028 ret
= LTTNG_ERR_SNAPSHOT_OUTPUT_EXIST
;
4032 new_output
= snapshot_output_alloc();
4034 ret
= LTTNG_ERR_NOMEM
;
4038 ret
= snapshot_output_init(session
, output
->max_size
, output
->name
,
4039 output
->ctrl_url
, output
->data_url
, session
->consumer
, new_output
,
4040 &session
->snapshot
);
4042 if (ret
== -ENOMEM
) {
4043 ret
= LTTNG_ERR_NOMEM
;
4045 ret
= LTTNG_ERR_INVALID
;
4051 snapshot_add_output(&session
->snapshot
, new_output
);
4053 *id
= new_output
->id
;
4060 snapshot_output_destroy(new_output
);
4066 * Command LTTNG_SNAPSHOT_DEL_OUTPUT from lib lttng ctl.
4068 * Return LTTNG_OK on success or else a LTTNG_ERR code.
4070 int cmd_snapshot_del_output(struct ltt_session
*session
,
4071 const struct lttng_snapshot_output
*output
)
4074 struct snapshot_output
*sout
= NULL
;
4082 * Permission denied to create an output if the session is not
4083 * set in no output mode.
4085 if (session
->output_traces
) {
4086 ret
= LTTNG_ERR_NOT_SNAPSHOT_SESSION
;
4091 DBG("Cmd snapshot del output id %" PRIu32
" for session %s", output
->id
,
4093 sout
= snapshot_find_output_by_id(output
->id
, &session
->snapshot
);
4094 } else if (*output
->name
!= '\0') {
4095 DBG("Cmd snapshot del output name %s for session %s", output
->name
,
4097 sout
= snapshot_find_output_by_name(output
->name
, &session
->snapshot
);
4100 ret
= LTTNG_ERR_INVALID
;
4104 snapshot_delete_output(&session
->snapshot
, sout
);
4105 snapshot_output_destroy(sout
);
4114 * Command LTTNG_SNAPSHOT_LIST_OUTPUT from lib lttng ctl.
4116 * If no output is available, outputs is untouched and 0 is returned.
4118 * Return the size of the newly allocated outputs or a negative LTTNG_ERR code.
4120 ssize_t
cmd_snapshot_list_outputs(struct ltt_session
*session
,
4121 struct lttng_snapshot_output
**outputs
)
4124 struct lttng_snapshot_output
*list
= NULL
;
4125 struct lttng_ht_iter iter
;
4126 struct snapshot_output
*output
;
4131 DBG("Cmd snapshot list outputs for session %s", session
->name
);
4134 * Permission denied to create an output if the session is not
4135 * set in no output mode.
4137 if (session
->output_traces
) {
4138 ret
= -LTTNG_ERR_NOT_SNAPSHOT_SESSION
;
4142 if (session
->snapshot
.nb_output
== 0) {
4147 list
= zmalloc(session
->snapshot
.nb_output
* sizeof(*list
));
4149 ret
= -LTTNG_ERR_NOMEM
;
4153 /* Copy list from session to the new list object. */
4155 cds_lfht_for_each_entry(session
->snapshot
.output_ht
->ht
, &iter
.iter
,
4156 output
, node
.node
) {
4157 assert(output
->consumer
);
4158 list
[idx
].id
= output
->id
;
4159 list
[idx
].max_size
= output
->max_size
;
4160 if (lttng_strncpy(list
[idx
].name
, output
->name
,
4161 sizeof(list
[idx
].name
))) {
4162 ret
= -LTTNG_ERR_INVALID
;
4165 if (output
->consumer
->type
== CONSUMER_DST_LOCAL
) {
4166 if (lttng_strncpy(list
[idx
].ctrl_url
,
4167 output
->consumer
->dst
.session_root_path
,
4168 sizeof(list
[idx
].ctrl_url
))) {
4169 ret
= -LTTNG_ERR_INVALID
;
4174 ret
= uri_to_str_url(&output
->consumer
->dst
.net
.control
,
4175 list
[idx
].ctrl_url
, sizeof(list
[idx
].ctrl_url
));
4177 ret
= -LTTNG_ERR_NOMEM
;
4182 ret
= uri_to_str_url(&output
->consumer
->dst
.net
.data
,
4183 list
[idx
].data_url
, sizeof(list
[idx
].data_url
));
4185 ret
= -LTTNG_ERR_NOMEM
;
4194 ret
= session
->snapshot
.nb_output
;
4203 * Check if we can regenerate the metadata for this session.
4204 * Only kernel, UST per-uid and non-live sessions are supported.
4206 * Return 0 if the metadata can be generated, a LTTNG_ERR code otherwise.
4209 int check_regenerate_metadata_support(struct ltt_session
*session
)
4215 if (session
->live_timer
!= 0) {
4216 ret
= LTTNG_ERR_LIVE_SESSION
;
4219 if (!session
->active
) {
4220 ret
= LTTNG_ERR_SESSION_NOT_STARTED
;
4223 if (session
->ust_session
) {
4224 switch (session
->ust_session
->buffer_type
) {
4225 case LTTNG_BUFFER_PER_UID
:
4227 case LTTNG_BUFFER_PER_PID
:
4228 ret
= LTTNG_ERR_PER_PID_SESSION
;
4232 ret
= LTTNG_ERR_UNK
;
4236 if (session
->consumer
->type
== CONSUMER_DST_NET
&&
4237 session
->consumer
->relay_minor_version
< 8) {
4238 ret
= LTTNG_ERR_RELAYD_VERSION_FAIL
;
4248 int clear_metadata_file(int fd
)
4253 lseek_ret
= lseek(fd
, 0, SEEK_SET
);
4254 if (lseek_ret
< 0) {
4260 ret
= ftruncate(fd
, 0);
4262 PERROR("ftruncate");
4271 int ust_regenerate_metadata(struct ltt_ust_session
*usess
)
4274 struct buffer_reg_uid
*uid_reg
= NULL
;
4275 struct buffer_reg_session
*session_reg
= NULL
;
4278 cds_list_for_each_entry(uid_reg
, &usess
->buffer_reg_uid_list
, lnode
) {
4279 struct ust_registry_session
*registry
;
4280 struct ust_registry_channel
*chan
;
4281 struct lttng_ht_iter iter_chan
;
4283 session_reg
= uid_reg
->registry
;
4284 registry
= session_reg
->reg
.ust
;
4286 pthread_mutex_lock(®istry
->lock
);
4287 registry
->metadata_len_sent
= 0;
4288 memset(registry
->metadata
, 0, registry
->metadata_alloc_len
);
4289 registry
->metadata_len
= 0;
4290 registry
->metadata_version
++;
4291 if (registry
->metadata_fd
> 0) {
4292 /* Clear the metadata file's content. */
4293 ret
= clear_metadata_file(registry
->metadata_fd
);
4295 pthread_mutex_unlock(®istry
->lock
);
4300 ret
= ust_metadata_session_statedump(registry
, NULL
,
4301 registry
->major
, registry
->minor
);
4303 pthread_mutex_unlock(®istry
->lock
);
4304 ERR("Failed to generate session metadata (err = %d)",
4308 cds_lfht_for_each_entry(registry
->channels
->ht
, &iter_chan
.iter
,
4310 struct ust_registry_event
*event
;
4311 struct lttng_ht_iter iter_event
;
4313 ret
= ust_metadata_channel_statedump(registry
, chan
);
4315 pthread_mutex_unlock(®istry
->lock
);
4316 ERR("Failed to generate channel metadata "
4320 cds_lfht_for_each_entry(chan
->ht
->ht
, &iter_event
.iter
,
4322 ret
= ust_metadata_event_statedump(registry
,
4325 pthread_mutex_unlock(®istry
->lock
);
4326 ERR("Failed to generate event metadata "
4332 pthread_mutex_unlock(®istry
->lock
);
4341 * Command LTTNG_REGENERATE_METADATA from the lttng-ctl library.
4343 * Ask the consumer to truncate the existing metadata file(s) and
4344 * then regenerate the metadata. Live and per-pid sessions are not
4345 * supported and return an error.
4347 * Return LTTNG_OK on success or else a LTTNG_ERR code.
4349 int cmd_regenerate_metadata(struct ltt_session
*session
)
4355 ret
= check_regenerate_metadata_support(session
);
4360 if (session
->kernel_session
) {
4361 ret
= kernctl_session_regenerate_metadata(
4362 session
->kernel_session
->fd
);
4364 ERR("Failed to regenerate the kernel metadata");
4369 if (session
->ust_session
) {
4370 ret
= ust_regenerate_metadata(session
->ust_session
);
4372 ERR("Failed to regenerate the UST metadata");
4376 DBG("Cmd metadata regenerate for session %s", session
->name
);
4384 * Command LTTNG_REGENERATE_STATEDUMP from the lttng-ctl library.
4386 * Ask the tracer to regenerate a new statedump.
4388 * Return LTTNG_OK on success or else a LTTNG_ERR code.
4390 int cmd_regenerate_statedump(struct ltt_session
*session
)
4396 if (!session
->active
) {
4397 ret
= LTTNG_ERR_SESSION_NOT_STARTED
;
4401 if (session
->kernel_session
) {
4402 ret
= kernctl_session_regenerate_statedump(
4403 session
->kernel_session
->fd
);
4405 * Currently, the statedump in kernel can only fail if out
4409 if (ret
== -ENOMEM
) {
4410 ret
= LTTNG_ERR_REGEN_STATEDUMP_NOMEM
;
4412 ret
= LTTNG_ERR_REGEN_STATEDUMP_FAIL
;
4414 ERR("Failed to regenerate the kernel statedump");
4419 if (session
->ust_session
) {
4420 ret
= ust_app_regenerate_statedump_all(session
->ust_session
);
4422 * Currently, the statedump in UST always returns 0.
4425 ret
= LTTNG_ERR_REGEN_STATEDUMP_FAIL
;
4426 ERR("Failed to regenerate the UST statedump");
4430 DBG("Cmd regenerate statedump for session %s", session
->name
);
4438 enum lttng_error_code
synchronize_tracer_notifier_register(
4439 struct notification_thread_handle
*notification_thread
,
4440 struct lttng_trigger
*trigger
, const struct lttng_credentials
*cmd_creds
)
4442 enum lttng_error_code ret_code
;
4443 const struct lttng_condition
*condition
=
4444 lttng_trigger_get_const_condition(trigger
);
4445 const char *trigger_name
;
4446 uid_t trigger_owner
;
4447 enum lttng_trigger_status trigger_status
;
4448 const enum lttng_domain_type trigger_domain
=
4449 lttng_trigger_get_underlying_domain_type_restriction(
4452 trigger_status
= lttng_trigger_get_owner_uid(trigger
, &trigger_owner
);
4453 assert(trigger_status
== LTTNG_TRIGGER_STATUS_OK
);
4456 assert(lttng_condition_get_type(condition
) ==
4457 LTTNG_CONDITION_TYPE_EVENT_RULE_MATCHES
);
4459 trigger_status
= lttng_trigger_get_name(trigger
, &trigger_name
);
4460 trigger_name
= trigger_status
== LTTNG_TRIGGER_STATUS_OK
?
4461 trigger_name
: "(anonymous)";
4463 session_lock_list();
4464 switch (trigger_domain
) {
4465 case LTTNG_DOMAIN_KERNEL
:
4467 ret_code
= kernel_register_event_notifier(trigger
, cmd_creds
);
4468 if (ret_code
!= LTTNG_OK
) {
4469 enum lttng_error_code notif_thread_unregister_ret
;
4471 notif_thread_unregister_ret
=
4472 notification_thread_command_unregister_trigger(
4473 notification_thread
, trigger
);
4475 if (notif_thread_unregister_ret
!= LTTNG_OK
) {
4476 /* Return the original error code. */
4477 ERR("Failed to unregister trigger from notification thread during error recovery: trigger name = '%s', trigger owner uid = %d, error code = %d",
4479 (int) trigger_owner
,
4485 case LTTNG_DOMAIN_UST
:
4486 ust_app_global_update_all_event_notifier_rules();
4488 case LTTNG_DOMAIN_JUL
:
4489 case LTTNG_DOMAIN_LOG4J
:
4490 case LTTNG_DOMAIN_PYTHON
:
4492 /* Agent domains. */
4493 struct agent
*agt
= agent_find_by_event_notifier_domain(
4497 agt
= agent_create(trigger_domain
);
4499 ret_code
= LTTNG_ERR_NOMEM
;
4500 goto end_unlock_session_list
;
4503 agent_add(agt
, the_trigger_agents_ht_by_domain
);
4506 ret_code
= trigger_agent_enable(trigger
, agt
);
4507 if (ret_code
!= LTTNG_OK
) {
4508 goto end_unlock_session_list
;
4513 case LTTNG_DOMAIN_NONE
:
4518 ret_code
= LTTNG_OK
;
4519 end_unlock_session_list
:
4520 session_unlock_list();
4524 enum lttng_error_code
cmd_register_trigger(const struct lttng_credentials
*cmd_creds
,
4525 struct lttng_trigger
*trigger
,
4526 bool is_trigger_anonymous
,
4527 struct notification_thread_handle
*notification_thread
,
4528 struct lttng_trigger
**return_trigger
)
4530 enum lttng_error_code ret_code
;
4531 const char *trigger_name
;
4532 uid_t trigger_owner
;
4533 enum lttng_trigger_status trigger_status
;
4535 trigger_status
= lttng_trigger_get_name(trigger
, &trigger_name
);
4536 trigger_name
= trigger_status
== LTTNG_TRIGGER_STATUS_OK
?
4537 trigger_name
: "(anonymous)";
4539 trigger_status
= lttng_trigger_get_owner_uid(
4540 trigger
, &trigger_owner
);
4541 assert(trigger_status
== LTTNG_TRIGGER_STATUS_OK
);
4543 DBG("Running register trigger command: trigger name = '%s', trigger owner uid = %d, command creds uid = %d",
4544 trigger_name
, (int) trigger_owner
,
4545 (int) lttng_credentials_get_uid(cmd_creds
));
4548 * Validate the trigger credentials against the command credentials.
4549 * Only the root user can register a trigger with non-matching
4552 if (!lttng_credentials_is_equal_uid(
4553 lttng_trigger_get_credentials(trigger
),
4555 if (lttng_credentials_get_uid(cmd_creds
) != 0) {
4556 ERR("Trigger credentials do not match the command credentials: trigger name = '%s', trigger owner uid = %d, command creds uid = %d",
4557 trigger_name
, (int) trigger_owner
,
4558 (int) lttng_credentials_get_uid(cmd_creds
));
4559 ret_code
= LTTNG_ERR_INVALID_TRIGGER
;
4565 * The bytecode generation also serves as a validation step for the
4566 * bytecode expressions.
4568 ret_code
= lttng_trigger_generate_bytecode(trigger
, cmd_creds
);
4569 if (ret_code
!= LTTNG_OK
) {
4570 ERR("Failed to generate bytecode of trigger: trigger name = '%s', trigger owner uid = %d, error code = %d",
4571 trigger_name
, (int) trigger_owner
, ret_code
);
4576 * A reference to the trigger is acquired by the notification thread.
4577 * It is safe to return the same trigger to the caller since it the
4578 * other user holds a reference.
4580 * The trigger is modified during the execution of the
4581 * "register trigger" command. However, by the time the command returns,
4582 * it is safe to use without any locking as its properties are
4585 ret_code
= notification_thread_command_register_trigger(
4586 notification_thread
, trigger
, is_trigger_anonymous
);
4587 if (ret_code
!= LTTNG_OK
) {
4588 DBG("Failed to register trigger to notification thread: trigger name = '%s', trigger owner uid = %d, error code = %d",
4589 trigger_name
, (int) trigger_owner
, ret_code
);
4593 trigger_status
= lttng_trigger_get_name(trigger
, &trigger_name
);
4594 trigger_name
= trigger_status
== LTTNG_TRIGGER_STATUS_OK
?
4595 trigger_name
: "(anonymous)";
4598 * Synchronize tracers if the trigger adds an event notifier.
4600 if (lttng_trigger_needs_tracer_notifier(trigger
)) {
4601 ret_code
= synchronize_tracer_notifier_register(notification_thread
,
4602 trigger
, cmd_creds
);
4603 if (ret_code
!= LTTNG_OK
) {
4604 ERR("Error registering tracer notifier: %s",
4605 lttng_strerror(-ret_code
));
4611 * Return an updated trigger to the client.
4613 * Since a modified version of the same trigger is returned, acquire a
4614 * reference to the trigger so the caller doesn't have to care if those
4615 * are distinct instances or not.
4617 if (ret_code
== LTTNG_OK
) {
4618 lttng_trigger_get(trigger
);
4619 *return_trigger
= trigger
;
4620 /* Ownership of trigger was transferred to caller. */
4628 enum lttng_error_code
synchronize_tracer_notifier_unregister(
4629 const struct lttng_trigger
*trigger
)
4631 enum lttng_error_code ret_code
;
4632 const struct lttng_condition
*condition
=
4633 lttng_trigger_get_const_condition(trigger
);
4634 const enum lttng_domain_type trigger_domain
=
4635 lttng_trigger_get_underlying_domain_type_restriction(
4639 assert(lttng_condition_get_type(condition
) ==
4640 LTTNG_CONDITION_TYPE_EVENT_RULE_MATCHES
);
4642 session_lock_list();
4643 switch (trigger_domain
) {
4644 case LTTNG_DOMAIN_KERNEL
:
4645 ret_code
= kernel_unregister_event_notifier(trigger
);
4646 if (ret_code
!= LTTNG_OK
) {
4647 goto end_unlock_session_list
;
4651 case LTTNG_DOMAIN_UST
:
4652 ust_app_global_update_all_event_notifier_rules();
4654 case LTTNG_DOMAIN_JUL
:
4655 case LTTNG_DOMAIN_LOG4J
:
4656 case LTTNG_DOMAIN_PYTHON
:
4658 /* Agent domains. */
4659 struct agent
*agt
= agent_find_by_event_notifier_domain(
4663 * This trigger was never registered in the first place. Calling
4664 * this function under those circumstances is an internal error.
4667 ret_code
= trigger_agent_disable(trigger
, agt
);
4668 if (ret_code
!= LTTNG_OK
) {
4669 goto end_unlock_session_list
;
4674 case LTTNG_DOMAIN_NONE
:
4679 ret_code
= LTTNG_OK
;
4681 end_unlock_session_list
:
4682 session_unlock_list();
4686 enum lttng_error_code
cmd_unregister_trigger(const struct lttng_credentials
*cmd_creds
,
4687 const struct lttng_trigger
*trigger
,
4688 struct notification_thread_handle
*notification_thread
)
4690 enum lttng_error_code ret_code
;
4691 const char *trigger_name
;
4692 uid_t trigger_owner
;
4693 enum lttng_trigger_status trigger_status
;
4694 struct lttng_trigger
*sessiond_trigger
= NULL
;
4696 trigger_status
= lttng_trigger_get_name(trigger
, &trigger_name
);
4697 trigger_name
= trigger_status
== LTTNG_TRIGGER_STATUS_OK
? trigger_name
: "(anonymous)";
4698 trigger_status
= lttng_trigger_get_owner_uid(trigger
, &trigger_owner
);
4699 assert(trigger_status
== LTTNG_TRIGGER_STATUS_OK
);
4701 DBG("Running unregister trigger command: trigger name = '%s', trigger owner uid = %d, command creds uid = %d",
4702 trigger_name
, (int) trigger_owner
,
4703 (int) lttng_credentials_get_uid(cmd_creds
));
4706 * Validate the trigger credentials against the command credentials.
4707 * Only the root user can unregister a trigger with non-matching
4710 if (!lttng_credentials_is_equal_uid(
4711 lttng_trigger_get_credentials(trigger
),
4713 if (lttng_credentials_get_uid(cmd_creds
) != 0) {
4714 ERR("Trigger credentials do not match the command credentials: trigger name = '%s', trigger owner uid = %d, command creds uid = %d",
4715 trigger_name
, (int) trigger_owner
,
4716 (int) lttng_credentials_get_uid(cmd_creds
));
4717 ret_code
= LTTNG_ERR_INVALID_TRIGGER
;
4722 /* Fetch the sessiond side trigger object. */
4723 ret_code
= notification_thread_command_get_trigger(
4724 notification_thread
, trigger
, &sessiond_trigger
);
4725 if (ret_code
!= LTTNG_OK
) {
4726 DBG("Failed to get trigger from notification thread during unregister: trigger name = '%s', trigger owner uid = %d, error code = %d",
4727 trigger_name
, (int) trigger_owner
, ret_code
);
4731 assert(sessiond_trigger
);
4734 * From this point on, no matter what, consider the trigger
4737 * We set the unregistered state of the sessiond side trigger object in
4738 * the client thread since we want to minimize the possibility of the
4739 * notification thread being stalled due to a long execution of an
4740 * action that required the trigger lock.
4742 lttng_trigger_set_as_unregistered(sessiond_trigger
);
4744 ret_code
= notification_thread_command_unregister_trigger(notification_thread
,
4746 if (ret_code
!= LTTNG_OK
) {
4747 DBG("Failed to unregister trigger from notification thread: trigger name = '%s', trigger owner uid = %d, error code = %d",
4748 trigger_name
, (int) trigger_owner
, ret_code
);
4753 * Synchronize tracers if the trigger removes an event notifier.
4754 * Do this even if the trigger unregistration failed to at least stop
4755 * the tracers from producing notifications associated with this
4758 if (lttng_trigger_needs_tracer_notifier(trigger
)) {
4759 ret_code
= synchronize_tracer_notifier_unregister(trigger
);
4760 if (ret_code
!= LTTNG_OK
) {
4761 ERR("Error unregistering trigger to tracer.");
4768 lttng_trigger_put(sessiond_trigger
);
4772 enum lttng_error_code
cmd_list_triggers(struct command_ctx
*cmd_ctx
,
4773 struct notification_thread_handle
*notification_thread
,
4774 struct lttng_triggers
**return_triggers
)
4777 enum lttng_error_code ret_code
;
4778 struct lttng_triggers
*triggers
= NULL
;
4780 /* Get the set of triggers from the notification thread. */
4781 ret_code
= notification_thread_command_list_triggers(
4782 notification_thread
, cmd_ctx
->creds
.uid
, &triggers
);
4783 if (ret_code
!= LTTNG_OK
) {
4787 ret
= lttng_triggers_remove_hidden_triggers(triggers
);
4789 ret_code
= LTTNG_ERR_UNK
;
4793 *return_triggers
= triggers
;
4795 ret_code
= LTTNG_OK
;
4797 lttng_triggers_destroy(triggers
);
4801 enum lttng_error_code
cmd_execute_error_query(const struct lttng_credentials
*cmd_creds
,
4802 const struct lttng_error_query
*query
,
4803 struct lttng_error_query_results
**_results
,
4804 struct notification_thread_handle
*notification_thread
)
4806 enum lttng_error_code ret_code
;
4807 const struct lttng_trigger
*query_target_trigger
;
4808 const struct lttng_action
*query_target_action
= NULL
;
4809 struct lttng_trigger
*matching_trigger
= NULL
;
4810 const char *trigger_name
;
4811 uid_t trigger_owner
;
4812 enum lttng_trigger_status trigger_status
;
4813 struct lttng_error_query_results
*results
= NULL
;
4815 switch (lttng_error_query_get_target_type(query
)) {
4816 case LTTNG_ERROR_QUERY_TARGET_TYPE_TRIGGER
:
4817 query_target_trigger
= lttng_error_query_trigger_borrow_target(query
);
4819 case LTTNG_ERROR_QUERY_TARGET_TYPE_CONDITION
:
4820 query_target_trigger
=
4821 lttng_error_query_condition_borrow_target(query
);
4823 case LTTNG_ERROR_QUERY_TARGET_TYPE_ACTION
:
4824 query_target_trigger
= lttng_error_query_action_borrow_trigger_target(
4831 assert(query_target_trigger
);
4833 ret_code
= notification_thread_command_get_trigger(notification_thread
,
4834 query_target_trigger
, &matching_trigger
);
4835 if (ret_code
!= LTTNG_OK
) {
4839 /* No longer needed. */
4840 query_target_trigger
= NULL
;
4842 if (lttng_error_query_get_target_type(query
) ==
4843 LTTNG_ERROR_QUERY_TARGET_TYPE_ACTION
) {
4844 /* Get the sessiond-side version of the target action. */
4845 query_target_action
=
4846 lttng_error_query_action_borrow_action_target(
4847 query
, matching_trigger
);
4850 trigger_status
= lttng_trigger_get_name(matching_trigger
, &trigger_name
);
4851 trigger_name
= trigger_status
== LTTNG_TRIGGER_STATUS_OK
?
4852 trigger_name
: "(anonymous)";
4853 trigger_status
= lttng_trigger_get_owner_uid(matching_trigger
,
4855 assert(trigger_status
== LTTNG_TRIGGER_STATUS_OK
);
4857 results
= lttng_error_query_results_create();
4859 ret_code
= LTTNG_ERR_NOMEM
;
4863 DBG("Running \"execute error query\" command: trigger name = '%s', trigger owner uid = %d, command creds uid = %d",
4864 trigger_name
, (int) trigger_owner
,
4865 (int) lttng_credentials_get_uid(cmd_creds
));
4868 * Validate the trigger credentials against the command credentials.
4869 * Only the root user can target a trigger with non-matching
4872 if (!lttng_credentials_is_equal_uid(
4873 lttng_trigger_get_credentials(matching_trigger
),
4875 if (lttng_credentials_get_uid(cmd_creds
) != 0) {
4876 ERR("Trigger credentials do not match the command credentials: trigger name = '%s', trigger owner uid = %d, command creds uid = %d",
4877 trigger_name
, (int) trigger_owner
,
4878 (int) lttng_credentials_get_uid(cmd_creds
));
4879 ret_code
= LTTNG_ERR_INVALID_TRIGGER
;
4884 switch (lttng_error_query_get_target_type(query
)) {
4885 case LTTNG_ERROR_QUERY_TARGET_TYPE_TRIGGER
:
4886 trigger_status
= lttng_trigger_add_error_results(
4887 matching_trigger
, results
);
4889 switch (trigger_status
) {
4890 case LTTNG_TRIGGER_STATUS_OK
:
4893 ret_code
= LTTNG_ERR_UNK
;
4898 case LTTNG_ERROR_QUERY_TARGET_TYPE_CONDITION
:
4900 trigger_status
= lttng_trigger_condition_add_error_results(
4901 matching_trigger
, results
);
4903 switch (trigger_status
) {
4904 case LTTNG_TRIGGER_STATUS_OK
:
4907 ret_code
= LTTNG_ERR_UNK
;
4913 case LTTNG_ERROR_QUERY_TARGET_TYPE_ACTION
:
4915 const enum lttng_action_status action_status
=
4916 lttng_action_add_error_query_results(
4917 query_target_action
, results
);
4919 switch (action_status
) {
4920 case LTTNG_ACTION_STATUS_OK
:
4923 ret_code
= LTTNG_ERR_UNK
;
4934 *_results
= results
;
4936 ret_code
= LTTNG_OK
;
4938 lttng_trigger_put(matching_trigger
);
4939 lttng_error_query_results_destroy(results
);
4944 * Send relayd sockets from snapshot output to consumer. Ignore request if the
4945 * snapshot output is *not* set with a remote destination.
4947 * Return LTTNG_OK on success or a LTTNG_ERR code.
4949 static enum lttng_error_code
set_relayd_for_snapshot(
4950 struct consumer_output
*output
,
4951 const struct ltt_session
*session
)
4953 enum lttng_error_code status
= LTTNG_OK
;
4954 struct lttng_ht_iter iter
;
4955 struct consumer_socket
*socket
;
4956 LTTNG_OPTIONAL(uint64_t) current_chunk_id
= {};
4957 const char *base_path
;
4962 DBG2("Set relayd object from snapshot output");
4964 if (session
->current_trace_chunk
) {
4965 enum lttng_trace_chunk_status chunk_status
=
4966 lttng_trace_chunk_get_id(
4967 session
->current_trace_chunk
,
4968 ¤t_chunk_id
.value
);
4970 if (chunk_status
== LTTNG_TRACE_CHUNK_STATUS_OK
) {
4971 current_chunk_id
.is_set
= true;
4973 ERR("Failed to get current trace chunk id");
4974 status
= LTTNG_ERR_UNK
;
4979 /* Ignore if snapshot consumer output is not network. */
4980 if (output
->type
!= CONSUMER_DST_NET
) {
4985 * The snapshot record URI base path overrides the session
4988 if (output
->dst
.net
.control
.subdir
[0] != '\0') {
4989 base_path
= output
->dst
.net
.control
.subdir
;
4991 base_path
= session
->base_path
;
4995 * For each consumer socket, create and send the relayd object of the
4999 cds_lfht_for_each_entry(output
->socks
->ht
, &iter
.iter
,
5000 socket
, node
.node
) {
5001 pthread_mutex_lock(socket
->lock
);
5002 status
= send_consumer_relayd_sockets(0, session
->id
,
5004 session
->name
, session
->hostname
,
5006 session
->live_timer
,
5007 current_chunk_id
.is_set
? ¤t_chunk_id
.value
: NULL
,
5008 session
->creation_time
,
5009 session
->name_contains_creation_time
);
5010 pthread_mutex_unlock(socket
->lock
);
5011 if (status
!= LTTNG_OK
) {
5023 * Record a kernel snapshot.
5025 * Return LTTNG_OK on success or a LTTNG_ERR code.
5027 static enum lttng_error_code
record_kernel_snapshot(
5028 struct ltt_kernel_session
*ksess
,
5029 const struct consumer_output
*output
,
5030 const struct ltt_session
*session
,
5031 int wait
, uint64_t nb_packets_per_stream
)
5033 enum lttng_error_code status
;
5039 status
= kernel_snapshot_record(
5040 ksess
, output
, wait
, nb_packets_per_stream
);
5045 * Record a UST snapshot.
5047 * Returns LTTNG_OK on success or a LTTNG_ERR error code.
5049 static enum lttng_error_code
record_ust_snapshot(struct ltt_ust_session
*usess
,
5050 const struct consumer_output
*output
,
5051 const struct ltt_session
*session
,
5052 int wait
, uint64_t nb_packets_per_stream
)
5054 enum lttng_error_code status
;
5060 status
= ust_app_snapshot_record(
5061 usess
, output
, wait
, nb_packets_per_stream
);
5066 uint64_t get_session_size_one_more_packet_per_stream(
5067 const struct ltt_session
*session
, uint64_t cur_nr_packets
)
5069 uint64_t tot_size
= 0;
5071 if (session
->kernel_session
) {
5072 struct ltt_kernel_channel
*chan
;
5073 const struct ltt_kernel_session
*ksess
=
5074 session
->kernel_session
;
5076 cds_list_for_each_entry(chan
, &ksess
->channel_list
.head
, list
) {
5077 if (cur_nr_packets
>= chan
->channel
->attr
.num_subbuf
) {
5079 * Don't take channel into account if we
5080 * already grab all its packets.
5084 tot_size
+= chan
->channel
->attr
.subbuf_size
5085 * chan
->stream_count
;
5089 if (session
->ust_session
) {
5090 const struct ltt_ust_session
*usess
= session
->ust_session
;
5092 tot_size
+= ust_app_get_size_one_more_packet_per_stream(usess
,
5100 * Calculate the number of packets we can grab from each stream that
5101 * fits within the overall snapshot max size.
5103 * Returns -1 on error, 0 means infinite number of packets, else > 0 is
5104 * the number of packets per stream.
5106 * TODO: this approach is not perfect: we consider the worse case
5107 * (packet filling the sub-buffers) as an upper bound, but we could do
5108 * better if we do this calculation while we actually grab the packet
5109 * content: we would know how much padding we don't actually store into
5112 * This algorithm is currently bounded by the number of packets per
5115 * Since we call this algorithm before actually grabbing the data, it's
5116 * an approximation: for instance, applications could appear/disappear
5117 * in between this call and actually grabbing data.
5120 int64_t get_session_nb_packets_per_stream(const struct ltt_session
*session
,
5124 uint64_t cur_nb_packets
= 0;
5127 return 0; /* Infinite */
5130 size_left
= max_size
;
5132 uint64_t one_more_packet_tot_size
;
5134 one_more_packet_tot_size
= get_session_size_one_more_packet_per_stream(
5135 session
, cur_nb_packets
);
5136 if (!one_more_packet_tot_size
) {
5137 /* We are already grabbing all packets. */
5140 size_left
-= one_more_packet_tot_size
;
5141 if (size_left
< 0) {
5146 if (!cur_nb_packets
&& size_left
!= max_size
) {
5147 /* Not enough room to grab one packet of each stream, error. */
5150 return cur_nb_packets
;
5154 enum lttng_error_code
snapshot_record(struct ltt_session
*session
,
5155 const struct snapshot_output
*snapshot_output
, int wait
)
5157 int64_t nb_packets_per_stream
;
5158 char snapshot_chunk_name
[LTTNG_NAME_MAX
];
5160 enum lttng_error_code ret_code
= LTTNG_OK
;
5161 struct lttng_trace_chunk
*snapshot_trace_chunk
;
5162 struct consumer_output
*original_ust_consumer_output
= NULL
;
5163 struct consumer_output
*original_kernel_consumer_output
= NULL
;
5164 struct consumer_output
*snapshot_ust_consumer_output
= NULL
;
5165 struct consumer_output
*snapshot_kernel_consumer_output
= NULL
;
5167 ret
= snprintf(snapshot_chunk_name
, sizeof(snapshot_chunk_name
),
5169 snapshot_output
->name
,
5170 snapshot_output
->datetime
,
5171 snapshot_output
->nb_snapshot
);
5172 if (ret
< 0 || ret
>= sizeof(snapshot_chunk_name
)) {
5173 ERR("Failed to format snapshot name");
5174 ret_code
= LTTNG_ERR_INVALID
;
5177 DBG("Recording snapshot \"%s\" for session \"%s\" with chunk name \"%s\"",
5178 snapshot_output
->name
, session
->name
,
5179 snapshot_chunk_name
);
5180 if (!session
->kernel_session
&& !session
->ust_session
) {
5181 ERR("Failed to record snapshot as no channels exist");
5182 ret_code
= LTTNG_ERR_NO_CHANNEL
;
5186 if (session
->kernel_session
) {
5187 original_kernel_consumer_output
=
5188 session
->kernel_session
->consumer
;
5189 snapshot_kernel_consumer_output
=
5190 consumer_copy_output(snapshot_output
->consumer
);
5191 strcpy(snapshot_kernel_consumer_output
->chunk_path
,
5192 snapshot_chunk_name
);
5194 /* Copy the original domain subdir. */
5195 strcpy(snapshot_kernel_consumer_output
->domain_subdir
,
5196 original_kernel_consumer_output
->domain_subdir
);
5198 ret
= consumer_copy_sockets(snapshot_kernel_consumer_output
,
5199 original_kernel_consumer_output
);
5201 ERR("Failed to copy consumer sockets from snapshot output configuration");
5202 ret_code
= LTTNG_ERR_NOMEM
;
5205 ret_code
= set_relayd_for_snapshot(
5206 snapshot_kernel_consumer_output
, session
);
5207 if (ret_code
!= LTTNG_OK
) {
5208 ERR("Failed to setup relay daemon for kernel tracer snapshot");
5211 session
->kernel_session
->consumer
=
5212 snapshot_kernel_consumer_output
;
5214 if (session
->ust_session
) {
5215 original_ust_consumer_output
= session
->ust_session
->consumer
;
5216 snapshot_ust_consumer_output
=
5217 consumer_copy_output(snapshot_output
->consumer
);
5218 strcpy(snapshot_ust_consumer_output
->chunk_path
,
5219 snapshot_chunk_name
);
5221 /* Copy the original domain subdir. */
5222 strcpy(snapshot_ust_consumer_output
->domain_subdir
,
5223 original_ust_consumer_output
->domain_subdir
);
5225 ret
= consumer_copy_sockets(snapshot_ust_consumer_output
,
5226 original_ust_consumer_output
);
5228 ERR("Failed to copy consumer sockets from snapshot output configuration");
5229 ret_code
= LTTNG_ERR_NOMEM
;
5232 ret_code
= set_relayd_for_snapshot(
5233 snapshot_ust_consumer_output
, session
);
5234 if (ret_code
!= LTTNG_OK
) {
5235 ERR("Failed to setup relay daemon for userspace tracer snapshot");
5238 session
->ust_session
->consumer
=
5239 snapshot_ust_consumer_output
;
5242 snapshot_trace_chunk
= session_create_new_trace_chunk(session
,
5243 snapshot_kernel_consumer_output
?:
5244 snapshot_ust_consumer_output
,
5245 consumer_output_get_base_path(
5246 snapshot_output
->consumer
),
5247 snapshot_chunk_name
);
5248 if (!snapshot_trace_chunk
) {
5249 ERR("Failed to create temporary trace chunk to record a snapshot of session \"%s\"",
5251 ret_code
= LTTNG_ERR_CREATE_DIR_FAIL
;
5254 assert(!session
->current_trace_chunk
);
5255 ret
= session_set_trace_chunk(session
, snapshot_trace_chunk
, NULL
);
5256 lttng_trace_chunk_put(snapshot_trace_chunk
);
5257 snapshot_trace_chunk
= NULL
;
5259 ERR("Failed to set temporary trace chunk to record a snapshot of session \"%s\"",
5261 ret_code
= LTTNG_ERR_CREATE_TRACE_CHUNK_FAIL_CONSUMER
;
5265 nb_packets_per_stream
= get_session_nb_packets_per_stream(session
,
5266 snapshot_output
->max_size
);
5267 if (nb_packets_per_stream
< 0) {
5268 ret_code
= LTTNG_ERR_MAX_SIZE_INVALID
;
5269 goto error_close_trace_chunk
;
5272 if (session
->kernel_session
) {
5273 ret_code
= record_kernel_snapshot(session
->kernel_session
,
5274 snapshot_kernel_consumer_output
, session
,
5275 wait
, nb_packets_per_stream
);
5276 if (ret_code
!= LTTNG_OK
) {
5277 goto error_close_trace_chunk
;
5281 if (session
->ust_session
) {
5282 ret_code
= record_ust_snapshot(session
->ust_session
,
5283 snapshot_ust_consumer_output
, session
,
5284 wait
, nb_packets_per_stream
);
5285 if (ret_code
!= LTTNG_OK
) {
5286 goto error_close_trace_chunk
;
5290 error_close_trace_chunk
:
5291 if (session_set_trace_chunk(session
, NULL
, &snapshot_trace_chunk
)) {
5292 ERR("Failed to release the current trace chunk of session \"%s\"",
5294 ret_code
= LTTNG_ERR_UNK
;
5297 if (session_close_trace_chunk(session
, snapshot_trace_chunk
,
5298 LTTNG_TRACE_CHUNK_COMMAND_TYPE_NO_OPERATION
, NULL
)) {
5300 * Don't goto end; make sure the chunk is closed for the session
5301 * to allow future snapshots.
5303 ERR("Failed to close snapshot trace chunk of session \"%s\"",
5305 ret_code
= LTTNG_ERR_CLOSE_TRACE_CHUNK_FAIL_CONSUMER
;
5308 lttng_trace_chunk_put(snapshot_trace_chunk
);
5309 snapshot_trace_chunk
= NULL
;
5311 if (original_ust_consumer_output
) {
5312 session
->ust_session
->consumer
= original_ust_consumer_output
;
5314 if (original_kernel_consumer_output
) {
5315 session
->kernel_session
->consumer
=
5316 original_kernel_consumer_output
;
5318 consumer_output_put(snapshot_ust_consumer_output
);
5319 consumer_output_put(snapshot_kernel_consumer_output
);
5324 * Command LTTNG_SNAPSHOT_RECORD from lib lttng ctl.
5326 * The wait parameter is ignored so this call always wait for the snapshot to
5327 * complete before returning.
5329 * Return LTTNG_OK on success or else a LTTNG_ERR code.
5331 int cmd_snapshot_record(struct ltt_session
*session
,
5332 const struct lttng_snapshot_output
*output
, int wait
)
5334 enum lttng_error_code cmd_ret
= LTTNG_OK
;
5336 unsigned int snapshot_success
= 0;
5338 struct snapshot_output
*tmp_output
= NULL
;
5343 DBG("Cmd snapshot record for session %s", session
->name
);
5345 /* Get the datetime for the snapshot output directory. */
5346 ret
= utils_get_current_time_str("%Y%m%d-%H%M%S", datetime
,
5349 cmd_ret
= LTTNG_ERR_INVALID
;
5354 * Permission denied to create an output if the session is not
5355 * set in no output mode.
5357 if (session
->output_traces
) {
5358 cmd_ret
= LTTNG_ERR_NOT_SNAPSHOT_SESSION
;
5362 /* The session needs to be started at least once. */
5363 if (!session
->has_been_started
) {
5364 cmd_ret
= LTTNG_ERR_START_SESSION_ONCE
;
5368 /* Use temporary output for the session. */
5369 if (*output
->ctrl_url
!= '\0') {
5370 tmp_output
= snapshot_output_alloc();
5372 cmd_ret
= LTTNG_ERR_NOMEM
;
5376 ret
= snapshot_output_init(session
, output
->max_size
,
5378 output
->ctrl_url
, output
->data_url
,
5382 if (ret
== -ENOMEM
) {
5383 cmd_ret
= LTTNG_ERR_NOMEM
;
5385 cmd_ret
= LTTNG_ERR_INVALID
;
5389 /* Use the global session count for the temporary snapshot. */
5390 tmp_output
->nb_snapshot
= session
->snapshot
.nb_snapshot
;
5392 /* Use the global datetime */
5393 memcpy(tmp_output
->datetime
, datetime
, sizeof(datetime
));
5394 cmd_ret
= snapshot_record(session
, tmp_output
, wait
);
5395 if (cmd_ret
!= LTTNG_OK
) {
5398 snapshot_success
= 1;
5400 struct snapshot_output
*sout
;
5401 struct lttng_ht_iter iter
;
5404 cds_lfht_for_each_entry(session
->snapshot
.output_ht
->ht
,
5405 &iter
.iter
, sout
, node
.node
) {
5406 struct snapshot_output output_copy
;
5409 * Make a local copy of the output and override output
5410 * parameters with those provided as part of the
5413 memcpy(&output_copy
, sout
, sizeof(output_copy
));
5415 if (output
->max_size
!= (uint64_t) -1ULL) {
5416 output_copy
.max_size
= output
->max_size
;
5419 output_copy
.nb_snapshot
= session
->snapshot
.nb_snapshot
;
5420 memcpy(output_copy
.datetime
, datetime
,
5423 /* Use temporary name. */
5424 if (*output
->name
!= '\0') {
5425 if (lttng_strncpy(output_copy
.name
,
5427 sizeof(output_copy
.name
))) {
5428 cmd_ret
= LTTNG_ERR_INVALID
;
5434 cmd_ret
= snapshot_record(session
, &output_copy
, wait
);
5435 if (cmd_ret
!= LTTNG_OK
) {
5439 snapshot_success
= 1;
5444 if (snapshot_success
) {
5445 session
->snapshot
.nb_snapshot
++;
5447 cmd_ret
= LTTNG_ERR_SNAPSHOT_FAIL
;
5452 snapshot_output_destroy(tmp_output
);
5458 * Command LTTNG_SET_SESSION_SHM_PATH processed by the client thread.
5460 int cmd_set_session_shm_path(struct ltt_session
*session
,
5461 const char *shm_path
)
5467 * Can only set shm path before session is started.
5469 if (session
->has_been_started
) {
5470 return LTTNG_ERR_SESSION_STARTED
;
5473 strncpy(session
->shm_path
, shm_path
,
5474 sizeof(session
->shm_path
));
5475 session
->shm_path
[sizeof(session
->shm_path
) - 1] = '\0';
5481 * Command LTTNG_ROTATE_SESSION from the lttng-ctl library.
5483 * Ask the consumer to rotate the session output directory.
5484 * The session lock must be held.
5486 * Returns LTTNG_OK on success or else a negative LTTng error code.
5488 int cmd_rotate_session(struct ltt_session
*session
,
5489 struct lttng_rotate_session_return
*rotate_return
,
5490 bool quiet_rotation
,
5491 enum lttng_trace_chunk_command_type command
)
5494 uint64_t ongoing_rotation_chunk_id
;
5495 enum lttng_error_code cmd_ret
= LTTNG_OK
;
5496 struct lttng_trace_chunk
*chunk_being_archived
= NULL
;
5497 struct lttng_trace_chunk
*new_trace_chunk
= NULL
;
5498 enum lttng_trace_chunk_status chunk_status
;
5499 bool failed_to_rotate
= false;
5500 enum lttng_error_code rotation_fail_code
= LTTNG_OK
;
5504 if (!session
->has_been_started
) {
5505 cmd_ret
= LTTNG_ERR_START_SESSION_ONCE
;
5510 * Explicit rotation is not supported for live sessions.
5511 * However, live sessions can perform a quiet rotation on
5513 * Rotation is not supported for snapshot traces (no output).
5515 if ((!quiet_rotation
&& session
->live_timer
) ||
5516 !session
->output_traces
) {
5517 cmd_ret
= LTTNG_ERR_ROTATION_NOT_AVAILABLE
;
5521 /* Unsupported feature in lttng-relayd before 2.11. */
5522 if (!quiet_rotation
&& session
->consumer
->type
== CONSUMER_DST_NET
&&
5523 (session
->consumer
->relay_major_version
== 2 &&
5524 session
->consumer
->relay_minor_version
< 11)) {
5525 cmd_ret
= LTTNG_ERR_ROTATION_NOT_AVAILABLE_RELAY
;
5529 /* Unsupported feature in lttng-modules before 2.8 (lack of sequence number). */
5530 if (session
->kernel_session
&& !kernel_supports_ring_buffer_packet_sequence_number()) {
5531 cmd_ret
= LTTNG_ERR_ROTATION_NOT_AVAILABLE_KERNEL
;
5535 if (session
->rotation_state
== LTTNG_ROTATION_STATE_ONGOING
) {
5536 DBG("Refusing to launch a rotation; a rotation is already in progress for session %s",
5538 cmd_ret
= LTTNG_ERR_ROTATION_PENDING
;
5543 * After a stop, we only allow one rotation to occur, the other ones are
5544 * useless until a new start.
5546 if (session
->rotated_after_last_stop
) {
5547 DBG("Session \"%s\" was already rotated after stop, refusing rotation",
5549 cmd_ret
= LTTNG_ERR_ROTATION_MULTIPLE_AFTER_STOP
;
5554 * After a stop followed by a clear, disallow following rotations a they would
5555 * generate empty chunks.
5557 if (session
->cleared_after_last_stop
) {
5558 DBG("Session \"%s\" was already cleared after stop, refusing rotation",
5560 cmd_ret
= LTTNG_ERR_ROTATION_AFTER_STOP_CLEAR
;
5564 if (session
->active
) {
5565 new_trace_chunk
= session_create_new_trace_chunk(session
, NULL
,
5567 if (!new_trace_chunk
) {
5568 cmd_ret
= LTTNG_ERR_CREATE_DIR_FAIL
;
5574 * The current trace chunk becomes the chunk being archived.
5576 * After this point, "chunk_being_archived" must absolutely
5577 * be closed on the consumer(s), otherwise it will never be
5578 * cleaned-up, which will result in a leak.
5580 ret
= session_set_trace_chunk(session
, new_trace_chunk
,
5581 &chunk_being_archived
);
5583 cmd_ret
= LTTNG_ERR_CREATE_TRACE_CHUNK_FAIL_CONSUMER
;
5587 if (session
->kernel_session
) {
5588 cmd_ret
= kernel_rotate_session(session
);
5589 if (cmd_ret
!= LTTNG_OK
) {
5590 failed_to_rotate
= true;
5591 rotation_fail_code
= cmd_ret
;
5594 if (session
->ust_session
) {
5595 cmd_ret
= ust_app_rotate_session(session
);
5596 if (cmd_ret
!= LTTNG_OK
) {
5597 failed_to_rotate
= true;
5598 rotation_fail_code
= cmd_ret
;
5602 if (!session
->active
) {
5603 session
->rotated_after_last_stop
= true;
5606 if (!chunk_being_archived
) {
5607 DBG("Rotating session \"%s\" from a \"NULL\" trace chunk to a new trace chunk, skipping completion check",
5609 if (failed_to_rotate
) {
5610 cmd_ret
= rotation_fail_code
;
5617 session
->rotation_state
= LTTNG_ROTATION_STATE_ONGOING
;
5618 chunk_status
= lttng_trace_chunk_get_id(chunk_being_archived
,
5619 &ongoing_rotation_chunk_id
);
5620 assert(chunk_status
== LTTNG_TRACE_CHUNK_STATUS_OK
);
5622 ret
= session_close_trace_chunk(session
, chunk_being_archived
,
5623 command
, session
->last_chunk_path
);
5625 cmd_ret
= LTTNG_ERR_CLOSE_TRACE_CHUNK_FAIL_CONSUMER
;
5629 if (failed_to_rotate
) {
5630 cmd_ret
= rotation_fail_code
;
5634 session
->quiet_rotation
= quiet_rotation
;
5635 ret
= timer_session_rotation_pending_check_start(session
,
5636 DEFAULT_ROTATE_PENDING_TIMER
);
5638 cmd_ret
= LTTNG_ERR_UNK
;
5642 if (rotate_return
) {
5643 rotate_return
->rotation_id
= ongoing_rotation_chunk_id
;
5646 session
->chunk_being_archived
= chunk_being_archived
;
5647 chunk_being_archived
= NULL
;
5648 if (!quiet_rotation
) {
5649 ret
= notification_thread_command_session_rotation_ongoing(
5650 the_notification_thread_handle
, session
->name
,
5651 session
->uid
, session
->gid
,
5652 ongoing_rotation_chunk_id
);
5653 if (ret
!= LTTNG_OK
) {
5654 ERR("Failed to notify notification thread that a session rotation is ongoing for session %s",
5660 DBG("Cmd rotate session %s, archive_id %" PRIu64
" sent",
5661 session
->name
, ongoing_rotation_chunk_id
);
5663 lttng_trace_chunk_put(new_trace_chunk
);
5664 lttng_trace_chunk_put(chunk_being_archived
);
5665 ret
= (cmd_ret
== LTTNG_OK
) ? cmd_ret
: -((int) cmd_ret
);
5668 if (session_reset_rotation_state(session
,
5669 LTTNG_ROTATION_STATE_ERROR
)) {
5670 ERR("Failed to reset rotation state of session \"%s\"",
5677 * Command LTTNG_ROTATION_GET_INFO from the lttng-ctl library.
5679 * Check if the session has finished its rotation.
5681 * Return LTTNG_OK on success or else an LTTNG_ERR code.
5683 int cmd_rotate_get_info(struct ltt_session
*session
,
5684 struct lttng_rotation_get_info_return
*info_return
,
5685 uint64_t rotation_id
)
5687 enum lttng_error_code cmd_ret
= LTTNG_OK
;
5688 enum lttng_rotation_state rotation_state
;
5690 DBG("Cmd rotate_get_info session %s, rotation id %" PRIu64
, session
->name
,
5691 session
->most_recent_chunk_id
.value
);
5693 if (session
->chunk_being_archived
) {
5694 enum lttng_trace_chunk_status chunk_status
;
5697 chunk_status
= lttng_trace_chunk_get_id(
5698 session
->chunk_being_archived
,
5700 assert(chunk_status
== LTTNG_TRACE_CHUNK_STATUS_OK
);
5702 rotation_state
= rotation_id
== chunk_id
?
5703 LTTNG_ROTATION_STATE_ONGOING
:
5704 LTTNG_ROTATION_STATE_EXPIRED
;
5706 if (session
->last_archived_chunk_id
.is_set
&&
5707 rotation_id
!= session
->last_archived_chunk_id
.value
) {
5708 rotation_state
= LTTNG_ROTATION_STATE_EXPIRED
;
5710 rotation_state
= session
->rotation_state
;
5714 switch (rotation_state
) {
5715 case LTTNG_ROTATION_STATE_NO_ROTATION
:
5716 DBG("Reporting that no rotation has occurred within the lifetime of session \"%s\"",
5719 case LTTNG_ROTATION_STATE_EXPIRED
:
5720 DBG("Reporting that the rotation state of rotation id %" PRIu64
" of session \"%s\" has expired",
5721 rotation_id
, session
->name
);
5723 case LTTNG_ROTATION_STATE_ONGOING
:
5724 DBG("Reporting that rotation id %" PRIu64
" of session \"%s\" is still pending",
5725 rotation_id
, session
->name
);
5727 case LTTNG_ROTATION_STATE_COMPLETED
:
5731 char *current_tracing_path_reply
;
5732 size_t current_tracing_path_reply_len
;
5734 DBG("Reporting that rotation id %" PRIu64
" of session \"%s\" is completed",
5735 rotation_id
, session
->name
);
5737 switch (session_get_consumer_destination_type(session
)) {
5738 case CONSUMER_DST_LOCAL
:
5739 current_tracing_path_reply
=
5740 info_return
->location
.local
.absolute_path
;
5741 current_tracing_path_reply_len
=
5742 sizeof(info_return
->location
.local
.absolute_path
);
5743 info_return
->location_type
=
5744 (int8_t) LTTNG_TRACE_ARCHIVE_LOCATION_TYPE_LOCAL
;
5745 fmt_ret
= asprintf(&chunk_path
,
5746 "%s/" DEFAULT_ARCHIVED_TRACE_CHUNKS_DIRECTORY
"/%s",
5747 session_get_base_path(session
),
5748 session
->last_archived_chunk_name
);
5749 if (fmt_ret
== -1) {
5750 PERROR("Failed to format the path of the last archived trace chunk");
5751 info_return
->status
= LTTNG_ROTATION_STATUS_ERROR
;
5752 cmd_ret
= LTTNG_ERR_UNK
;
5756 case CONSUMER_DST_NET
:
5758 uint16_t ctrl_port
, data_port
;
5760 current_tracing_path_reply
=
5761 info_return
->location
.relay
.relative_path
;
5762 current_tracing_path_reply_len
=
5763 sizeof(info_return
->location
.relay
.relative_path
);
5764 /* Currently the only supported relay protocol. */
5765 info_return
->location
.relay
.protocol
=
5766 (int8_t) LTTNG_TRACE_ARCHIVE_LOCATION_RELAY_PROTOCOL_TYPE_TCP
;
5768 fmt_ret
= lttng_strncpy(info_return
->location
.relay
.host
,
5769 session_get_net_consumer_hostname(session
),
5770 sizeof(info_return
->location
.relay
.host
));
5772 ERR("Failed to copy host name to rotate_get_info reply");
5773 info_return
->status
= LTTNG_ROTATION_STATUS_ERROR
;
5774 cmd_ret
= LTTNG_ERR_SET_URL
;
5778 session_get_net_consumer_ports(session
, &ctrl_port
, &data_port
);
5779 info_return
->location
.relay
.ports
.control
= ctrl_port
;
5780 info_return
->location
.relay
.ports
.data
= data_port
;
5781 info_return
->location_type
=
5782 (int8_t) LTTNG_TRACE_ARCHIVE_LOCATION_TYPE_RELAY
;
5783 chunk_path
= strdup(session
->last_chunk_path
);
5785 ERR("Failed to allocate the path of the last archived trace chunk");
5786 info_return
->status
= LTTNG_ROTATION_STATUS_ERROR
;
5787 cmd_ret
= LTTNG_ERR_UNK
;
5796 fmt_ret
= lttng_strncpy(current_tracing_path_reply
,
5797 chunk_path
, current_tracing_path_reply_len
);
5800 ERR("Failed to copy path of the last archived trace chunk to rotate_get_info reply");
5801 info_return
->status
= LTTNG_ROTATION_STATUS_ERROR
;
5802 cmd_ret
= LTTNG_ERR_UNK
;
5808 case LTTNG_ROTATION_STATE_ERROR
:
5809 DBG("Reporting that an error occurred during rotation %" PRIu64
" of session \"%s\"",
5810 rotation_id
, session
->name
);
5818 info_return
->status
= (int32_t) rotation_state
;
5823 * Command LTTNG_ROTATION_SET_SCHEDULE from the lttng-ctl library.
5825 * Configure the automatic rotation parameters.
5826 * 'activate' to true means activate the rotation schedule type with 'new_value'.
5827 * 'activate' to false means deactivate the rotation schedule and validate that
5828 * 'new_value' has the same value as the currently active value.
5830 * Return LTTNG_OK on success or else a positive LTTNG_ERR code.
5832 int cmd_rotation_set_schedule(struct ltt_session
*session
,
5833 bool activate
, enum lttng_rotation_schedule_type schedule_type
,
5835 struct notification_thread_handle
*notification_thread_handle
)
5838 uint64_t *parameter_value
;
5842 DBG("Cmd rotate set schedule session %s", session
->name
);
5844 if (session
->live_timer
|| !session
->output_traces
) {
5845 DBG("Failing ROTATION_SET_SCHEDULE command as the rotation feature is not available for this session");
5846 ret
= LTTNG_ERR_ROTATION_NOT_AVAILABLE
;
5850 switch (schedule_type
) {
5851 case LTTNG_ROTATION_SCHEDULE_TYPE_SIZE_THRESHOLD
:
5852 parameter_value
= &session
->rotate_size
;
5854 case LTTNG_ROTATION_SCHEDULE_TYPE_PERIODIC
:
5855 parameter_value
= &session
->rotate_timer_period
;
5856 if (new_value
>= UINT_MAX
) {
5857 DBG("Failing ROTATION_SET_SCHEDULE command as the value requested for a periodic rotation schedule is invalid: %" PRIu64
" > %u (UINT_MAX)",
5858 new_value
, UINT_MAX
);
5859 ret
= LTTNG_ERR_INVALID
;
5864 WARN("Failing ROTATION_SET_SCHEDULE command on unknown schedule type");
5865 ret
= LTTNG_ERR_INVALID
;
5869 /* Improper use of the API. */
5870 if (new_value
== -1ULL) {
5871 WARN("Failing ROTATION_SET_SCHEDULE command as the value requested is -1");
5872 ret
= LTTNG_ERR_INVALID
;
5877 * As indicated in struct ltt_session's comments, a value of == 0 means
5878 * this schedule rotation type is not in use.
5880 * Reject the command if we were asked to activate a schedule that was
5883 if (activate
&& *parameter_value
!= 0) {
5884 DBG("Failing ROTATION_SET_SCHEDULE (activate) command as the schedule is already active");
5885 ret
= LTTNG_ERR_ROTATION_SCHEDULE_SET
;
5890 * Reject the command if we were asked to deactivate a schedule that was
5893 if (!activate
&& *parameter_value
== 0) {
5894 DBG("Failing ROTATION_SET_SCHEDULE (deactivate) command as the schedule is already inactive");
5895 ret
= LTTNG_ERR_ROTATION_SCHEDULE_NOT_SET
;
5900 * Reject the command if we were asked to deactivate a schedule that
5903 if (!activate
&& *parameter_value
!= new_value
) {
5904 DBG("Failing ROTATION_SET_SCHEDULE (deactivate) command as an inexistant schedule was provided");
5905 ret
= LTTNG_ERR_ROTATION_SCHEDULE_NOT_SET
;
5909 *parameter_value
= activate
? new_value
: 0;
5911 switch (schedule_type
) {
5912 case LTTNG_ROTATION_SCHEDULE_TYPE_PERIODIC
:
5913 if (activate
&& session
->active
) {
5915 * Only start the timer if the session is active,
5916 * otherwise it will be started when the session starts.
5918 ret
= timer_session_rotation_schedule_timer_start(
5919 session
, new_value
);
5921 ERR("Failed to enable session rotation timer in ROTATION_SET_SCHEDULE command");
5922 ret
= LTTNG_ERR_UNK
;
5926 ret
= timer_session_rotation_schedule_timer_stop(
5929 ERR("Failed to disable session rotation timer in ROTATION_SET_SCHEDULE command");
5930 ret
= LTTNG_ERR_UNK
;
5935 case LTTNG_ROTATION_SCHEDULE_TYPE_SIZE_THRESHOLD
:
5937 ret
= subscribe_session_consumed_size_rotation(session
,
5938 new_value
, notification_thread_handle
);
5940 ERR("Failed to enable consumed-size notification in ROTATION_SET_SCHEDULE command");
5941 ret
= LTTNG_ERR_UNK
;
5945 ret
= unsubscribe_session_consumed_size_rotation(session
,
5946 notification_thread_handle
);
5948 ERR("Failed to disable consumed-size notification in ROTATION_SET_SCHEDULE command");
5949 ret
= LTTNG_ERR_UNK
;
5956 /* Would have been caught before. */
5968 /* Wait for a given path to be removed before continuing. */
5969 static enum lttng_error_code
wait_on_path(void *path_data
)
5971 const char *shm_path
= path_data
;
5973 DBG("Waiting for the shm path at %s to be removed before completing session destruction",
5979 ret
= stat(shm_path
, &st
);
5981 if (errno
!= ENOENT
) {
5982 PERROR("stat() returned an error while checking for the existence of the shm path");
5984 DBG("shm path no longer exists, completing the destruction of session");
5988 if (!S_ISDIR(st
.st_mode
)) {
5989 ERR("The type of shm path %s returned by stat() is not a directory; aborting the wait for shm path removal",
5994 usleep(SESSION_DESTROY_SHM_PATH_CHECK_DELAY_US
);
6000 * Returns a pointer to a handler to run on completion of a command.
6001 * Returns NULL if no handler has to be run for the last command executed.
6003 const struct cmd_completion_handler
*cmd_pop_completion_handler(void)
6005 struct cmd_completion_handler
*handler
= current_completion_handler
;
6007 current_completion_handler
= NULL
;
6012 * Init command subsystem.
6017 * Set network sequence index to 1 for streams to match a relayd
6018 * socket on the consumer side.
6020 pthread_mutex_lock(&relayd_net_seq_idx_lock
);
6021 relayd_net_seq_idx
= 1;
6022 pthread_mutex_unlock(&relayd_net_seq_idx_lock
);
6024 DBG("Command subsystem initialized");