4 * Holds LTTng per-session event registry.
6 * Copyright (C) 2010-2012 Mathieu Desnoyers <mathieu.desnoyers@efficios.com>
8 * This library is free software; you can redistribute it and/or
9 * modify it under the terms of the GNU Lesser General Public
10 * License as published by the Free Software Foundation; only
11 * version 2.1 of the License.
13 * This library is distributed in the hope that it will be useful,
14 * but WITHOUT ANY WARRANTY; without even the implied warranty of
15 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
16 * Lesser General Public License for more details.
18 * You should have received a copy of the GNU Lesser General Public
19 * License along with this library; if not, write to the Free Software
20 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
25 #include <urcu/list.h>
26 #include <urcu/hlist.h>
35 #include <lttng/ust-endian.h>
39 #include <urcu/compiler.h>
40 #include <urcu/uatomic.h>
41 #include <urcu/arch.h>
43 #include <lttng/tracepoint.h>
44 #include <lttng/ust-events.h>
46 #include <usterr-signal-safe.h>
48 #include <lttng/ust-ctl.h>
52 #include "lttng-ust-uuid.h"
54 #include "tracepoint-internal.h"
55 #include "lttng-tracer.h"
56 #include "lttng-tracer-core.h"
57 #include "lttng-ust-statedump.h"
59 #include "../libringbuffer/shm.h"
63 * All operations within this file are called by the communication
64 * thread, under ust_lock protection.
67 static CDS_LIST_HEAD(sessions
);
69 struct cds_list_head
*_lttng_get_sessions(void)
74 static void _lttng_event_destroy(struct lttng_event
*event
);
77 void lttng_session_lazy_sync_enablers(struct lttng_session
*session
);
79 void lttng_session_sync_enablers(struct lttng_session
*session
);
81 void lttng_enabler_destroy(struct lttng_enabler
*enabler
);
84 * Called with ust lock held.
86 int lttng_session_active(void)
88 struct lttng_session
*iter
;
90 cds_list_for_each_entry(iter
, &sessions
, node
) {
98 int lttng_loglevel_match(int loglevel
,
99 unsigned int has_loglevel
,
100 enum lttng_ust_loglevel_type req_type
,
104 loglevel
= TRACE_DEFAULT
;
106 case LTTNG_UST_LOGLEVEL_RANGE
:
107 if (loglevel
<= req_loglevel
108 || (req_loglevel
== -1 && loglevel
<= TRACE_DEBUG
))
112 case LTTNG_UST_LOGLEVEL_SINGLE
:
113 if (loglevel
== req_loglevel
114 || (req_loglevel
== -1 && loglevel
<= TRACE_DEBUG
))
118 case LTTNG_UST_LOGLEVEL_ALL
:
120 if (loglevel
<= TRACE_DEBUG
)
127 void synchronize_trace(void)
132 struct lttng_session
*lttng_session_create(void)
134 struct lttng_session
*session
;
137 session
= zmalloc(sizeof(struct lttng_session
));
140 CDS_INIT_LIST_HEAD(&session
->chan_head
);
141 CDS_INIT_LIST_HEAD(&session
->events_head
);
142 CDS_INIT_LIST_HEAD(&session
->enablers_head
);
143 for (i
= 0; i
< LTTNG_UST_EVENT_HT_SIZE
; i
++)
144 CDS_INIT_HLIST_HEAD(&session
->events_ht
.table
[i
]);
145 cds_list_add(&session
->node
, &sessions
);
150 * Only used internally at session destruction.
153 void _lttng_channel_unmap(struct lttng_channel
*lttng_chan
)
155 struct channel
*chan
;
156 struct lttng_ust_shm_handle
*handle
;
158 cds_list_del(<tng_chan
->node
);
159 lttng_destroy_context(lttng_chan
->ctx
);
160 chan
= lttng_chan
->chan
;
161 handle
= lttng_chan
->handle
;
163 * note: lttng_chan is private data contained within handle. It
164 * will be freed along with the handle.
166 channel_destroy(chan
, handle
, 0);
170 void register_event(struct lttng_event
*event
)
173 const struct lttng_event_desc
*desc
;
175 assert(event
->registered
== 0);
177 ret
= __tracepoint_probe_register_queue_release(desc
->name
,
178 desc
->probe_callback
,
179 event
, desc
->signature
);
182 event
->registered
= 1;
186 void unregister_event(struct lttng_event
*event
)
189 const struct lttng_event_desc
*desc
;
191 assert(event
->registered
== 1);
193 ret
= __tracepoint_probe_unregister_queue_release(desc
->name
,
194 desc
->probe_callback
,
198 event
->registered
= 0;
202 * Only used internally at session destruction.
205 void _lttng_event_unregister(struct lttng_event
*event
)
207 if (event
->registered
)
208 unregister_event(event
);
211 void lttng_session_destroy(struct lttng_session
*session
)
213 struct lttng_channel
*chan
, *tmpchan
;
214 struct lttng_event
*event
, *tmpevent
;
215 struct lttng_enabler
*enabler
, *tmpenabler
;
217 CMM_ACCESS_ONCE(session
->active
) = 0;
218 cds_list_for_each_entry(event
, &session
->events_head
, node
) {
219 _lttng_event_unregister(event
);
221 synchronize_trace(); /* Wait for in-flight events to complete */
222 __tracepoint_probe_prune_release_queue();
223 cds_list_for_each_entry_safe(enabler
, tmpenabler
,
224 &session
->enablers_head
, node
)
225 lttng_enabler_destroy(enabler
);
226 cds_list_for_each_entry_safe(event
, tmpevent
,
227 &session
->events_head
, node
)
228 _lttng_event_destroy(event
);
229 cds_list_for_each_entry_safe(chan
, tmpchan
, &session
->chan_head
, node
)
230 _lttng_channel_unmap(chan
);
231 cds_list_del(&session
->node
);
235 int lttng_session_enable(struct lttng_session
*session
)
238 struct lttng_channel
*chan
;
241 if (session
->active
) {
246 notify_socket
= lttng_get_notify_socket(session
->owner
);
247 if (notify_socket
< 0)
248 return notify_socket
;
250 /* Set transient enabler state to "enabled" */
254 * Snapshot the number of events per channel to know the type of header
257 cds_list_for_each_entry(chan
, &session
->chan_head
, node
) {
258 const struct lttng_ctx
*ctx
;
259 const struct lttng_ctx_field
*fields
= NULL
;
260 size_t nr_fields
= 0;
263 /* don't change it if session stop/restart */
264 if (chan
->header_type
)
268 nr_fields
= ctx
->nr_fields
;
269 fields
= ctx
->fields
;
271 ret
= ustcomm_register_channel(notify_socket
,
279 DBG("Error (%d) registering channel to sessiond", ret
);
282 if (chan_id
!= chan
->id
) {
283 DBG("Error: channel registration id (%u) does not match id assigned at creation (%u)",
289 /* We need to sync enablers with session before activation. */
290 lttng_session_sync_enablers(session
);
292 /* Set atomically the state to "active" */
293 CMM_ACCESS_ONCE(session
->active
) = 1;
294 CMM_ACCESS_ONCE(session
->been_active
) = 1;
296 session
->statedump_pending
= 1;
297 lttng_ust_sockinfo_session_enabled(session
->owner
);
302 int lttng_session_disable(struct lttng_session
*session
)
306 if (!session
->active
) {
310 /* Set atomically the state to "inactive" */
311 CMM_ACCESS_ONCE(session
->active
) = 0;
313 /* Set transient enabler state to "disabled" */
315 lttng_session_sync_enablers(session
);
320 int lttng_channel_enable(struct lttng_channel
*channel
)
324 if (channel
->enabled
) {
328 /* Set transient enabler state to "enabled" */
330 lttng_session_sync_enablers(channel
->session
);
331 /* Set atomically the state to "enabled" */
332 CMM_ACCESS_ONCE(channel
->enabled
) = 1;
337 int lttng_channel_disable(struct lttng_channel
*channel
)
341 if (!channel
->enabled
) {
345 /* Set atomically the state to "disabled" */
346 CMM_ACCESS_ONCE(channel
->enabled
) = 0;
347 /* Set transient enabler state to "enabled" */
349 lttng_session_sync_enablers(channel
->session
);
355 * Supports event creation while tracing session is active.
358 int lttng_event_create(const struct lttng_event_desc
*desc
,
359 struct lttng_channel
*chan
)
361 const char *event_name
= desc
->name
;
362 struct lttng_event
*event
;
363 struct lttng_session
*session
= chan
->session
;
364 struct cds_hlist_head
*head
;
365 struct cds_hlist_node
*node
;
367 size_t name_len
= strlen(event_name
);
369 int notify_socket
, loglevel
;
372 hash
= jhash(event_name
, name_len
, 0);
373 head
= &chan
->session
->events_ht
.table
[hash
& (LTTNG_UST_EVENT_HT_SIZE
- 1)];
374 cds_hlist_for_each_entry(event
, node
, head
, hlist
) {
376 if (!strncmp(event
->desc
->name
, desc
->name
,
377 LTTNG_UST_SYM_NAME_LEN
- 1)
378 && chan
== event
->chan
) {
384 notify_socket
= lttng_get_notify_socket(session
->owner
);
385 if (notify_socket
< 0) {
391 * Check if loglevel match. Refuse to connect event if not.
393 event
= zmalloc(sizeof(struct lttng_event
));
400 /* Event will be enabled by enabler sync. */
402 event
->registered
= 0;
403 CDS_INIT_LIST_HEAD(&event
->bytecode_runtime_head
);
404 CDS_INIT_LIST_HEAD(&event
->enablers_ref_head
);
408 loglevel
= *(*event
->desc
->loglevel
);
410 loglevel
= TRACE_DEFAULT
;
411 if (desc
->u
.ext
.model_emf_uri
)
412 uri
= *(desc
->u
.ext
.model_emf_uri
);
416 /* Fetch event ID from sessiond */
417 ret
= ustcomm_register_event(notify_socket
,
428 DBG("Error (%d) registering event to sessiond", ret
);
429 goto sessiond_register_error
;
432 /* Populate lttng_event structure before tracepoint registration. */
434 cds_list_add(&event
->node
, &chan
->session
->events_head
);
435 cds_hlist_add_head(&event
->hlist
, head
);
438 sessiond_register_error
:
447 int lttng_desc_match_wildcard_enabler(const struct lttng_event_desc
*desc
,
448 struct lttng_enabler
*enabler
)
451 unsigned int has_loglevel
= 0;
453 assert(enabler
->type
== LTTNG_ENABLER_WILDCARD
);
454 /* Compare excluding final '*' */
455 if (strncmp(desc
->name
, enabler
->event_param
.name
,
456 strlen(enabler
->event_param
.name
) - 1))
458 if (desc
->loglevel
) {
459 loglevel
= *(*desc
->loglevel
);
462 if (!lttng_loglevel_match(loglevel
,
464 enabler
->event_param
.loglevel_type
,
465 enabler
->event_param
.loglevel
))
471 int lttng_desc_match_event_enabler(const struct lttng_event_desc
*desc
,
472 struct lttng_enabler
*enabler
)
475 unsigned int has_loglevel
= 0;
477 assert(enabler
->type
== LTTNG_ENABLER_EVENT
);
478 if (strcmp(desc
->name
, enabler
->event_param
.name
))
480 if (desc
->loglevel
) {
481 loglevel
= *(*desc
->loglevel
);
484 if (!lttng_loglevel_match(loglevel
,
486 enabler
->event_param
.loglevel_type
,
487 enabler
->event_param
.loglevel
))
493 int lttng_desc_match_enabler(const struct lttng_event_desc
*desc
,
494 struct lttng_enabler
*enabler
)
496 struct lttng_ust_excluder_node
*excluder
;
498 /* If event matches with an excluder, return 'does not match' */
499 cds_list_for_each_entry(excluder
, &enabler
->excluder_head
, node
) {
502 for (count
= 0; count
< excluder
->excluder
.count
; count
++) {
506 excluder_name
= (char *) (excluder
->excluder
.names
)
507 + count
* LTTNG_UST_SYM_NAME_LEN
;
508 len
= strnlen(excluder_name
, LTTNG_UST_SYM_NAME_LEN
);
509 if (len
> 0 && excluder_name
[len
- 1] == '*') {
510 found
= !strncmp(desc
->name
, excluder_name
,
513 found
= !strncmp(desc
->name
, excluder_name
,
514 LTTNG_UST_SYM_NAME_LEN
- 1);
521 switch (enabler
->type
) {
522 case LTTNG_ENABLER_WILDCARD
:
523 return lttng_desc_match_wildcard_enabler(desc
, enabler
);
524 case LTTNG_ENABLER_EVENT
:
525 return lttng_desc_match_event_enabler(desc
, enabler
);
532 int lttng_event_match_enabler(struct lttng_event
*event
,
533 struct lttng_enabler
*enabler
)
535 if (lttng_desc_match_enabler(event
->desc
, enabler
)
536 && event
->chan
== enabler
->chan
)
543 struct lttng_enabler_ref
* lttng_event_enabler_ref(struct lttng_event
*event
,
544 struct lttng_enabler
*enabler
)
546 struct lttng_enabler_ref
*enabler_ref
;
548 cds_list_for_each_entry(enabler_ref
,
549 &event
->enablers_ref_head
, node
) {
550 if (enabler_ref
->ref
== enabler
)
557 * Create struct lttng_event if it is missing and present in the list of
561 void lttng_create_event_if_missing(struct lttng_enabler
*enabler
)
563 struct lttng_session
*session
= enabler
->chan
->session
;
564 struct lttng_probe_desc
*probe_desc
;
565 const struct lttng_event_desc
*desc
;
566 struct lttng_event
*event
;
568 struct cds_list_head
*probe_list
;
570 probe_list
= lttng_get_probe_list_head();
572 * For each probe event, if we find that a probe event matches
573 * our enabler, create an associated lttng_event if not
576 cds_list_for_each_entry(probe_desc
, probe_list
, head
) {
577 for (i
= 0; i
< probe_desc
->nr_events
; i
++) {
579 struct cds_hlist_head
*head
;
580 struct cds_hlist_node
*node
;
581 const char *event_name
;
585 desc
= probe_desc
->event_desc
[i
];
586 if (!lttng_desc_match_enabler(desc
, enabler
))
588 event_name
= desc
->name
;
589 name_len
= strlen(event_name
);
592 * Check if already created.
594 hash
= jhash(event_name
, name_len
, 0);
595 head
= &session
->events_ht
.table
[hash
& (LTTNG_UST_EVENT_HT_SIZE
- 1)];
596 cds_hlist_for_each_entry(event
, node
, head
, hlist
) {
597 if (event
->desc
== desc
598 && event
->chan
== enabler
->chan
)
605 * We need to create an event for this
608 ret
= lttng_event_create(probe_desc
->event_desc
[i
],
611 DBG("Unable to create event %s, error %d\n",
612 probe_desc
->event_desc
[i
]->name
, ret
);
619 * Create events associated with an enabler (if not already present),
620 * and add backward reference from the event to the enabler.
623 int lttng_enabler_ref_events(struct lttng_enabler
*enabler
)
625 struct lttng_session
*session
= enabler
->chan
->session
;
626 struct lttng_event
*event
;
628 /* First ensure that probe events are created for this enabler. */
629 lttng_create_event_if_missing(enabler
);
631 /* For each event matching enabler in session event list. */
632 cds_list_for_each_entry(event
, &session
->events_head
, node
) {
633 struct lttng_enabler_ref
*enabler_ref
;
635 if (!lttng_event_match_enabler(event
, enabler
))
638 enabler_ref
= lttng_event_enabler_ref(event
, enabler
);
641 * If no backward ref, create it.
642 * Add backward ref from event to enabler.
644 enabler_ref
= zmalloc(sizeof(*enabler_ref
));
647 enabler_ref
->ref
= enabler
;
648 cds_list_add(&enabler_ref
->node
,
649 &event
->enablers_ref_head
);
653 * Link filter bytecodes if not linked yet.
655 lttng_enabler_event_link_bytecode(event
, enabler
);
657 /* TODO: merge event context. */
663 * Called at library load: connect the probe on all enablers matching
665 * Called with session mutex held.
667 int lttng_fix_pending_events(void)
669 struct lttng_session
*session
;
671 cds_list_for_each_entry(session
, &sessions
, node
) {
672 lttng_session_lazy_sync_enablers(session
);
678 * For each session of the owner thread, execute pending statedump.
679 * Only dump state for the sessions owned by the caller thread, because
680 * we don't keep ust_lock across the entire iteration.
682 void lttng_handle_pending_statedump(void *owner
)
684 struct lttng_session
*session
;
686 /* Execute state dump */
687 do_lttng_ust_statedump(owner
);
689 /* Clear pending state dump */
693 cds_list_for_each_entry(session
, &sessions
, node
) {
694 if (session
->owner
!= owner
)
696 if (!session
->statedump_pending
)
698 session
->statedump_pending
= 0;
706 * Only used internally at session destruction.
709 void _lttng_event_destroy(struct lttng_event
*event
)
711 struct lttng_enabler_ref
*enabler_ref
, *tmp_enabler_ref
;
713 cds_list_del(&event
->node
);
714 lttng_destroy_context(event
->ctx
);
715 lttng_free_event_filter_runtime(event
);
716 /* Free event enabler refs */
717 cds_list_for_each_entry_safe(enabler_ref
, tmp_enabler_ref
,
718 &event
->enablers_ref_head
, node
)
723 void lttng_ust_events_exit(void)
725 struct lttng_session
*session
, *tmpsession
;
727 cds_list_for_each_entry_safe(session
, tmpsession
, &sessions
, node
)
728 lttng_session_destroy(session
);
732 * Enabler management.
734 struct lttng_enabler
*lttng_enabler_create(enum lttng_enabler_type type
,
735 struct lttng_ust_event
*event_param
,
736 struct lttng_channel
*chan
)
738 struct lttng_enabler
*enabler
;
740 enabler
= zmalloc(sizeof(*enabler
));
743 enabler
->type
= type
;
744 CDS_INIT_LIST_HEAD(&enabler
->filter_bytecode_head
);
745 CDS_INIT_LIST_HEAD(&enabler
->excluder_head
);
746 memcpy(&enabler
->event_param
, event_param
,
747 sizeof(enabler
->event_param
));
748 enabler
->chan
= chan
;
750 enabler
->enabled
= 0;
751 cds_list_add(&enabler
->node
, &enabler
->chan
->session
->enablers_head
);
752 lttng_session_lazy_sync_enablers(enabler
->chan
->session
);
756 int lttng_enabler_enable(struct lttng_enabler
*enabler
)
758 enabler
->enabled
= 1;
759 lttng_session_lazy_sync_enablers(enabler
->chan
->session
);
763 int lttng_enabler_disable(struct lttng_enabler
*enabler
)
765 enabler
->enabled
= 0;
766 lttng_session_lazy_sync_enablers(enabler
->chan
->session
);
770 int lttng_enabler_attach_bytecode(struct lttng_enabler
*enabler
,
771 struct lttng_ust_filter_bytecode_node
*bytecode
)
773 bytecode
->enabler
= enabler
;
774 cds_list_add_tail(&bytecode
->node
, &enabler
->filter_bytecode_head
);
775 lttng_session_lazy_sync_enablers(enabler
->chan
->session
);
779 int lttng_enabler_attach_exclusion(struct lttng_enabler
*enabler
,
780 struct lttng_ust_excluder_node
*excluder
)
782 excluder
->enabler
= enabler
;
783 cds_list_add_tail(&excluder
->node
, &enabler
->excluder_head
);
784 lttng_session_lazy_sync_enablers(enabler
->chan
->session
);
788 int lttng_attach_context(struct lttng_ust_context
*context_param
,
789 struct lttng_ctx
**ctx
, struct lttng_session
*session
)
792 * We cannot attach a context after trace has been started for a
793 * session because the metadata does not allow expressing this
794 * information outside of the original channel scope.
796 if (session
->been_active
)
799 switch (context_param
->ctx
) {
800 case LTTNG_UST_CONTEXT_PTHREAD_ID
:
801 return lttng_add_pthread_id_to_ctx(ctx
);
802 case LTTNG_UST_CONTEXT_PERF_THREAD_COUNTER
:
804 struct lttng_ust_perf_counter_ctx
*perf_ctx_param
;
806 perf_ctx_param
= &context_param
->u
.perf_counter
;
807 return lttng_add_perf_counter_to_ctx(
808 perf_ctx_param
->type
,
809 perf_ctx_param
->config
,
810 perf_ctx_param
->name
,
813 case LTTNG_UST_CONTEXT_VTID
:
814 return lttng_add_vtid_to_ctx(ctx
);
815 case LTTNG_UST_CONTEXT_VPID
:
816 return lttng_add_vpid_to_ctx(ctx
);
817 case LTTNG_UST_CONTEXT_PROCNAME
:
818 return lttng_add_procname_to_ctx(ctx
);
819 case LTTNG_UST_CONTEXT_IP
:
820 return lttng_add_ip_to_ctx(ctx
);
821 case LTTNG_UST_CONTEXT_CPU_ID
:
822 return lttng_add_cpu_id_to_ctx(ctx
);
828 int lttng_enabler_attach_context(struct lttng_enabler
*enabler
,
829 struct lttng_ust_context
*context_param
)
831 #if 0 // disabled for now.
832 struct lttng_session
*session
= enabler
->chan
->session
;
835 ret
= lttng_attach_context(context_param
, &enabler
->ctx
,
839 lttng_session_lazy_sync_enablers(enabler
->chan
->session
);
845 void lttng_enabler_destroy(struct lttng_enabler
*enabler
)
847 struct lttng_ust_filter_bytecode_node
*filter_node
, *tmp_filter_node
;
848 struct lttng_ust_excluder_node
*excluder_node
, *tmp_excluder_node
;
850 /* Destroy filter bytecode */
851 cds_list_for_each_entry_safe(filter_node
, tmp_filter_node
,
852 &enabler
->filter_bytecode_head
, node
) {
856 /* Destroy excluders */
857 cds_list_for_each_entry_safe(excluder_node
, tmp_excluder_node
,
858 &enabler
->excluder_head
, node
) {
862 /* Destroy contexts */
863 lttng_destroy_context(enabler
->ctx
);
865 cds_list_del(&enabler
->node
);
870 * lttng_session_sync_enablers should be called just before starting a
874 void lttng_session_sync_enablers(struct lttng_session
*session
)
876 struct lttng_enabler
*enabler
;
877 struct lttng_event
*event
;
879 cds_list_for_each_entry(enabler
, &session
->enablers_head
, node
)
880 lttng_enabler_ref_events(enabler
);
882 * For each event, if at least one of its enablers is enabled,
883 * and its channel and session transient states are enabled, we
884 * enable the event, else we disable it.
886 cds_list_for_each_entry(event
, &session
->events_head
, node
) {
887 struct lttng_enabler_ref
*enabler_ref
;
888 struct lttng_bytecode_runtime
*runtime
;
889 int enabled
= 0, has_enablers_without_bytecode
= 0;
892 cds_list_for_each_entry(enabler_ref
,
893 &event
->enablers_ref_head
, node
) {
894 if (enabler_ref
->ref
->enabled
) {
900 * Enabled state is based on union of enablers, with
901 * intesection of session and channel transient enable
904 enabled
= enabled
&& session
->tstate
&& event
->chan
->tstate
;
906 CMM_STORE_SHARED(event
->enabled
, enabled
);
908 * Sync tracepoint registration with event enabled
912 if (!event
->registered
)
913 register_event(event
);
915 if (event
->registered
)
916 unregister_event(event
);
919 /* Check if has enablers without bytecode enabled */
920 cds_list_for_each_entry(enabler_ref
,
921 &event
->enablers_ref_head
, node
) {
922 if (enabler_ref
->ref
->enabled
923 && cds_list_empty(&enabler_ref
->ref
->filter_bytecode_head
)) {
924 has_enablers_without_bytecode
= 1;
928 event
->has_enablers_without_bytecode
=
929 has_enablers_without_bytecode
;
932 cds_list_for_each_entry(runtime
,
933 &event
->bytecode_runtime_head
, node
) {
934 lttng_filter_sync_state(runtime
);
937 __tracepoint_probe_prune_release_queue();
941 * Apply enablers to session events, adding events to session if need
942 * be. It is required after each modification applied to an active
943 * session, and right before session "start".
944 * "lazy" sync means we only sync if required.
947 void lttng_session_lazy_sync_enablers(struct lttng_session
*session
)
949 /* We can skip if session is not active */
950 if (!session
->active
)
952 lttng_session_sync_enablers(session
);