a84316a9358caeda6be0138a439c3828dd979b8e
[lttng-ust.git] / liblttng-ust / lttng-ust-comm.c
1 /*
2 * lttng-ust-comm.c
3 *
4 * Copyright (C) 2011 David Goulet <david.goulet@polymtl.ca>
5 * Copyright (C) 2011 Mathieu Desnoyers <mathieu.desnoyers@efficios.com>
6 *
7 * This library is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU Lesser General Public
9 * License as published by the Free Software Foundation; only
10 * version 2.1 of the License.
11 *
12 * This library is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15 * Lesser General Public License for more details.
16 *
17 * You should have received a copy of the GNU Lesser General Public
18 * License along with this library; if not, write to the Free Software
19 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
20 */
21
22 #define _LGPL_SOURCE
23 #define _GNU_SOURCE
24 #include <sys/types.h>
25 #include <sys/socket.h>
26 #include <sys/mman.h>
27 #include <sys/stat.h>
28 #include <sys/types.h>
29 #include <sys/wait.h>
30 #include <fcntl.h>
31 #include <unistd.h>
32 #include <errno.h>
33 #include <pthread.h>
34 #include <semaphore.h>
35 #include <time.h>
36 #include <assert.h>
37 #include <signal.h>
38 #include <urcu/uatomic.h>
39 #include <urcu/futex.h>
40 #include <urcu/compiler.h>
41
42 #include <lttng/ust-events.h>
43 #include <lttng/ust-abi.h>
44 #include <lttng/ust.h>
45 #include <lttng/ust-error.h>
46 #include <lttng/ust-ctl.h>
47 #include <urcu/tls-compat.h>
48 #include <ust-comm.h>
49 #include <usterr-signal-safe.h>
50 #include <helper.h>
51 #include "tracepoint-internal.h"
52 #include "lttng-tracer-core.h"
53 #include "compat.h"
54 #include "../libringbuffer/tlsfixup.h"
55 #include "lttng-ust-statedump.h"
56 #include "clock.h"
57 #include "../libringbuffer/getcpu.h"
58 #include "getenv.h"
59
60 /*
61 * Has lttng ust comm constructor been called ?
62 */
63 static int initialized;
64
65 /*
66 * The ust_lock/ust_unlock lock is used as a communication thread mutex.
67 * Held when handling a command, also held by fork() to deal with
68 * removal of threads, and by exit path.
69 *
70 * The UST lock is the centralized mutex across UST tracing control and
71 * probe registration.
72 *
73 * ust_exit_mutex must never nest in ust_mutex.
74 *
75 * ust_fork_mutex must never nest in ust_mutex.
76 *
77 * ust_mutex_nest is a per-thread nesting counter, allowing the perf
78 * counter lazy initialization called by events within the statedump,
79 * which traces while the ust_mutex is held.
80 *
81 * ust_lock nests within the dynamic loader lock (within glibc) because
82 * it is taken within the library constructor.
83 */
84 static pthread_mutex_t ust_mutex = PTHREAD_MUTEX_INITIALIZER;
85
86 /* Allow nesting the ust_mutex within the same thread. */
87 static DEFINE_URCU_TLS(int, ust_mutex_nest);
88
89 /*
90 * ust_exit_mutex protects thread_active variable wrt thread exit. It
91 * cannot be done by ust_mutex because pthread_cancel(), which takes an
92 * internal libc lock, cannot nest within ust_mutex.
93 *
94 * It never nests within a ust_mutex.
95 */
96 static pthread_mutex_t ust_exit_mutex = PTHREAD_MUTEX_INITIALIZER;
97
98 /*
99 * ust_fork_mutex protects base address statedump tracing against forks. It
100 * prevents the dynamic loader lock to be taken (by base address statedump
101 * tracing) while a fork is happening, thus preventing deadlock issues with
102 * the dynamic loader lock.
103 */
104 static pthread_mutex_t ust_fork_mutex = PTHREAD_MUTEX_INITIALIZER;
105
106 /* Should the ust comm thread quit ? */
107 static int lttng_ust_comm_should_quit;
108
109 /*
110 * This variable can be tested by applications to check whether
111 * lttng-ust is loaded. They simply have to define their own
112 * "lttng_ust_loaded" weak symbol, and test it. It is set to 1 by the
113 * library constructor.
114 */
115 int lttng_ust_loaded __attribute__((weak));
116
117 /*
118 * Return 0 on success, -1 if should quit.
119 * The lock is taken in both cases.
120 * Signal-safe.
121 */
122 int ust_lock(void)
123 {
124 sigset_t sig_all_blocked, orig_mask;
125 int ret, oldstate;
126
127 ret = pthread_setcancelstate(PTHREAD_CANCEL_DISABLE, &oldstate);
128 if (ret) {
129 ERR("pthread_setcancelstate: %s", strerror(ret));
130 }
131 if (oldstate != PTHREAD_CANCEL_ENABLE) {
132 ERR("pthread_setcancelstate: unexpected oldstate");
133 }
134 sigfillset(&sig_all_blocked);
135 ret = pthread_sigmask(SIG_SETMASK, &sig_all_blocked, &orig_mask);
136 if (ret) {
137 ERR("pthread_sigmask: %s", strerror(ret));
138 }
139 if (!URCU_TLS(ust_mutex_nest)++)
140 pthread_mutex_lock(&ust_mutex);
141 ret = pthread_sigmask(SIG_SETMASK, &orig_mask, NULL);
142 if (ret) {
143 ERR("pthread_sigmask: %s", strerror(ret));
144 }
145 if (lttng_ust_comm_should_quit) {
146 return -1;
147 } else {
148 return 0;
149 }
150 }
151
152 /*
153 * ust_lock_nocheck() can be used in constructors/destructors, because
154 * they are already nested within the dynamic loader lock, and therefore
155 * have exclusive access against execution of liblttng-ust destructor.
156 * Signal-safe.
157 */
158 void ust_lock_nocheck(void)
159 {
160 sigset_t sig_all_blocked, orig_mask;
161 int ret, oldstate;
162
163 ret = pthread_setcancelstate(PTHREAD_CANCEL_DISABLE, &oldstate);
164 if (ret) {
165 ERR("pthread_setcancelstate: %s", strerror(ret));
166 }
167 if (oldstate != PTHREAD_CANCEL_ENABLE) {
168 ERR("pthread_setcancelstate: unexpected oldstate");
169 }
170 sigfillset(&sig_all_blocked);
171 ret = pthread_sigmask(SIG_SETMASK, &sig_all_blocked, &orig_mask);
172 if (ret) {
173 ERR("pthread_sigmask: %s", strerror(ret));
174 }
175 if (!URCU_TLS(ust_mutex_nest)++)
176 pthread_mutex_lock(&ust_mutex);
177 ret = pthread_sigmask(SIG_SETMASK, &orig_mask, NULL);
178 if (ret) {
179 ERR("pthread_sigmask: %s", strerror(ret));
180 }
181 }
182
183 /*
184 * Signal-safe.
185 */
186 void ust_unlock(void)
187 {
188 sigset_t sig_all_blocked, orig_mask;
189 int ret, oldstate;
190
191 sigfillset(&sig_all_blocked);
192 ret = pthread_sigmask(SIG_SETMASK, &sig_all_blocked, &orig_mask);
193 if (ret) {
194 ERR("pthread_sigmask: %s", strerror(ret));
195 }
196 if (!--URCU_TLS(ust_mutex_nest))
197 pthread_mutex_unlock(&ust_mutex);
198 ret = pthread_sigmask(SIG_SETMASK, &orig_mask, NULL);
199 if (ret) {
200 ERR("pthread_sigmask: %s", strerror(ret));
201 }
202 ret = pthread_setcancelstate(PTHREAD_CANCEL_ENABLE, &oldstate);
203 if (ret) {
204 ERR("pthread_setcancelstate: %s", strerror(ret));
205 }
206 if (oldstate != PTHREAD_CANCEL_DISABLE) {
207 ERR("pthread_setcancelstate: unexpected oldstate");
208 }
209 }
210
211 /*
212 * Wait for either of these before continuing to the main
213 * program:
214 * - the register_done message from sessiond daemon
215 * (will let the sessiond daemon enable sessions before main
216 * starts.)
217 * - sessiond daemon is not reachable.
218 * - timeout (ensuring applications are resilient to session
219 * daemon problems).
220 */
221 static sem_t constructor_wait;
222 /*
223 * Doing this for both the global and local sessiond.
224 */
225 static int sem_count = { 2 };
226
227 /*
228 * Counting nesting within lttng-ust. Used to ensure that calling fork()
229 * from liblttng-ust does not execute the pre/post fork handlers.
230 */
231 static DEFINE_URCU_TLS(int, lttng_ust_nest_count);
232
233 /*
234 * Info about socket and associated listener thread.
235 */
236 struct sock_info {
237 const char *name;
238 pthread_t ust_listener; /* listener thread */
239 int root_handle;
240 int constructor_sem_posted;
241 int allowed;
242 int global;
243 int thread_active;
244
245 char sock_path[PATH_MAX];
246 int socket;
247 int notify_socket;
248
249 char wait_shm_path[PATH_MAX];
250 char *wait_shm_mmap;
251 /* Keep track of lazy state dump not performed yet. */
252 int statedump_pending;
253 };
254
255 /* Socket from app (connect) to session daemon (listen) for communication */
256 struct sock_info global_apps = {
257 .name = "global",
258 .global = 1,
259
260 .root_handle = -1,
261 .allowed = 1,
262 .thread_active = 0,
263
264 .sock_path = LTTNG_DEFAULT_RUNDIR "/" LTTNG_UST_SOCK_FILENAME,
265 .socket = -1,
266 .notify_socket = -1,
267
268 .wait_shm_path = "/" LTTNG_UST_WAIT_FILENAME,
269
270 .statedump_pending = 0,
271 };
272
273 /* TODO: allow global_apps_sock_path override */
274
275 struct sock_info local_apps = {
276 .name = "local",
277 .global = 0,
278 .root_handle = -1,
279 .allowed = 0, /* Check setuid bit first */
280 .thread_active = 0,
281
282 .socket = -1,
283 .notify_socket = -1,
284
285 .statedump_pending = 0,
286 };
287
288 static int wait_poll_fallback;
289
290 static const char *cmd_name_mapping[] = {
291 [ LTTNG_UST_RELEASE ] = "Release",
292 [ LTTNG_UST_SESSION ] = "Create Session",
293 [ LTTNG_UST_TRACER_VERSION ] = "Get Tracer Version",
294
295 [ LTTNG_UST_TRACEPOINT_LIST ] = "Create Tracepoint List",
296 [ LTTNG_UST_WAIT_QUIESCENT ] = "Wait for Quiescent State",
297 [ LTTNG_UST_REGISTER_DONE ] = "Registration Done",
298 [ LTTNG_UST_TRACEPOINT_FIELD_LIST ] = "Create Tracepoint Field List",
299
300 /* Session FD commands */
301 [ LTTNG_UST_CHANNEL ] = "Create Channel",
302 [ LTTNG_UST_SESSION_START ] = "Start Session",
303 [ LTTNG_UST_SESSION_STOP ] = "Stop Session",
304
305 /* Channel FD commands */
306 [ LTTNG_UST_STREAM ] = "Create Stream",
307 [ LTTNG_UST_EVENT ] = "Create Event",
308
309 /* Event and Channel FD commands */
310 [ LTTNG_UST_CONTEXT ] = "Create Context",
311 [ LTTNG_UST_FLUSH_BUFFER ] = "Flush Buffer",
312
313 /* Event, Channel and Session commands */
314 [ LTTNG_UST_ENABLE ] = "Enable",
315 [ LTTNG_UST_DISABLE ] = "Disable",
316
317 /* Tracepoint list commands */
318 [ LTTNG_UST_TRACEPOINT_LIST_GET ] = "List Next Tracepoint",
319 [ LTTNG_UST_TRACEPOINT_FIELD_LIST_GET ] = "List Next Tracepoint Field",
320
321 /* Event FD commands */
322 [ LTTNG_UST_FILTER ] = "Create Filter",
323 [ LTTNG_UST_EXCLUSION ] = "Add exclusions to event",
324 };
325
326 static const char *str_timeout;
327 static int got_timeout_env;
328
329 extern void lttng_ring_buffer_client_overwrite_init(void);
330 extern void lttng_ring_buffer_client_overwrite_rt_init(void);
331 extern void lttng_ring_buffer_client_discard_init(void);
332 extern void lttng_ring_buffer_client_discard_rt_init(void);
333 extern void lttng_ring_buffer_metadata_client_init(void);
334 extern void lttng_ring_buffer_client_overwrite_exit(void);
335 extern void lttng_ring_buffer_client_overwrite_rt_exit(void);
336 extern void lttng_ring_buffer_client_discard_exit(void);
337 extern void lttng_ring_buffer_client_discard_rt_exit(void);
338 extern void lttng_ring_buffer_metadata_client_exit(void);
339
340 ssize_t lttng_ust_read(int fd, void *buf, size_t len)
341 {
342 ssize_t ret;
343 size_t copied = 0, to_copy = len;
344
345 do {
346 ret = read(fd, buf + copied, to_copy);
347 if (ret > 0) {
348 copied += ret;
349 to_copy -= ret;
350 }
351 } while ((ret > 0 && to_copy > 0)
352 || (ret < 0 && errno == EINTR));
353 if (ret > 0) {
354 ret = copied;
355 }
356 return ret;
357 }
358 /*
359 * Returns the HOME directory path. Caller MUST NOT free(3) the returned
360 * pointer.
361 */
362 static
363 const char *get_lttng_home_dir(void)
364 {
365 const char *val;
366
367 val = (const char *) lttng_secure_getenv("LTTNG_HOME");
368 if (val != NULL) {
369 return val;
370 }
371 return (const char *) lttng_secure_getenv("HOME");
372 }
373
374 /*
375 * Force a read (imply TLS fixup for dlopen) of TLS variables.
376 */
377 static
378 void lttng_fixup_nest_count_tls(void)
379 {
380 asm volatile ("" : : "m" (URCU_TLS(lttng_ust_nest_count)));
381 }
382
383 static
384 void lttng_fixup_ust_mutex_nest_tls(void)
385 {
386 asm volatile ("" : : "m" (URCU_TLS(ust_mutex_nest)));
387 }
388
389 /*
390 * Fixup urcu bp TLS.
391 */
392 static
393 void lttng_fixup_urcu_bp_tls(void)
394 {
395 rcu_read_lock();
396 rcu_read_unlock();
397 }
398
399 int lttng_get_notify_socket(void *owner)
400 {
401 struct sock_info *info = owner;
402
403 return info->notify_socket;
404 }
405
406 static
407 void print_cmd(int cmd, int handle)
408 {
409 const char *cmd_name = "Unknown";
410
411 if (cmd >= 0 && cmd < LTTNG_ARRAY_SIZE(cmd_name_mapping)
412 && cmd_name_mapping[cmd]) {
413 cmd_name = cmd_name_mapping[cmd];
414 }
415 DBG("Message Received \"%s\" (%d), Handle \"%s\" (%d)",
416 cmd_name, cmd,
417 lttng_ust_obj_get_name(handle), handle);
418 }
419
420 static
421 int setup_local_apps(void)
422 {
423 const char *home_dir;
424 uid_t uid;
425
426 uid = getuid();
427 /*
428 * Disallow per-user tracing for setuid binaries.
429 */
430 if (uid != geteuid()) {
431 assert(local_apps.allowed == 0);
432 return 0;
433 }
434 home_dir = get_lttng_home_dir();
435 if (!home_dir) {
436 WARN("HOME environment variable not set. Disabling LTTng-UST per-user tracing.");
437 assert(local_apps.allowed == 0);
438 return -ENOENT;
439 }
440 local_apps.allowed = 1;
441 snprintf(local_apps.sock_path, PATH_MAX, "%s/%s/%s",
442 home_dir,
443 LTTNG_DEFAULT_HOME_RUNDIR,
444 LTTNG_UST_SOCK_FILENAME);
445 snprintf(local_apps.wait_shm_path, PATH_MAX, "/%s-%u",
446 LTTNG_UST_WAIT_FILENAME,
447 uid);
448 return 0;
449 }
450
451 /*
452 * Get notify_sock timeout, in ms.
453 * -1: wait forever. 0: don't wait. >0: timeout, in ms.
454 */
455 static
456 long get_timeout(void)
457 {
458 long constructor_delay_ms = LTTNG_UST_DEFAULT_CONSTRUCTOR_TIMEOUT_MS;
459
460 if (!got_timeout_env) {
461 str_timeout = getenv("LTTNG_UST_REGISTER_TIMEOUT");
462 got_timeout_env = 1;
463 }
464 if (str_timeout)
465 constructor_delay_ms = strtol(str_timeout, NULL, 10);
466 /* All negative values are considered as "-1". */
467 if (constructor_delay_ms < -1)
468 constructor_delay_ms = -1;
469 return constructor_delay_ms;
470 }
471
472 static
473 long get_notify_sock_timeout(void)
474 {
475 return get_timeout();
476 }
477
478 /*
479 * Return values: -1: wait forever. 0: don't wait. 1: timeout wait.
480 */
481 static
482 int get_constructor_timeout(struct timespec *constructor_timeout)
483 {
484 long constructor_delay_ms;
485 int ret;
486
487 constructor_delay_ms = get_timeout();
488
489 switch (constructor_delay_ms) {
490 case -1:/* fall-through */
491 case 0:
492 return constructor_delay_ms;
493 default:
494 break;
495 }
496
497 /*
498 * If we are unable to find the current time, don't wait.
499 */
500 ret = clock_gettime(CLOCK_REALTIME, constructor_timeout);
501 if (ret) {
502 /* Don't wait. */
503 return 0;
504 }
505 constructor_timeout->tv_sec += constructor_delay_ms / 1000UL;
506 constructor_timeout->tv_nsec +=
507 (constructor_delay_ms % 1000UL) * 1000000UL;
508 if (constructor_timeout->tv_nsec >= 1000000000UL) {
509 constructor_timeout->tv_sec++;
510 constructor_timeout->tv_nsec -= 1000000000UL;
511 }
512 /* Timeout wait (constructor_delay_ms). */
513 return 1;
514 }
515
516 static
517 int register_to_sessiond(int socket, enum ustctl_socket_type type)
518 {
519 return ustcomm_send_reg_msg(socket,
520 type,
521 CAA_BITS_PER_LONG,
522 lttng_alignof(uint8_t) * CHAR_BIT,
523 lttng_alignof(uint16_t) * CHAR_BIT,
524 lttng_alignof(uint32_t) * CHAR_BIT,
525 lttng_alignof(uint64_t) * CHAR_BIT,
526 lttng_alignof(unsigned long) * CHAR_BIT);
527 }
528
529 static
530 int send_reply(int sock, struct ustcomm_ust_reply *lur)
531 {
532 ssize_t len;
533
534 len = ustcomm_send_unix_sock(sock, lur, sizeof(*lur));
535 switch (len) {
536 case sizeof(*lur):
537 DBG("message successfully sent");
538 return 0;
539 default:
540 if (len == -ECONNRESET) {
541 DBG("remote end closed connection");
542 return 0;
543 }
544 if (len < 0)
545 return len;
546 DBG("incorrect message size: %zd", len);
547 return -EINVAL;
548 }
549 }
550
551 static
552 int handle_register_done(struct sock_info *sock_info)
553 {
554 int ret;
555
556 if (sock_info->constructor_sem_posted)
557 return 0;
558 sock_info->constructor_sem_posted = 1;
559 if (uatomic_read(&sem_count) <= 0) {
560 return 0;
561 }
562 ret = uatomic_add_return(&sem_count, -1);
563 if (ret == 0) {
564 ret = sem_post(&constructor_wait);
565 assert(!ret);
566 }
567 return 0;
568 }
569
570 /*
571 * Only execute pending statedump after the constructor semaphore has
572 * been posted by each listener thread. This means statedump will only
573 * be performed after the "registration done" command is received from
574 * each session daemon the application is connected to.
575 *
576 * This ensures we don't run into deadlock issues with the dynamic
577 * loader mutex, which is held while the constructor is called and
578 * waiting on the constructor semaphore. All operations requiring this
579 * dynamic loader lock need to be postponed using this mechanism.
580 */
581 static
582 void handle_pending_statedump(struct sock_info *sock_info)
583 {
584 int ctor_passed = sock_info->constructor_sem_posted;
585
586 if (ctor_passed && sock_info->statedump_pending) {
587 sock_info->statedump_pending = 0;
588 pthread_mutex_lock(&ust_fork_mutex);
589 lttng_handle_pending_statedump(sock_info);
590 pthread_mutex_unlock(&ust_fork_mutex);
591 }
592 }
593
594 static
595 int handle_message(struct sock_info *sock_info,
596 int sock, struct ustcomm_ust_msg *lum)
597 {
598 int ret = 0;
599 const struct lttng_ust_objd_ops *ops;
600 struct ustcomm_ust_reply lur;
601 union ust_args args;
602 char ctxstr[LTTNG_UST_SYM_NAME_LEN]; /* App context string. */
603 ssize_t len;
604
605 memset(&lur, 0, sizeof(lur));
606
607 if (ust_lock()) {
608 ret = -LTTNG_UST_ERR_EXITING;
609 goto error;
610 }
611
612 ops = objd_ops(lum->handle);
613 if (!ops) {
614 ret = -ENOENT;
615 goto error;
616 }
617
618 switch (lum->cmd) {
619 case LTTNG_UST_REGISTER_DONE:
620 if (lum->handle == LTTNG_UST_ROOT_HANDLE)
621 ret = handle_register_done(sock_info);
622 else
623 ret = -EINVAL;
624 break;
625 case LTTNG_UST_RELEASE:
626 if (lum->handle == LTTNG_UST_ROOT_HANDLE)
627 ret = -EPERM;
628 else
629 ret = lttng_ust_objd_unref(lum->handle, 1);
630 break;
631 case LTTNG_UST_FILTER:
632 {
633 /* Receive filter data */
634 struct lttng_ust_filter_bytecode_node *bytecode;
635
636 if (lum->u.filter.data_size > FILTER_BYTECODE_MAX_LEN) {
637 ERR("Filter data size is too large: %u bytes",
638 lum->u.filter.data_size);
639 ret = -EINVAL;
640 goto error;
641 }
642
643 if (lum->u.filter.reloc_offset > lum->u.filter.data_size) {
644 ERR("Filter reloc offset %u is not within data",
645 lum->u.filter.reloc_offset);
646 ret = -EINVAL;
647 goto error;
648 }
649
650 bytecode = zmalloc(sizeof(*bytecode) + lum->u.filter.data_size);
651 if (!bytecode) {
652 ret = -ENOMEM;
653 goto error;
654 }
655 len = ustcomm_recv_unix_sock(sock, bytecode->bc.data,
656 lum->u.filter.data_size);
657 switch (len) {
658 case 0: /* orderly shutdown */
659 ret = 0;
660 free(bytecode);
661 goto error;
662 default:
663 if (len == lum->u.filter.data_size) {
664 DBG("filter data received");
665 break;
666 } else if (len < 0) {
667 DBG("Receive failed from lttng-sessiond with errno %d", (int) -len);
668 if (len == -ECONNRESET) {
669 ERR("%s remote end closed connection", sock_info->name);
670 ret = len;
671 free(bytecode);
672 goto error;
673 }
674 ret = len;
675 free(bytecode);
676 goto error;
677 } else {
678 DBG("incorrect filter data message size: %zd", len);
679 ret = -EINVAL;
680 free(bytecode);
681 goto error;
682 }
683 }
684 bytecode->bc.len = lum->u.filter.data_size;
685 bytecode->bc.reloc_offset = lum->u.filter.reloc_offset;
686 bytecode->bc.seqnum = lum->u.filter.seqnum;
687 if (ops->cmd) {
688 ret = ops->cmd(lum->handle, lum->cmd,
689 (unsigned long) bytecode,
690 &args, sock_info);
691 if (ret) {
692 free(bytecode);
693 }
694 /* don't free bytecode if everything went fine. */
695 } else {
696 ret = -ENOSYS;
697 free(bytecode);
698 }
699 break;
700 }
701 case LTTNG_UST_EXCLUSION:
702 {
703 /* Receive exclusion names */
704 struct lttng_ust_excluder_node *node;
705 unsigned int count;
706
707 count = lum->u.exclusion.count;
708 if (count == 0) {
709 /* There are no names to read */
710 ret = 0;
711 goto error;
712 }
713 node = zmalloc(sizeof(*node) +
714 count * LTTNG_UST_SYM_NAME_LEN);
715 if (!node) {
716 ret = -ENOMEM;
717 goto error;
718 }
719 node->excluder.count = count;
720 len = ustcomm_recv_unix_sock(sock, node->excluder.names,
721 count * LTTNG_UST_SYM_NAME_LEN);
722 switch (len) {
723 case 0: /* orderly shutdown */
724 ret = 0;
725 free(node);
726 goto error;
727 default:
728 if (len == count * LTTNG_UST_SYM_NAME_LEN) {
729 DBG("Exclusion data received");
730 break;
731 } else if (len < 0) {
732 DBG("Receive failed from lttng-sessiond with errno %d", (int) -len);
733 if (len == -ECONNRESET) {
734 ERR("%s remote end closed connection", sock_info->name);
735 ret = len;
736 free(node);
737 goto error;
738 }
739 ret = len;
740 free(node);
741 goto error;
742 } else {
743 DBG("Incorrect exclusion data message size: %zd", len);
744 ret = -EINVAL;
745 free(node);
746 goto error;
747 }
748 }
749 if (ops->cmd) {
750 ret = ops->cmd(lum->handle, lum->cmd,
751 (unsigned long) node,
752 &args, sock_info);
753 if (ret) {
754 free(node);
755 }
756 /* Don't free exclusion data if everything went fine. */
757 } else {
758 ret = -ENOSYS;
759 free(node);
760 }
761 break;
762 }
763 case LTTNG_UST_CHANNEL:
764 {
765 void *chan_data;
766 int wakeup_fd;
767
768 len = ustcomm_recv_channel_from_sessiond(sock,
769 &chan_data, lum->u.channel.len,
770 &wakeup_fd);
771 switch (len) {
772 case 0: /* orderly shutdown */
773 ret = 0;
774 goto error;
775 default:
776 if (len == lum->u.channel.len) {
777 DBG("channel data received");
778 break;
779 } else if (len < 0) {
780 DBG("Receive failed from lttng-sessiond with errno %d", (int) -len);
781 if (len == -ECONNRESET) {
782 ERR("%s remote end closed connection", sock_info->name);
783 ret = len;
784 goto error;
785 }
786 ret = len;
787 goto error;
788 } else {
789 DBG("incorrect channel data message size: %zd", len);
790 ret = -EINVAL;
791 goto error;
792 }
793 }
794 args.channel.chan_data = chan_data;
795 args.channel.wakeup_fd = wakeup_fd;
796 if (ops->cmd)
797 ret = ops->cmd(lum->handle, lum->cmd,
798 (unsigned long) &lum->u,
799 &args, sock_info);
800 else
801 ret = -ENOSYS;
802 break;
803 }
804 case LTTNG_UST_STREAM:
805 {
806 /* Receive shm_fd, wakeup_fd */
807 ret = ustcomm_recv_stream_from_sessiond(sock,
808 &lum->u.stream.len,
809 &args.stream.shm_fd,
810 &args.stream.wakeup_fd);
811 if (ret) {
812 goto error;
813 }
814 if (ops->cmd)
815 ret = ops->cmd(lum->handle, lum->cmd,
816 (unsigned long) &lum->u,
817 &args, sock_info);
818 else
819 ret = -ENOSYS;
820 break;
821 }
822 case LTTNG_UST_CONTEXT:
823 switch (lum->u.context.ctx) {
824 case LTTNG_UST_CONTEXT_APP_CONTEXT:
825 {
826 char *p;
827 size_t ctxlen, recvlen;
828
829 ctxlen = strlen("$app.") + lum->u.context.u.app_ctx.provider_name_len - 1
830 + strlen(":") + lum->u.context.u.app_ctx.ctx_name_len;
831 if (ctxlen >= LTTNG_UST_SYM_NAME_LEN) {
832 ERR("Application context string length size is too large: %zu bytes",
833 ctxlen);
834 ret = -EINVAL;
835 goto error;
836 }
837 strcpy(ctxstr, "$app.");
838 p = &ctxstr[strlen("$app.")];
839 recvlen = ctxlen - strlen("$app.");
840 len = ustcomm_recv_unix_sock(sock, p, recvlen);
841 switch (len) {
842 case 0: /* orderly shutdown */
843 ret = 0;
844 goto error;
845 default:
846 if (len == recvlen) {
847 DBG("app context data received");
848 break;
849 } else if (len < 0) {
850 DBG("Receive failed from lttng-sessiond with errno %d", (int) -len);
851 if (len == -ECONNRESET) {
852 ERR("%s remote end closed connection", sock_info->name);
853 ret = len;
854 goto error;
855 }
856 ret = len;
857 goto error;
858 } else {
859 DBG("incorrect app context data message size: %zd", len);
860 ret = -EINVAL;
861 goto error;
862 }
863 }
864 /* Put : between provider and ctxname. */
865 p[lum->u.context.u.app_ctx.provider_name_len - 1] = ':';
866 args.app_context.ctxname = ctxstr;
867 break;
868 }
869 default:
870 break;
871 }
872 if (ops->cmd) {
873 ret = ops->cmd(lum->handle, lum->cmd,
874 (unsigned long) &lum->u,
875 &args, sock_info);
876 } else {
877 ret = -ENOSYS;
878 }
879 break;
880 default:
881 if (ops->cmd)
882 ret = ops->cmd(lum->handle, lum->cmd,
883 (unsigned long) &lum->u,
884 &args, sock_info);
885 else
886 ret = -ENOSYS;
887 break;
888 }
889
890 lur.handle = lum->handle;
891 lur.cmd = lum->cmd;
892 lur.ret_val = ret;
893 if (ret >= 0) {
894 lur.ret_code = LTTNG_UST_OK;
895 } else {
896 /*
897 * Use -LTTNG_UST_ERR as wildcard for UST internal
898 * error that are not caused by the transport, except if
899 * we already have a more precise error message to
900 * report.
901 */
902 if (ret > -LTTNG_UST_ERR) {
903 /* Translate code to UST error. */
904 switch (ret) {
905 case -EEXIST:
906 lur.ret_code = -LTTNG_UST_ERR_EXIST;
907 break;
908 case -EINVAL:
909 lur.ret_code = -LTTNG_UST_ERR_INVAL;
910 break;
911 case -ENOENT:
912 lur.ret_code = -LTTNG_UST_ERR_NOENT;
913 break;
914 case -EPERM:
915 lur.ret_code = -LTTNG_UST_ERR_PERM;
916 break;
917 case -ENOSYS:
918 lur.ret_code = -LTTNG_UST_ERR_NOSYS;
919 break;
920 default:
921 lur.ret_code = -LTTNG_UST_ERR;
922 break;
923 }
924 } else {
925 lur.ret_code = ret;
926 }
927 }
928 if (ret >= 0) {
929 switch (lum->cmd) {
930 case LTTNG_UST_TRACER_VERSION:
931 lur.u.version = lum->u.version;
932 break;
933 case LTTNG_UST_TRACEPOINT_LIST_GET:
934 memcpy(&lur.u.tracepoint, &lum->u.tracepoint, sizeof(lur.u.tracepoint));
935 break;
936 }
937 }
938 DBG("Return value: %d", lur.ret_val);
939 ret = send_reply(sock, &lur);
940 if (ret < 0) {
941 DBG("error sending reply");
942 goto error;
943 }
944
945 /*
946 * LTTNG_UST_TRACEPOINT_FIELD_LIST_GET needs to send the field
947 * after the reply.
948 */
949 if (lur.ret_code == LTTNG_UST_OK) {
950 switch (lum->cmd) {
951 case LTTNG_UST_TRACEPOINT_FIELD_LIST_GET:
952 len = ustcomm_send_unix_sock(sock,
953 &args.field_list.entry,
954 sizeof(args.field_list.entry));
955 if (len < 0) {
956 ret = len;
957 goto error;
958 }
959 if (len != sizeof(args.field_list.entry)) {
960 ret = -EINVAL;
961 goto error;
962 }
963 }
964 }
965
966 error:
967 ust_unlock();
968
969 /*
970 * Performed delayed statedump operations outside of the UST
971 * lock. We need to take the dynamic loader lock before we take
972 * the UST lock internally within handle_pending_statedump().
973 */
974 handle_pending_statedump(sock_info);
975
976 return ret;
977 }
978
979 static
980 void cleanup_sock_info(struct sock_info *sock_info, int exiting)
981 {
982 int ret;
983
984 if (sock_info->root_handle != -1) {
985 ret = lttng_ust_objd_unref(sock_info->root_handle, 1);
986 if (ret) {
987 ERR("Error unref root handle");
988 }
989 sock_info->root_handle = -1;
990 }
991 sock_info->constructor_sem_posted = 0;
992
993 /*
994 * wait_shm_mmap, socket and notify socket are used by listener
995 * threads outside of the ust lock, so we cannot tear them down
996 * ourselves, because we cannot join on these threads. Leave
997 * responsibility of cleaning up these resources to the OS
998 * process exit.
999 */
1000 if (exiting)
1001 return;
1002
1003 if (sock_info->socket != -1) {
1004 ret = ustcomm_close_unix_sock(sock_info->socket);
1005 if (ret) {
1006 ERR("Error closing ust cmd socket");
1007 }
1008 sock_info->socket = -1;
1009 }
1010 if (sock_info->notify_socket != -1) {
1011 ret = ustcomm_close_unix_sock(sock_info->notify_socket);
1012 if (ret) {
1013 ERR("Error closing ust notify socket");
1014 }
1015 sock_info->notify_socket = -1;
1016 }
1017 if (sock_info->wait_shm_mmap) {
1018 long page_size;
1019
1020 page_size = sysconf(_SC_PAGE_SIZE);
1021 if (page_size <= 0) {
1022 if (!page_size) {
1023 errno = EINVAL;
1024 }
1025 PERROR("Error in sysconf(_SC_PAGE_SIZE)");
1026 } else {
1027 ret = munmap(sock_info->wait_shm_mmap, page_size);
1028 if (ret) {
1029 ERR("Error unmapping wait shm");
1030 }
1031 }
1032 sock_info->wait_shm_mmap = NULL;
1033 }
1034 }
1035
1036 /*
1037 * Using fork to set umask in the child process (not multi-thread safe).
1038 * We deal with the shm_open vs ftruncate race (happening when the
1039 * sessiond owns the shm and does not let everybody modify it, to ensure
1040 * safety against shm_unlink) by simply letting the mmap fail and
1041 * retrying after a few seconds.
1042 * For global shm, everybody has rw access to it until the sessiond
1043 * starts.
1044 */
1045 static
1046 int get_wait_shm(struct sock_info *sock_info, size_t mmap_size)
1047 {
1048 int wait_shm_fd, ret;
1049 pid_t pid;
1050
1051 /*
1052 * Try to open read-only.
1053 */
1054 wait_shm_fd = shm_open(sock_info->wait_shm_path, O_RDONLY, 0);
1055 if (wait_shm_fd >= 0) {
1056 int32_t tmp_read;
1057 ssize_t len;
1058 size_t bytes_read = 0;
1059
1060 /*
1061 * Try to read the fd. If unable to do so, try opening
1062 * it in write mode.
1063 */
1064 do {
1065 len = read(wait_shm_fd,
1066 &((char *) &tmp_read)[bytes_read],
1067 sizeof(tmp_read) - bytes_read);
1068 if (len > 0) {
1069 bytes_read += len;
1070 }
1071 } while ((len < 0 && errno == EINTR)
1072 || (len > 0 && bytes_read < sizeof(tmp_read)));
1073 if (bytes_read != sizeof(tmp_read)) {
1074 ret = close(wait_shm_fd);
1075 if (ret) {
1076 ERR("close wait_shm_fd");
1077 }
1078 goto open_write;
1079 }
1080 goto end;
1081 } else if (wait_shm_fd < 0 && errno != ENOENT) {
1082 /*
1083 * Real-only open did not work, and it's not because the
1084 * entry was not present. It's a failure that prohibits
1085 * using shm.
1086 */
1087 ERR("Error opening shm %s", sock_info->wait_shm_path);
1088 goto end;
1089 }
1090
1091 open_write:
1092 /*
1093 * If the open failed because the file did not exist, or because
1094 * the file was not truncated yet, try creating it ourself.
1095 */
1096 URCU_TLS(lttng_ust_nest_count)++;
1097 pid = fork();
1098 URCU_TLS(lttng_ust_nest_count)--;
1099 if (pid > 0) {
1100 int status;
1101
1102 /*
1103 * Parent: wait for child to return, in which case the
1104 * shared memory map will have been created.
1105 */
1106 pid = wait(&status);
1107 if (pid < 0 || !WIFEXITED(status) || WEXITSTATUS(status) != 0) {
1108 wait_shm_fd = -1;
1109 goto end;
1110 }
1111 /*
1112 * Try to open read-only again after creation.
1113 */
1114 wait_shm_fd = shm_open(sock_info->wait_shm_path, O_RDONLY, 0);
1115 if (wait_shm_fd < 0) {
1116 /*
1117 * Real-only open did not work. It's a failure
1118 * that prohibits using shm.
1119 */
1120 ERR("Error opening shm %s", sock_info->wait_shm_path);
1121 goto end;
1122 }
1123 goto end;
1124 } else if (pid == 0) {
1125 int create_mode;
1126
1127 /* Child */
1128 create_mode = S_IRUSR | S_IWUSR | S_IRGRP;
1129 if (sock_info->global)
1130 create_mode |= S_IROTH | S_IWGRP | S_IWOTH;
1131 /*
1132 * We're alone in a child process, so we can modify the
1133 * process-wide umask.
1134 */
1135 umask(~create_mode);
1136 /*
1137 * Try creating shm (or get rw access).
1138 * We don't do an exclusive open, because we allow other
1139 * processes to create+ftruncate it concurrently.
1140 */
1141 wait_shm_fd = shm_open(sock_info->wait_shm_path,
1142 O_RDWR | O_CREAT, create_mode);
1143 if (wait_shm_fd >= 0) {
1144 ret = ftruncate(wait_shm_fd, mmap_size);
1145 if (ret) {
1146 PERROR("ftruncate");
1147 _exit(EXIT_FAILURE);
1148 }
1149 _exit(EXIT_SUCCESS);
1150 }
1151 /*
1152 * For local shm, we need to have rw access to accept
1153 * opening it: this means the local sessiond will be
1154 * able to wake us up. For global shm, we open it even
1155 * if rw access is not granted, because the root.root
1156 * sessiond will be able to override all rights and wake
1157 * us up.
1158 */
1159 if (!sock_info->global && errno != EACCES) {
1160 ERR("Error opening shm %s", sock_info->wait_shm_path);
1161 _exit(EXIT_FAILURE);
1162 }
1163 /*
1164 * The shm exists, but we cannot open it RW. Report
1165 * success.
1166 */
1167 _exit(EXIT_SUCCESS);
1168 } else {
1169 return -1;
1170 }
1171 end:
1172 if (wait_shm_fd >= 0 && !sock_info->global) {
1173 struct stat statbuf;
1174
1175 /*
1176 * Ensure that our user is the owner of the shm file for
1177 * local shm. If we do not own the file, it means our
1178 * sessiond will not have access to wake us up (there is
1179 * probably a rogue process trying to fake our
1180 * sessiond). Fallback to polling method in this case.
1181 */
1182 ret = fstat(wait_shm_fd, &statbuf);
1183 if (ret) {
1184 PERROR("fstat");
1185 goto error_close;
1186 }
1187 if (statbuf.st_uid != getuid())
1188 goto error_close;
1189 }
1190 return wait_shm_fd;
1191
1192 error_close:
1193 ret = close(wait_shm_fd);
1194 if (ret) {
1195 PERROR("Error closing fd");
1196 }
1197 return -1;
1198 }
1199
1200 static
1201 char *get_map_shm(struct sock_info *sock_info)
1202 {
1203 long page_size;
1204 int wait_shm_fd, ret;
1205 char *wait_shm_mmap;
1206
1207 page_size = sysconf(_SC_PAGE_SIZE);
1208 if (page_size <= 0) {
1209 if (!page_size) {
1210 errno = EINVAL;
1211 }
1212 PERROR("Error in sysconf(_SC_PAGE_SIZE)");
1213 goto error;
1214 }
1215
1216 wait_shm_fd = get_wait_shm(sock_info, page_size);
1217 if (wait_shm_fd < 0) {
1218 goto error;
1219 }
1220 wait_shm_mmap = mmap(NULL, page_size, PROT_READ,
1221 MAP_SHARED, wait_shm_fd, 0);
1222 /* close shm fd immediately after taking the mmap reference */
1223 ret = close(wait_shm_fd);
1224 if (ret) {
1225 PERROR("Error closing fd");
1226 }
1227 if (wait_shm_mmap == MAP_FAILED) {
1228 DBG("mmap error (can be caused by race with sessiond). Fallback to poll mode.");
1229 goto error;
1230 }
1231 return wait_shm_mmap;
1232
1233 error:
1234 return NULL;
1235 }
1236
1237 static
1238 void wait_for_sessiond(struct sock_info *sock_info)
1239 {
1240 if (ust_lock()) {
1241 goto quit;
1242 }
1243 if (wait_poll_fallback) {
1244 goto error;
1245 }
1246 if (!sock_info->wait_shm_mmap) {
1247 sock_info->wait_shm_mmap = get_map_shm(sock_info);
1248 if (!sock_info->wait_shm_mmap)
1249 goto error;
1250 }
1251 ust_unlock();
1252
1253 DBG("Waiting for %s apps sessiond", sock_info->name);
1254 /* Wait for futex wakeup */
1255 if (uatomic_read((int32_t *) sock_info->wait_shm_mmap))
1256 goto end_wait;
1257
1258 while (futex_async((int32_t *) sock_info->wait_shm_mmap,
1259 FUTEX_WAIT, 0, NULL, NULL, 0)) {
1260 switch (errno) {
1261 case EWOULDBLOCK:
1262 /* Value already changed. */
1263 goto end_wait;
1264 case EINTR:
1265 /* Retry if interrupted by signal. */
1266 break; /* Get out of switch. */
1267 case EFAULT:
1268 wait_poll_fallback = 1;
1269 DBG(
1270 "Linux kernels 2.6.33 to 3.0 (with the exception of stable versions) "
1271 "do not support FUTEX_WAKE on read-only memory mappings correctly. "
1272 "Please upgrade your kernel "
1273 "(fix is commit 9ea71503a8ed9184d2d0b8ccc4d269d05f7940ae in Linux kernel "
1274 "mainline). LTTng-UST will use polling mode fallback.");
1275 if (ust_debug())
1276 PERROR("futex");
1277 goto end_wait;
1278 }
1279 }
1280 end_wait:
1281 return;
1282
1283 quit:
1284 ust_unlock();
1285 return;
1286
1287 error:
1288 ust_unlock();
1289 return;
1290 }
1291
1292 /*
1293 * This thread does not allocate any resource, except within
1294 * handle_message, within mutex protection. This mutex protects against
1295 * fork and exit.
1296 * The other moment it allocates resources is at socket connection, which
1297 * is also protected by the mutex.
1298 */
1299 static
1300 void *ust_listener_thread(void *arg)
1301 {
1302 struct sock_info *sock_info = arg;
1303 int sock, ret, prev_connect_failed = 0, has_waited = 0;
1304 long timeout;
1305
1306 /*
1307 * If available, add '-ust' to the end of this thread's
1308 * process name
1309 */
1310 ret = lttng_ust_setustprocname();
1311 if (ret) {
1312 ERR("Unable to set UST process name");
1313 }
1314
1315 /* Restart trying to connect to the session daemon */
1316 restart:
1317 if (prev_connect_failed) {
1318 /* Wait for sessiond availability with pipe */
1319 wait_for_sessiond(sock_info);
1320 if (has_waited) {
1321 has_waited = 0;
1322 /*
1323 * Sleep for 5 seconds before retrying after a
1324 * sequence of failure / wait / failure. This
1325 * deals with a killed or broken session daemon.
1326 */
1327 sleep(5);
1328 } else {
1329 has_waited = 1;
1330 }
1331 prev_connect_failed = 0;
1332 }
1333
1334 if (sock_info->socket != -1) {
1335 ret = ustcomm_close_unix_sock(sock_info->socket);
1336 if (ret) {
1337 ERR("Error closing %s ust cmd socket",
1338 sock_info->name);
1339 }
1340 sock_info->socket = -1;
1341 }
1342 if (sock_info->notify_socket != -1) {
1343 ret = ustcomm_close_unix_sock(sock_info->notify_socket);
1344 if (ret) {
1345 ERR("Error closing %s ust notify socket",
1346 sock_info->name);
1347 }
1348 sock_info->notify_socket = -1;
1349 }
1350
1351 /*
1352 * Register. We need to perform both connect and sending
1353 * registration message before doing the next connect otherwise
1354 * we may reach unix socket connect queue max limits and block
1355 * on the 2nd connect while the session daemon is awaiting the
1356 * first connect registration message.
1357 */
1358 /* Connect cmd socket */
1359 ret = ustcomm_connect_unix_sock(sock_info->sock_path);
1360 if (ret < 0) {
1361 DBG("Info: sessiond not accepting connections to %s apps socket", sock_info->name);
1362 prev_connect_failed = 1;
1363
1364 if (ust_lock()) {
1365 goto quit;
1366 }
1367
1368 /*
1369 * If we cannot find the sessiond daemon, don't delay
1370 * constructor execution.
1371 */
1372 ret = handle_register_done(sock_info);
1373 assert(!ret);
1374 ust_unlock();
1375 goto restart;
1376 }
1377 sock_info->socket = ret;
1378
1379 if (ust_lock()) {
1380 goto quit;
1381 }
1382
1383 /*
1384 * Create only one root handle per listener thread for the whole
1385 * process lifetime, so we ensure we get ID which is statically
1386 * assigned to the root handle.
1387 */
1388 if (sock_info->root_handle == -1) {
1389 ret = lttng_abi_create_root_handle();
1390 if (ret < 0) {
1391 ERR("Error creating root handle");
1392 goto quit;
1393 }
1394 sock_info->root_handle = ret;
1395 }
1396
1397 ret = register_to_sessiond(sock_info->socket, USTCTL_SOCKET_CMD);
1398 if (ret < 0) {
1399 ERR("Error registering to %s ust cmd socket",
1400 sock_info->name);
1401 prev_connect_failed = 1;
1402 /*
1403 * If we cannot register to the sessiond daemon, don't
1404 * delay constructor execution.
1405 */
1406 ret = handle_register_done(sock_info);
1407 assert(!ret);
1408 ust_unlock();
1409 goto restart;
1410 }
1411
1412 ust_unlock();
1413
1414 /* Connect notify socket */
1415 ret = ustcomm_connect_unix_sock(sock_info->sock_path);
1416 if (ret < 0) {
1417 DBG("Info: sessiond not accepting connections to %s apps socket", sock_info->name);
1418 prev_connect_failed = 1;
1419
1420 if (ust_lock()) {
1421 goto quit;
1422 }
1423
1424 /*
1425 * If we cannot find the sessiond daemon, don't delay
1426 * constructor execution.
1427 */
1428 ret = handle_register_done(sock_info);
1429 assert(!ret);
1430 ust_unlock();
1431 goto restart;
1432 }
1433 sock_info->notify_socket = ret;
1434
1435 timeout = get_notify_sock_timeout();
1436 if (timeout >= 0) {
1437 /*
1438 * Give at least 10ms to sessiond to reply to
1439 * notifications.
1440 */
1441 if (timeout < 10)
1442 timeout = 10;
1443 ret = ustcomm_setsockopt_rcv_timeout(sock_info->notify_socket,
1444 timeout);
1445 if (ret < 0) {
1446 WARN("Error setting socket receive timeout");
1447 }
1448 ret = ustcomm_setsockopt_snd_timeout(sock_info->notify_socket,
1449 timeout);
1450 if (ret < 0) {
1451 WARN("Error setting socket send timeout");
1452 }
1453 } else if (timeout < -1) {
1454 WARN("Unsupported timeout value %ld", timeout);
1455 }
1456
1457 if (ust_lock()) {
1458 goto quit;
1459 }
1460
1461 ret = register_to_sessiond(sock_info->notify_socket,
1462 USTCTL_SOCKET_NOTIFY);
1463 if (ret < 0) {
1464 ERR("Error registering to %s ust notify socket",
1465 sock_info->name);
1466 prev_connect_failed = 1;
1467 /*
1468 * If we cannot register to the sessiond daemon, don't
1469 * delay constructor execution.
1470 */
1471 ret = handle_register_done(sock_info);
1472 assert(!ret);
1473 ust_unlock();
1474 goto restart;
1475 }
1476 sock = sock_info->socket;
1477
1478 ust_unlock();
1479
1480 for (;;) {
1481 ssize_t len;
1482 struct ustcomm_ust_msg lum;
1483
1484 len = ustcomm_recv_unix_sock(sock, &lum, sizeof(lum));
1485 switch (len) {
1486 case 0: /* orderly shutdown */
1487 DBG("%s lttng-sessiond has performed an orderly shutdown", sock_info->name);
1488 if (ust_lock()) {
1489 goto quit;
1490 }
1491 /*
1492 * Either sessiond has shutdown or refused us by closing the socket.
1493 * In either case, we don't want to delay construction execution,
1494 * and we need to wait before retry.
1495 */
1496 prev_connect_failed = 1;
1497 /*
1498 * If we cannot register to the sessiond daemon, don't
1499 * delay constructor execution.
1500 */
1501 ret = handle_register_done(sock_info);
1502 assert(!ret);
1503 ust_unlock();
1504 goto end;
1505 case sizeof(lum):
1506 print_cmd(lum.cmd, lum.handle);
1507 ret = handle_message(sock_info, sock, &lum);
1508 if (ret) {
1509 ERR("Error handling message for %s socket",
1510 sock_info->name);
1511 /*
1512 * Close socket if protocol error is
1513 * detected.
1514 */
1515 goto end;
1516 }
1517 continue;
1518 default:
1519 if (len < 0) {
1520 DBG("Receive failed from lttng-sessiond with errno %d", (int) -len);
1521 } else {
1522 DBG("incorrect message size (%s socket): %zd", sock_info->name, len);
1523 }
1524 if (len == -ECONNRESET) {
1525 DBG("%s remote end closed connection", sock_info->name);
1526 goto end;
1527 }
1528 goto end;
1529 }
1530
1531 }
1532 end:
1533 if (ust_lock()) {
1534 goto quit;
1535 }
1536 /* Cleanup socket handles before trying to reconnect */
1537 lttng_ust_objd_table_owner_cleanup(sock_info);
1538 ust_unlock();
1539 goto restart; /* try to reconnect */
1540
1541 quit:
1542 ust_unlock();
1543
1544 pthread_mutex_lock(&ust_exit_mutex);
1545 sock_info->thread_active = 0;
1546 pthread_mutex_unlock(&ust_exit_mutex);
1547 return NULL;
1548 }
1549
1550 /*
1551 * Weak symbol to call when the ust malloc wrapper is not loaded.
1552 */
1553 __attribute__((weak))
1554 void lttng_ust_malloc_wrapper_init(void)
1555 {
1556 }
1557
1558 /*
1559 * sessiond monitoring thread: monitor presence of global and per-user
1560 * sessiond by polling the application common named pipe.
1561 */
1562 void __attribute__((constructor)) lttng_ust_init(void)
1563 {
1564 struct timespec constructor_timeout;
1565 sigset_t sig_all_blocked, orig_parent_mask;
1566 pthread_attr_t thread_attr;
1567 int timeout_mode;
1568 int ret;
1569
1570 if (uatomic_xchg(&initialized, 1) == 1)
1571 return;
1572
1573 /*
1574 * Fixup interdependency between TLS fixup mutex (which happens
1575 * to be the dynamic linker mutex) and ust_lock, taken within
1576 * the ust lock.
1577 */
1578 lttng_fixup_urcu_bp_tls();
1579 lttng_fixup_ringbuffer_tls();
1580 lttng_fixup_vtid_tls();
1581 lttng_fixup_nest_count_tls();
1582 lttng_fixup_procname_tls();
1583 lttng_fixup_ust_mutex_nest_tls();
1584
1585 lttng_ust_loaded = 1;
1586
1587 /*
1588 * We want precise control over the order in which we construct
1589 * our sub-libraries vs starting to receive commands from
1590 * sessiond (otherwise leading to errors when trying to create
1591 * sessiond before the init functions are completed).
1592 */
1593 init_usterr();
1594 init_tracepoint();
1595 lttng_ust_clock_init();
1596 lttng_ust_getcpu_init();
1597 lttng_ust_statedump_init();
1598 lttng_ring_buffer_metadata_client_init();
1599 lttng_ring_buffer_client_overwrite_init();
1600 lttng_ring_buffer_client_overwrite_rt_init();
1601 lttng_ring_buffer_client_discard_init();
1602 lttng_ring_buffer_client_discard_rt_init();
1603 lttng_perf_counter_init();
1604 /*
1605 * Invoke ust malloc wrapper init before starting other threads.
1606 */
1607 lttng_ust_malloc_wrapper_init();
1608
1609 timeout_mode = get_constructor_timeout(&constructor_timeout);
1610
1611 ret = sem_init(&constructor_wait, 0, 0);
1612 if (ret) {
1613 PERROR("sem_init");
1614 }
1615
1616 ret = setup_local_apps();
1617 if (ret) {
1618 DBG("local apps setup returned %d", ret);
1619 }
1620
1621 /* A new thread created by pthread_create inherits the signal mask
1622 * from the parent. To avoid any signal being received by the
1623 * listener thread, we block all signals temporarily in the parent,
1624 * while we create the listener thread.
1625 */
1626 sigfillset(&sig_all_blocked);
1627 ret = pthread_sigmask(SIG_SETMASK, &sig_all_blocked, &orig_parent_mask);
1628 if (ret) {
1629 ERR("pthread_sigmask: %s", strerror(ret));
1630 }
1631
1632 ret = pthread_attr_init(&thread_attr);
1633 if (ret) {
1634 ERR("pthread_attr_init: %s", strerror(ret));
1635 }
1636 ret = pthread_attr_setdetachstate(&thread_attr, PTHREAD_CREATE_DETACHED);
1637 if (ret) {
1638 ERR("pthread_attr_setdetachstate: %s", strerror(ret));
1639 }
1640
1641 pthread_mutex_lock(&ust_exit_mutex);
1642 ret = pthread_create(&global_apps.ust_listener, &thread_attr,
1643 ust_listener_thread, &global_apps);
1644 if (ret) {
1645 ERR("pthread_create global: %s", strerror(ret));
1646 }
1647 global_apps.thread_active = 1;
1648 pthread_mutex_unlock(&ust_exit_mutex);
1649
1650 if (local_apps.allowed) {
1651 pthread_mutex_lock(&ust_exit_mutex);
1652 ret = pthread_create(&local_apps.ust_listener, &thread_attr,
1653 ust_listener_thread, &local_apps);
1654 if (ret) {
1655 ERR("pthread_create local: %s", strerror(ret));
1656 }
1657 local_apps.thread_active = 1;
1658 pthread_mutex_unlock(&ust_exit_mutex);
1659 } else {
1660 handle_register_done(&local_apps);
1661 }
1662 ret = pthread_attr_destroy(&thread_attr);
1663 if (ret) {
1664 ERR("pthread_attr_destroy: %s", strerror(ret));
1665 }
1666
1667 /* Restore original signal mask in parent */
1668 ret = pthread_sigmask(SIG_SETMASK, &orig_parent_mask, NULL);
1669 if (ret) {
1670 ERR("pthread_sigmask: %s", strerror(ret));
1671 }
1672
1673 switch (timeout_mode) {
1674 case 1: /* timeout wait */
1675 do {
1676 ret = sem_timedwait(&constructor_wait,
1677 &constructor_timeout);
1678 } while (ret < 0 && errno == EINTR);
1679 if (ret < 0) {
1680 switch (errno) {
1681 case ETIMEDOUT:
1682 ERR("Timed out waiting for lttng-sessiond");
1683 break;
1684 case EINVAL:
1685 PERROR("sem_timedwait");
1686 break;
1687 default:
1688 ERR("Unexpected error \"%s\" returned by sem_timedwait",
1689 strerror(errno));
1690 }
1691 }
1692 break;
1693 case -1:/* wait forever */
1694 do {
1695 ret = sem_wait(&constructor_wait);
1696 } while (ret < 0 && errno == EINTR);
1697 if (ret < 0) {
1698 switch (errno) {
1699 case EINVAL:
1700 PERROR("sem_wait");
1701 break;
1702 default:
1703 ERR("Unexpected error \"%s\" returned by sem_wait",
1704 strerror(errno));
1705 }
1706 }
1707 break;
1708 case 0: /* no timeout */
1709 break;
1710 }
1711 }
1712
1713 static
1714 void lttng_ust_cleanup(int exiting)
1715 {
1716 cleanup_sock_info(&global_apps, exiting);
1717 cleanup_sock_info(&local_apps, exiting);
1718 local_apps.allowed = 0;
1719 /*
1720 * The teardown in this function all affect data structures
1721 * accessed under the UST lock by the listener thread. This
1722 * lock, along with the lttng_ust_comm_should_quit flag, ensure
1723 * that none of these threads are accessing this data at this
1724 * point.
1725 */
1726 lttng_ust_abi_exit();
1727 lttng_ust_events_exit();
1728 lttng_perf_counter_exit();
1729 lttng_ring_buffer_client_discard_rt_exit();
1730 lttng_ring_buffer_client_discard_exit();
1731 lttng_ring_buffer_client_overwrite_rt_exit();
1732 lttng_ring_buffer_client_overwrite_exit();
1733 lttng_ring_buffer_metadata_client_exit();
1734 lttng_ust_statedump_destroy();
1735 exit_tracepoint();
1736 if (!exiting) {
1737 /* Reinitialize values for fork */
1738 sem_count = 2;
1739 lttng_ust_comm_should_quit = 0;
1740 initialized = 0;
1741 }
1742 }
1743
1744 void __attribute__((destructor)) lttng_ust_exit(void)
1745 {
1746 int ret;
1747
1748 /*
1749 * Using pthread_cancel here because:
1750 * A) we don't want to hang application teardown.
1751 * B) the thread is not allocating any resource.
1752 */
1753
1754 /*
1755 * Require the communication thread to quit. Synchronize with
1756 * mutexes to ensure it is not in a mutex critical section when
1757 * pthread_cancel is later called.
1758 */
1759 ust_lock_nocheck();
1760 lttng_ust_comm_should_quit = 1;
1761 ust_unlock();
1762
1763 pthread_mutex_lock(&ust_exit_mutex);
1764 /* cancel threads */
1765 if (global_apps.thread_active) {
1766 ret = pthread_cancel(global_apps.ust_listener);
1767 if (ret) {
1768 ERR("Error cancelling global ust listener thread: %s",
1769 strerror(ret));
1770 } else {
1771 global_apps.thread_active = 0;
1772 }
1773 }
1774 if (local_apps.thread_active) {
1775 ret = pthread_cancel(local_apps.ust_listener);
1776 if (ret) {
1777 ERR("Error cancelling local ust listener thread: %s",
1778 strerror(ret));
1779 } else {
1780 local_apps.thread_active = 0;
1781 }
1782 }
1783 pthread_mutex_unlock(&ust_exit_mutex);
1784
1785 /*
1786 * Do NOT join threads: use of sys_futex makes it impossible to
1787 * join the threads without using async-cancel, but async-cancel
1788 * is delivered by a signal, which could hit the target thread
1789 * anywhere in its code path, including while the ust_lock() is
1790 * held, causing a deadlock for the other thread. Let the OS
1791 * cleanup the threads if there are stalled in a syscall.
1792 */
1793 lttng_ust_cleanup(1);
1794 }
1795
1796 /*
1797 * We exclude the worker threads across fork and clone (except
1798 * CLONE_VM), because these system calls only keep the forking thread
1799 * running in the child. Therefore, we don't want to call fork or clone
1800 * in the middle of an tracepoint or ust tracing state modification.
1801 * Holding this mutex protects these structures across fork and clone.
1802 */
1803 void ust_before_fork(sigset_t *save_sigset)
1804 {
1805 /*
1806 * Disable signals. This is to avoid that the child intervenes
1807 * before it is properly setup for tracing. It is safer to
1808 * disable all signals, because then we know we are not breaking
1809 * anything by restoring the original mask.
1810 */
1811 sigset_t all_sigs;
1812 int ret;
1813
1814 if (URCU_TLS(lttng_ust_nest_count))
1815 return;
1816 /* Disable signals */
1817 sigfillset(&all_sigs);
1818 ret = sigprocmask(SIG_BLOCK, &all_sigs, save_sigset);
1819 if (ret == -1) {
1820 PERROR("sigprocmask");
1821 }
1822
1823 pthread_mutex_lock(&ust_fork_mutex);
1824
1825 ust_lock_nocheck();
1826 rcu_bp_before_fork();
1827 }
1828
1829 static void ust_after_fork_common(sigset_t *restore_sigset)
1830 {
1831 int ret;
1832
1833 DBG("process %d", getpid());
1834 ust_unlock();
1835
1836 pthread_mutex_unlock(&ust_fork_mutex);
1837
1838 /* Restore signals */
1839 ret = sigprocmask(SIG_SETMASK, restore_sigset, NULL);
1840 if (ret == -1) {
1841 PERROR("sigprocmask");
1842 }
1843 }
1844
1845 void ust_after_fork_parent(sigset_t *restore_sigset)
1846 {
1847 if (URCU_TLS(lttng_ust_nest_count))
1848 return;
1849 DBG("process %d", getpid());
1850 rcu_bp_after_fork_parent();
1851 /* Release mutexes and reenable signals */
1852 ust_after_fork_common(restore_sigset);
1853 }
1854
1855 /*
1856 * After fork, in the child, we need to cleanup all the leftover state,
1857 * except the worker thread which already magically disappeared thanks
1858 * to the weird Linux fork semantics. After tyding up, we call
1859 * lttng_ust_init() again to start over as a new PID.
1860 *
1861 * This is meant for forks() that have tracing in the child between the
1862 * fork and following exec call (if there is any).
1863 */
1864 void ust_after_fork_child(sigset_t *restore_sigset)
1865 {
1866 if (URCU_TLS(lttng_ust_nest_count))
1867 return;
1868 lttng_context_vtid_reset();
1869 DBG("process %d", getpid());
1870 /* Release urcu mutexes */
1871 rcu_bp_after_fork_child();
1872 lttng_ust_cleanup(0);
1873 /* Release mutexes and reenable signals */
1874 ust_after_fork_common(restore_sigset);
1875 lttng_ust_init();
1876 }
1877
1878 void lttng_ust_sockinfo_session_enabled(void *owner)
1879 {
1880 struct sock_info *sock_info = owner;
1881 sock_info->statedump_pending = 1;
1882 }
This page took 0.087362 seconds and 3 git commands to generate.