1 /* SPDX-License-Identifier: MIT
3 * lttng-filter-interpreter.c
5 * LTTng modules filter interpreter.
7 * Copyright (C) 2010-2016 Mathieu Desnoyers <mathieu.desnoyers@efficios.com>
10 #include <wrapper/compiler_attributes.h>
11 #include <wrapper/uaccess.h>
12 #include <wrapper/objtool.h>
13 #include <wrapper/types.h>
14 #include <linux/swab.h>
16 #include <lttng-filter.h>
17 #include <lttng-string-utils.h>
19 LTTNG_STACK_FRAME_NON_STANDARD(lttng_filter_interpret_bytecode
);
22 * get_char should be called with page fault handler disabled if it is expected
23 * to handle user-space read.
26 char get_char(const struct estack_entry
*reg
, size_t offset
)
28 if (unlikely(offset
>= reg
->u
.s
.seq_len
))
33 /* Handle invalid access as end of string. */
34 if (unlikely(!lttng_access_ok(VERIFY_READ
,
35 reg
->u
.s
.user_str
+ offset
,
38 /* Handle fault (nonzero return value) as end of string. */
39 if (unlikely(__copy_from_user_inatomic(&c
,
40 reg
->u
.s
.user_str
+ offset
,
45 return reg
->u
.s
.str
[offset
];
51 * -2: unknown escape char.
55 int parse_char(struct estack_entry
*reg
, char *c
, size_t *offset
)
60 *c
= get_char(reg
, *offset
);
76 char get_char_at_cb(size_t at
, void *data
)
78 return get_char(data
, at
);
82 int stack_star_glob_match(struct estack
*stack
, int top
, const char *cmp_type
)
84 bool has_user
= false;
86 struct estack_entry
*pattern_reg
;
87 struct estack_entry
*candidate_reg
;
89 /* Disable the page fault handler when reading from userspace. */
90 if (estack_bx(stack
, top
)->u
.s
.user
91 || estack_ax(stack
, top
)->u
.s
.user
) {
96 /* Find out which side is the pattern vs. the candidate. */
97 if (estack_ax(stack
, top
)->u
.s
.literal_type
== ESTACK_STRING_LITERAL_TYPE_STAR_GLOB
) {
98 pattern_reg
= estack_ax(stack
, top
);
99 candidate_reg
= estack_bx(stack
, top
);
101 pattern_reg
= estack_bx(stack
, top
);
102 candidate_reg
= estack_ax(stack
, top
);
105 /* Perform the match operation. */
106 result
= !strutils_star_glob_match_char_cb(get_char_at_cb
,
107 pattern_reg
, get_char_at_cb
, candidate_reg
);
115 int stack_strcmp(struct estack
*stack
, int top
, const char *cmp_type
)
117 size_t offset_bx
= 0, offset_ax
= 0;
118 int diff
, has_user
= 0;
120 if (estack_bx(stack
, top
)->u
.s
.user
121 || estack_ax(stack
, top
)->u
.s
.user
) {
129 char char_bx
, char_ax
;
131 char_bx
= get_char(estack_bx(stack
, top
), offset_bx
);
132 char_ax
= get_char(estack_ax(stack
, top
), offset_ax
);
134 if (unlikely(char_bx
== '\0')) {
135 if (char_ax
== '\0') {
139 if (estack_ax(stack
, top
)->u
.s
.literal_type
==
140 ESTACK_STRING_LITERAL_TYPE_PLAIN
) {
141 ret
= parse_char(estack_ax(stack
, top
),
142 &char_ax
, &offset_ax
);
152 if (unlikely(char_ax
== '\0')) {
153 if (estack_bx(stack
, top
)->u
.s
.literal_type
==
154 ESTACK_STRING_LITERAL_TYPE_PLAIN
) {
155 ret
= parse_char(estack_bx(stack
, top
),
156 &char_bx
, &offset_bx
);
165 if (estack_bx(stack
, top
)->u
.s
.literal_type
==
166 ESTACK_STRING_LITERAL_TYPE_PLAIN
) {
167 ret
= parse_char(estack_bx(stack
, top
),
168 &char_bx
, &offset_bx
);
172 } else if (ret
== -2) {
175 /* else compare both char */
177 if (estack_ax(stack
, top
)->u
.s
.literal_type
==
178 ESTACK_STRING_LITERAL_TYPE_PLAIN
) {
179 ret
= parse_char(estack_ax(stack
, top
),
180 &char_ax
, &offset_ax
);
184 } else if (ret
== -2) {
201 diff
= char_bx
- char_ax
;
213 uint64_t lttng_filter_false(void *filter_data
,
214 struct lttng_probe_ctx
*lttng_probe_ctx
,
215 const char *filter_stack_data
)
220 #ifdef INTERPRETER_USE_SWITCH
223 * Fallback for compilers that do not support taking address of labels.
227 start_pc = &bytecode->data[0]; \
228 for (pc = next_pc = start_pc; pc - start_pc < bytecode->len; \
230 dbg_printk("Executing op %s (%u)\n", \
231 lttng_filter_print_op((unsigned int) *(filter_opcode_t *) pc), \
232 (unsigned int) *(filter_opcode_t *) pc); \
233 switch (*(filter_opcode_t *) pc) {
235 #define OP(name) case name
245 * Dispatch-table based interpreter.
249 start_pc = &bytecode->code[0]; \
250 pc = next_pc = start_pc; \
251 if (unlikely(pc - start_pc >= bytecode->len)) \
253 goto *dispatch[*(filter_opcode_t *) pc];
260 goto *dispatch[*(filter_opcode_t *) pc];
266 static int context_get_index(struct lttng_probe_ctx
*lttng_probe_ctx
,
267 struct load_ptr
*ptr
,
271 struct lttng_ctx_field
*ctx_field
;
272 struct lttng_event_field
*field
;
273 union lttng_ctx_value v
;
275 ctx_field
= <tng_static_ctx
->fields
[idx
];
276 field
= &ctx_field
->event_field
;
277 ptr
->type
= LOAD_OBJECT
;
278 /* field is only used for types nested within variants. */
281 switch (field
->type
.atype
) {
283 ctx_field
->get_value(ctx_field
, lttng_probe_ctx
, &v
);
284 if (field
->type
.u
.basic
.integer
.signedness
) {
285 ptr
->object_type
= OBJECT_TYPE_S64
;
287 ptr
->ptr
= &ptr
->u
.s64
;
289 ptr
->object_type
= OBJECT_TYPE_U64
;
290 ptr
->u
.u64
= v
.s64
; /* Cast. */
291 ptr
->ptr
= &ptr
->u
.u64
;
296 const struct lttng_integer_type
*itype
=
297 &field
->type
.u
.basic
.enumeration
.container_type
;
299 ctx_field
->get_value(ctx_field
, lttng_probe_ctx
, &v
);
300 if (itype
->signedness
) {
301 ptr
->object_type
= OBJECT_TYPE_S64
;
303 ptr
->ptr
= &ptr
->u
.s64
;
305 ptr
->object_type
= OBJECT_TYPE_U64
;
306 ptr
->u
.u64
= v
.s64
; /* Cast. */
307 ptr
->ptr
= &ptr
->u
.u64
;
312 if (field
->type
.u
.array
.elem_type
.atype
!= atype_integer
) {
313 printk(KERN_WARNING
"Array nesting only supports integer types.\n");
316 if (field
->type
.u
.array
.elem_type
.u
.basic
.integer
.encoding
== lttng_encode_none
) {
317 printk(KERN_WARNING
"Only string arrays are supported for contexts.\n");
320 ptr
->object_type
= OBJECT_TYPE_STRING
;
321 ctx_field
->get_value(ctx_field
, lttng_probe_ctx
, &v
);
325 if (field
->type
.u
.sequence
.elem_type
.atype
!= atype_integer
) {
326 printk(KERN_WARNING
"Sequence nesting only supports integer types.\n");
329 if (field
->type
.u
.sequence
.elem_type
.u
.basic
.integer
.encoding
== lttng_encode_none
) {
330 printk(KERN_WARNING
"Only string sequences are supported for contexts.\n");
333 ptr
->object_type
= OBJECT_TYPE_STRING
;
334 ctx_field
->get_value(ctx_field
, lttng_probe_ctx
, &v
);
337 case atype_array_bitfield
:
338 printk(KERN_WARNING
"Bitfield array type is not supported.\n");
340 case atype_sequence_bitfield
:
341 printk(KERN_WARNING
"Bitfield sequence type is not supported.\n");
344 ptr
->object_type
= OBJECT_TYPE_STRING
;
345 ctx_field
->get_value(ctx_field
, lttng_probe_ctx
, &v
);
349 printk(KERN_WARNING
"Structure type cannot be loaded.\n");
352 printk(KERN_WARNING
"Unknown type: %d", (int) field
->type
.atype
);
358 static int dynamic_get_index(struct lttng_probe_ctx
*lttng_probe_ctx
,
359 struct bytecode_runtime
*runtime
,
360 uint64_t index
, struct estack_entry
*stack_top
)
363 const struct filter_get_index_data
*gid
;
366 * Types nested within variants need to perform dynamic lookup
367 * based on the field descriptions. LTTng-UST does not implement
370 if (stack_top
->u
.ptr
.field
)
372 gid
= (const struct filter_get_index_data
*) &runtime
->data
[index
];
373 switch (stack_top
->u
.ptr
.type
) {
375 switch (stack_top
->u
.ptr
.object_type
) {
376 case OBJECT_TYPE_ARRAY
:
380 WARN_ON_ONCE(gid
->offset
>= gid
->array_len
);
381 /* Skip count (unsigned long) */
382 ptr
= *(const char **) (stack_top
->u
.ptr
.ptr
+ sizeof(unsigned long));
383 ptr
= ptr
+ gid
->offset
;
384 stack_top
->u
.ptr
.ptr
= ptr
;
385 stack_top
->u
.ptr
.object_type
= gid
->elem
.type
;
386 stack_top
->u
.ptr
.rev_bo
= gid
->elem
.rev_bo
;
387 /* field is only used for types nested within variants. */
388 stack_top
->u
.ptr
.field
= NULL
;
391 case OBJECT_TYPE_SEQUENCE
:
396 ptr
= *(const char **) (stack_top
->u
.ptr
.ptr
+ sizeof(unsigned long));
397 ptr_seq_len
= *(unsigned long *) stack_top
->u
.ptr
.ptr
;
398 if (gid
->offset
>= gid
->elem
.len
* ptr_seq_len
) {
402 ptr
= ptr
+ gid
->offset
;
403 stack_top
->u
.ptr
.ptr
= ptr
;
404 stack_top
->u
.ptr
.object_type
= gid
->elem
.type
;
405 stack_top
->u
.ptr
.rev_bo
= gid
->elem
.rev_bo
;
406 /* field is only used for types nested within variants. */
407 stack_top
->u
.ptr
.field
= NULL
;
410 case OBJECT_TYPE_STRUCT
:
411 printk(KERN_WARNING
"Nested structures are not supported yet.\n");
414 case OBJECT_TYPE_VARIANT
:
416 printk(KERN_WARNING
"Unexpected get index type %d",
417 (int) stack_top
->u
.ptr
.object_type
);
422 case LOAD_ROOT_CONTEXT
:
424 case LOAD_ROOT_APP_CONTEXT
:
426 ret
= context_get_index(lttng_probe_ctx
,
434 case LOAD_ROOT_PAYLOAD
:
435 stack_top
->u
.ptr
.ptr
+= gid
->offset
;
436 if (gid
->elem
.type
== OBJECT_TYPE_STRING
)
437 stack_top
->u
.ptr
.ptr
= *(const char * const *) stack_top
->u
.ptr
.ptr
;
438 stack_top
->u
.ptr
.object_type
= gid
->elem
.type
;
439 stack_top
->u
.ptr
.type
= LOAD_OBJECT
;
440 /* field is only used for types nested within variants. */
441 stack_top
->u
.ptr
.field
= NULL
;
450 static int dynamic_load_field(struct estack_entry
*stack_top
)
454 switch (stack_top
->u
.ptr
.type
) {
457 case LOAD_ROOT_CONTEXT
:
458 case LOAD_ROOT_APP_CONTEXT
:
459 case LOAD_ROOT_PAYLOAD
:
461 dbg_printk("Filter warning: cannot load root, missing field name.\n");
465 switch (stack_top
->u
.ptr
.object_type
) {
467 dbg_printk("op load field s8\n");
468 stack_top
->u
.v
= *(int8_t *) stack_top
->u
.ptr
.ptr
;
470 case OBJECT_TYPE_S16
:
474 dbg_printk("op load field s16\n");
475 tmp
= *(int16_t *) stack_top
->u
.ptr
.ptr
;
476 if (stack_top
->u
.ptr
.rev_bo
)
478 stack_top
->u
.v
= tmp
;
481 case OBJECT_TYPE_S32
:
485 dbg_printk("op load field s32\n");
486 tmp
= *(int32_t *) stack_top
->u
.ptr
.ptr
;
487 if (stack_top
->u
.ptr
.rev_bo
)
489 stack_top
->u
.v
= tmp
;
492 case OBJECT_TYPE_S64
:
496 dbg_printk("op load field s64\n");
497 tmp
= *(int64_t *) stack_top
->u
.ptr
.ptr
;
498 if (stack_top
->u
.ptr
.rev_bo
)
500 stack_top
->u
.v
= tmp
;
504 dbg_printk("op load field u8\n");
505 stack_top
->u
.v
= *(uint8_t *) stack_top
->u
.ptr
.ptr
;
507 case OBJECT_TYPE_U16
:
511 dbg_printk("op load field s16\n");
512 tmp
= *(uint16_t *) stack_top
->u
.ptr
.ptr
;
513 if (stack_top
->u
.ptr
.rev_bo
)
515 stack_top
->u
.v
= tmp
;
518 case OBJECT_TYPE_U32
:
522 dbg_printk("op load field u32\n");
523 tmp
= *(uint32_t *) stack_top
->u
.ptr
.ptr
;
524 if (stack_top
->u
.ptr
.rev_bo
)
526 stack_top
->u
.v
= tmp
;
529 case OBJECT_TYPE_U64
:
533 dbg_printk("op load field u64\n");
534 tmp
= *(uint64_t *) stack_top
->u
.ptr
.ptr
;
535 if (stack_top
->u
.ptr
.rev_bo
)
537 stack_top
->u
.v
= tmp
;
540 case OBJECT_TYPE_STRING
:
544 dbg_printk("op load field string\n");
545 str
= (const char *) stack_top
->u
.ptr
.ptr
;
546 stack_top
->u
.s
.str
= str
;
547 if (unlikely(!stack_top
->u
.s
.str
)) {
548 dbg_printk("Filter warning: loading a NULL string.\n");
552 stack_top
->u
.s
.seq_len
= LTTNG_SIZE_MAX
;
553 stack_top
->u
.s
.literal_type
=
554 ESTACK_STRING_LITERAL_TYPE_NONE
;
557 case OBJECT_TYPE_STRING_SEQUENCE
:
561 dbg_printk("op load field string sequence\n");
562 ptr
= stack_top
->u
.ptr
.ptr
;
563 stack_top
->u
.s
.seq_len
= *(unsigned long *) ptr
;
564 stack_top
->u
.s
.str
= *(const char **) (ptr
+ sizeof(unsigned long));
565 if (unlikely(!stack_top
->u
.s
.str
)) {
566 dbg_printk("Filter warning: loading a NULL sequence.\n");
570 stack_top
->u
.s
.literal_type
=
571 ESTACK_STRING_LITERAL_TYPE_NONE
;
574 case OBJECT_TYPE_DYNAMIC
:
576 * Dynamic types in context are looked up
577 * by context get index.
581 case OBJECT_TYPE_DOUBLE
:
584 case OBJECT_TYPE_SEQUENCE
:
585 case OBJECT_TYPE_ARRAY
:
586 case OBJECT_TYPE_STRUCT
:
587 case OBJECT_TYPE_VARIANT
:
588 printk(KERN_WARNING
"Sequences, arrays, struct and variant cannot be loaded (nested types).\n");
600 #define DBG_USER_STR_CUTOFF 32
603 * In debug mode, print user string (truncated, if necessary).
606 void dbg_load_ref_user_str_printk(const struct estack_entry
*user_str_reg
)
610 char user_str
[DBG_USER_STR_CUTOFF
];
614 last_char
= get_char(user_str_reg
, pos
);
615 user_str
[pos
] = last_char
;
617 } while (last_char
!= '\0' && pos
< sizeof(user_str
));
620 user_str
[sizeof(user_str
) - 1] = '\0';
621 dbg_printk("load field ref user string: '%s%s'\n", user_str
,
622 last_char
!= '\0' ? "[...]" : "");
626 void dbg_load_ref_user_str_printk(const struct estack_entry
*user_str_reg
)
632 * Return 0 (discard), or raise the 0x1 flag (log event).
633 * Currently, other flags are kept for future extensions and have no
636 uint64_t lttng_filter_interpret_bytecode(void *filter_data
,
637 struct lttng_probe_ctx
*lttng_probe_ctx
,
638 const char *filter_stack_data
)
640 struct bytecode_runtime
*bytecode
= filter_data
;
641 void *pc
, *next_pc
, *start_pc
;
644 struct estack _stack
;
645 struct estack
*stack
= &_stack
;
646 register int64_t ax
= 0, bx
= 0;
647 register int top
= FILTER_STACK_EMPTY
;
648 #ifndef INTERPRETER_USE_SWITCH
649 static void *dispatch
[NR_FILTER_OPS
] = {
650 [ FILTER_OP_UNKNOWN
] = &&LABEL_FILTER_OP_UNKNOWN
,
652 [ FILTER_OP_RETURN
] = &&LABEL_FILTER_OP_RETURN
,
655 [ FILTER_OP_MUL
] = &&LABEL_FILTER_OP_MUL
,
656 [ FILTER_OP_DIV
] = &&LABEL_FILTER_OP_DIV
,
657 [ FILTER_OP_MOD
] = &&LABEL_FILTER_OP_MOD
,
658 [ FILTER_OP_PLUS
] = &&LABEL_FILTER_OP_PLUS
,
659 [ FILTER_OP_MINUS
] = &&LABEL_FILTER_OP_MINUS
,
660 [ FILTER_OP_BIT_RSHIFT
] = &&LABEL_FILTER_OP_BIT_RSHIFT
,
661 [ FILTER_OP_BIT_LSHIFT
] = &&LABEL_FILTER_OP_BIT_LSHIFT
,
662 [ FILTER_OP_BIT_AND
] = &&LABEL_FILTER_OP_BIT_AND
,
663 [ FILTER_OP_BIT_OR
] = &&LABEL_FILTER_OP_BIT_OR
,
664 [ FILTER_OP_BIT_XOR
] = &&LABEL_FILTER_OP_BIT_XOR
,
666 /* binary comparators */
667 [ FILTER_OP_EQ
] = &&LABEL_FILTER_OP_EQ
,
668 [ FILTER_OP_NE
] = &&LABEL_FILTER_OP_NE
,
669 [ FILTER_OP_GT
] = &&LABEL_FILTER_OP_GT
,
670 [ FILTER_OP_LT
] = &&LABEL_FILTER_OP_LT
,
671 [ FILTER_OP_GE
] = &&LABEL_FILTER_OP_GE
,
672 [ FILTER_OP_LE
] = &&LABEL_FILTER_OP_LE
,
674 /* string binary comparator */
675 [ FILTER_OP_EQ_STRING
] = &&LABEL_FILTER_OP_EQ_STRING
,
676 [ FILTER_OP_NE_STRING
] = &&LABEL_FILTER_OP_NE_STRING
,
677 [ FILTER_OP_GT_STRING
] = &&LABEL_FILTER_OP_GT_STRING
,
678 [ FILTER_OP_LT_STRING
] = &&LABEL_FILTER_OP_LT_STRING
,
679 [ FILTER_OP_GE_STRING
] = &&LABEL_FILTER_OP_GE_STRING
,
680 [ FILTER_OP_LE_STRING
] = &&LABEL_FILTER_OP_LE_STRING
,
682 /* globbing pattern binary comparator */
683 [ FILTER_OP_EQ_STAR_GLOB_STRING
] = &&LABEL_FILTER_OP_EQ_STAR_GLOB_STRING
,
684 [ FILTER_OP_NE_STAR_GLOB_STRING
] = &&LABEL_FILTER_OP_NE_STAR_GLOB_STRING
,
686 /* s64 binary comparator */
687 [ FILTER_OP_EQ_S64
] = &&LABEL_FILTER_OP_EQ_S64
,
688 [ FILTER_OP_NE_S64
] = &&LABEL_FILTER_OP_NE_S64
,
689 [ FILTER_OP_GT_S64
] = &&LABEL_FILTER_OP_GT_S64
,
690 [ FILTER_OP_LT_S64
] = &&LABEL_FILTER_OP_LT_S64
,
691 [ FILTER_OP_GE_S64
] = &&LABEL_FILTER_OP_GE_S64
,
692 [ FILTER_OP_LE_S64
] = &&LABEL_FILTER_OP_LE_S64
,
694 /* double binary comparator */
695 [ FILTER_OP_EQ_DOUBLE
] = &&LABEL_FILTER_OP_EQ_DOUBLE
,
696 [ FILTER_OP_NE_DOUBLE
] = &&LABEL_FILTER_OP_NE_DOUBLE
,
697 [ FILTER_OP_GT_DOUBLE
] = &&LABEL_FILTER_OP_GT_DOUBLE
,
698 [ FILTER_OP_LT_DOUBLE
] = &&LABEL_FILTER_OP_LT_DOUBLE
,
699 [ FILTER_OP_GE_DOUBLE
] = &&LABEL_FILTER_OP_GE_DOUBLE
,
700 [ FILTER_OP_LE_DOUBLE
] = &&LABEL_FILTER_OP_LE_DOUBLE
,
702 /* Mixed S64-double binary comparators */
703 [ FILTER_OP_EQ_DOUBLE_S64
] = &&LABEL_FILTER_OP_EQ_DOUBLE_S64
,
704 [ FILTER_OP_NE_DOUBLE_S64
] = &&LABEL_FILTER_OP_NE_DOUBLE_S64
,
705 [ FILTER_OP_GT_DOUBLE_S64
] = &&LABEL_FILTER_OP_GT_DOUBLE_S64
,
706 [ FILTER_OP_LT_DOUBLE_S64
] = &&LABEL_FILTER_OP_LT_DOUBLE_S64
,
707 [ FILTER_OP_GE_DOUBLE_S64
] = &&LABEL_FILTER_OP_GE_DOUBLE_S64
,
708 [ FILTER_OP_LE_DOUBLE_S64
] = &&LABEL_FILTER_OP_LE_DOUBLE_S64
,
710 [ FILTER_OP_EQ_S64_DOUBLE
] = &&LABEL_FILTER_OP_EQ_S64_DOUBLE
,
711 [ FILTER_OP_NE_S64_DOUBLE
] = &&LABEL_FILTER_OP_NE_S64_DOUBLE
,
712 [ FILTER_OP_GT_S64_DOUBLE
] = &&LABEL_FILTER_OP_GT_S64_DOUBLE
,
713 [ FILTER_OP_LT_S64_DOUBLE
] = &&LABEL_FILTER_OP_LT_S64_DOUBLE
,
714 [ FILTER_OP_GE_S64_DOUBLE
] = &&LABEL_FILTER_OP_GE_S64_DOUBLE
,
715 [ FILTER_OP_LE_S64_DOUBLE
] = &&LABEL_FILTER_OP_LE_S64_DOUBLE
,
718 [ FILTER_OP_UNARY_PLUS
] = &&LABEL_FILTER_OP_UNARY_PLUS
,
719 [ FILTER_OP_UNARY_MINUS
] = &&LABEL_FILTER_OP_UNARY_MINUS
,
720 [ FILTER_OP_UNARY_NOT
] = &&LABEL_FILTER_OP_UNARY_NOT
,
721 [ FILTER_OP_UNARY_PLUS_S64
] = &&LABEL_FILTER_OP_UNARY_PLUS_S64
,
722 [ FILTER_OP_UNARY_MINUS_S64
] = &&LABEL_FILTER_OP_UNARY_MINUS_S64
,
723 [ FILTER_OP_UNARY_NOT_S64
] = &&LABEL_FILTER_OP_UNARY_NOT_S64
,
724 [ FILTER_OP_UNARY_PLUS_DOUBLE
] = &&LABEL_FILTER_OP_UNARY_PLUS_DOUBLE
,
725 [ FILTER_OP_UNARY_MINUS_DOUBLE
] = &&LABEL_FILTER_OP_UNARY_MINUS_DOUBLE
,
726 [ FILTER_OP_UNARY_NOT_DOUBLE
] = &&LABEL_FILTER_OP_UNARY_NOT_DOUBLE
,
729 [ FILTER_OP_AND
] = &&LABEL_FILTER_OP_AND
,
730 [ FILTER_OP_OR
] = &&LABEL_FILTER_OP_OR
,
733 [ FILTER_OP_LOAD_FIELD_REF
] = &&LABEL_FILTER_OP_LOAD_FIELD_REF
,
734 [ FILTER_OP_LOAD_FIELD_REF_STRING
] = &&LABEL_FILTER_OP_LOAD_FIELD_REF_STRING
,
735 [ FILTER_OP_LOAD_FIELD_REF_SEQUENCE
] = &&LABEL_FILTER_OP_LOAD_FIELD_REF_SEQUENCE
,
736 [ FILTER_OP_LOAD_FIELD_REF_S64
] = &&LABEL_FILTER_OP_LOAD_FIELD_REF_S64
,
737 [ FILTER_OP_LOAD_FIELD_REF_DOUBLE
] = &&LABEL_FILTER_OP_LOAD_FIELD_REF_DOUBLE
,
739 /* load from immediate operand */
740 [ FILTER_OP_LOAD_STRING
] = &&LABEL_FILTER_OP_LOAD_STRING
,
741 [ FILTER_OP_LOAD_STAR_GLOB_STRING
] = &&LABEL_FILTER_OP_LOAD_STAR_GLOB_STRING
,
742 [ FILTER_OP_LOAD_S64
] = &&LABEL_FILTER_OP_LOAD_S64
,
743 [ FILTER_OP_LOAD_DOUBLE
] = &&LABEL_FILTER_OP_LOAD_DOUBLE
,
746 [ FILTER_OP_CAST_TO_S64
] = &&LABEL_FILTER_OP_CAST_TO_S64
,
747 [ FILTER_OP_CAST_DOUBLE_TO_S64
] = &&LABEL_FILTER_OP_CAST_DOUBLE_TO_S64
,
748 [ FILTER_OP_CAST_NOP
] = &&LABEL_FILTER_OP_CAST_NOP
,
750 /* get context ref */
751 [ FILTER_OP_GET_CONTEXT_REF
] = &&LABEL_FILTER_OP_GET_CONTEXT_REF
,
752 [ FILTER_OP_GET_CONTEXT_REF_STRING
] = &&LABEL_FILTER_OP_GET_CONTEXT_REF_STRING
,
753 [ FILTER_OP_GET_CONTEXT_REF_S64
] = &&LABEL_FILTER_OP_GET_CONTEXT_REF_S64
,
754 [ FILTER_OP_GET_CONTEXT_REF_DOUBLE
] = &&LABEL_FILTER_OP_GET_CONTEXT_REF_DOUBLE
,
756 /* load userspace field ref */
757 [ FILTER_OP_LOAD_FIELD_REF_USER_STRING
] = &&LABEL_FILTER_OP_LOAD_FIELD_REF_USER_STRING
,
758 [ FILTER_OP_LOAD_FIELD_REF_USER_SEQUENCE
] = &&LABEL_FILTER_OP_LOAD_FIELD_REF_USER_SEQUENCE
,
760 /* Instructions for recursive traversal through composed types. */
761 [ FILTER_OP_GET_CONTEXT_ROOT
] = &&LABEL_FILTER_OP_GET_CONTEXT_ROOT
,
762 [ FILTER_OP_GET_APP_CONTEXT_ROOT
] = &&LABEL_FILTER_OP_GET_APP_CONTEXT_ROOT
,
763 [ FILTER_OP_GET_PAYLOAD_ROOT
] = &&LABEL_FILTER_OP_GET_PAYLOAD_ROOT
,
765 [ FILTER_OP_GET_SYMBOL
] = &&LABEL_FILTER_OP_GET_SYMBOL
,
766 [ FILTER_OP_GET_SYMBOL_FIELD
] = &&LABEL_FILTER_OP_GET_SYMBOL_FIELD
,
767 [ FILTER_OP_GET_INDEX_U16
] = &&LABEL_FILTER_OP_GET_INDEX_U16
,
768 [ FILTER_OP_GET_INDEX_U64
] = &&LABEL_FILTER_OP_GET_INDEX_U64
,
770 [ FILTER_OP_LOAD_FIELD
] = &&LABEL_FILTER_OP_LOAD_FIELD
,
771 [ FILTER_OP_LOAD_FIELD_S8
] = &&LABEL_FILTER_OP_LOAD_FIELD_S8
,
772 [ FILTER_OP_LOAD_FIELD_S16
] = &&LABEL_FILTER_OP_LOAD_FIELD_S16
,
773 [ FILTER_OP_LOAD_FIELD_S32
] = &&LABEL_FILTER_OP_LOAD_FIELD_S32
,
774 [ FILTER_OP_LOAD_FIELD_S64
] = &&LABEL_FILTER_OP_LOAD_FIELD_S64
,
775 [ FILTER_OP_LOAD_FIELD_U8
] = &&LABEL_FILTER_OP_LOAD_FIELD_U8
,
776 [ FILTER_OP_LOAD_FIELD_U16
] = &&LABEL_FILTER_OP_LOAD_FIELD_U16
,
777 [ FILTER_OP_LOAD_FIELD_U32
] = &&LABEL_FILTER_OP_LOAD_FIELD_U32
,
778 [ FILTER_OP_LOAD_FIELD_U64
] = &&LABEL_FILTER_OP_LOAD_FIELD_U64
,
779 [ FILTER_OP_LOAD_FIELD_STRING
] = &&LABEL_FILTER_OP_LOAD_FIELD_STRING
,
780 [ FILTER_OP_LOAD_FIELD_SEQUENCE
] = &&LABEL_FILTER_OP_LOAD_FIELD_SEQUENCE
,
781 [ FILTER_OP_LOAD_FIELD_DOUBLE
] = &&LABEL_FILTER_OP_LOAD_FIELD_DOUBLE
,
783 [ FILTER_OP_UNARY_BIT_NOT
] = &&LABEL_FILTER_OP_UNARY_BIT_NOT
,
785 [ FILTER_OP_RETURN_S64
] = &&LABEL_FILTER_OP_RETURN_S64
,
787 #endif /* #ifndef INTERPRETER_USE_SWITCH */
791 OP(FILTER_OP_UNKNOWN
):
792 OP(FILTER_OP_LOAD_FIELD_REF
):
793 OP(FILTER_OP_GET_CONTEXT_REF
):
794 #ifdef INTERPRETER_USE_SWITCH
796 #endif /* INTERPRETER_USE_SWITCH */
797 printk(KERN_WARNING
"unknown bytecode op %u\n",
798 (unsigned int) *(filter_opcode_t
*) pc
);
802 OP(FILTER_OP_RETURN
):
803 OP(FILTER_OP_RETURN_S64
):
804 /* LTTNG_FILTER_DISCARD or LTTNG_FILTER_RECORD_FLAG */
805 retval
= !!estack_ax_v
;
815 printk(KERN_WARNING
"unsupported bytecode op %u\n",
816 (unsigned int) *(filter_opcode_t
*) pc
);
826 printk(KERN_WARNING
"unsupported non-specialized bytecode op %u\n",
827 (unsigned int) *(filter_opcode_t
*) pc
);
831 OP(FILTER_OP_EQ_STRING
):
835 res
= (stack_strcmp(stack
, top
, "==") == 0);
836 estack_pop(stack
, top
, ax
, bx
);
838 next_pc
+= sizeof(struct binary_op
);
841 OP(FILTER_OP_NE_STRING
):
845 res
= (stack_strcmp(stack
, top
, "!=") != 0);
846 estack_pop(stack
, top
, ax
, bx
);
848 next_pc
+= sizeof(struct binary_op
);
851 OP(FILTER_OP_GT_STRING
):
855 res
= (stack_strcmp(stack
, top
, ">") > 0);
856 estack_pop(stack
, top
, ax
, bx
);
858 next_pc
+= sizeof(struct binary_op
);
861 OP(FILTER_OP_LT_STRING
):
865 res
= (stack_strcmp(stack
, top
, "<") < 0);
866 estack_pop(stack
, top
, ax
, bx
);
868 next_pc
+= sizeof(struct binary_op
);
871 OP(FILTER_OP_GE_STRING
):
875 res
= (stack_strcmp(stack
, top
, ">=") >= 0);
876 estack_pop(stack
, top
, ax
, bx
);
878 next_pc
+= sizeof(struct binary_op
);
881 OP(FILTER_OP_LE_STRING
):
885 res
= (stack_strcmp(stack
, top
, "<=") <= 0);
886 estack_pop(stack
, top
, ax
, bx
);
888 next_pc
+= sizeof(struct binary_op
);
892 OP(FILTER_OP_EQ_STAR_GLOB_STRING
):
896 res
= (stack_star_glob_match(stack
, top
, "==") == 0);
897 estack_pop(stack
, top
, ax
, bx
);
899 next_pc
+= sizeof(struct binary_op
);
902 OP(FILTER_OP_NE_STAR_GLOB_STRING
):
906 res
= (stack_star_glob_match(stack
, top
, "!=") != 0);
907 estack_pop(stack
, top
, ax
, bx
);
909 next_pc
+= sizeof(struct binary_op
);
913 OP(FILTER_OP_EQ_S64
):
917 res
= (estack_bx_v
== estack_ax_v
);
918 estack_pop(stack
, top
, ax
, bx
);
920 next_pc
+= sizeof(struct binary_op
);
923 OP(FILTER_OP_NE_S64
):
927 res
= (estack_bx_v
!= estack_ax_v
);
928 estack_pop(stack
, top
, ax
, bx
);
930 next_pc
+= sizeof(struct binary_op
);
933 OP(FILTER_OP_GT_S64
):
937 res
= (estack_bx_v
> estack_ax_v
);
938 estack_pop(stack
, top
, ax
, bx
);
940 next_pc
+= sizeof(struct binary_op
);
943 OP(FILTER_OP_LT_S64
):
947 res
= (estack_bx_v
< estack_ax_v
);
948 estack_pop(stack
, top
, ax
, bx
);
950 next_pc
+= sizeof(struct binary_op
);
953 OP(FILTER_OP_GE_S64
):
957 res
= (estack_bx_v
>= estack_ax_v
);
958 estack_pop(stack
, top
, ax
, bx
);
960 next_pc
+= sizeof(struct binary_op
);
963 OP(FILTER_OP_LE_S64
):
967 res
= (estack_bx_v
<= estack_ax_v
);
968 estack_pop(stack
, top
, ax
, bx
);
970 next_pc
+= sizeof(struct binary_op
);
974 OP(FILTER_OP_EQ_DOUBLE
):
975 OP(FILTER_OP_NE_DOUBLE
):
976 OP(FILTER_OP_GT_DOUBLE
):
977 OP(FILTER_OP_LT_DOUBLE
):
978 OP(FILTER_OP_GE_DOUBLE
):
979 OP(FILTER_OP_LE_DOUBLE
):
985 /* Mixed S64-double binary comparators */
986 OP(FILTER_OP_EQ_DOUBLE_S64
):
987 OP(FILTER_OP_NE_DOUBLE_S64
):
988 OP(FILTER_OP_GT_DOUBLE_S64
):
989 OP(FILTER_OP_LT_DOUBLE_S64
):
990 OP(FILTER_OP_GE_DOUBLE_S64
):
991 OP(FILTER_OP_LE_DOUBLE_S64
):
992 OP(FILTER_OP_EQ_S64_DOUBLE
):
993 OP(FILTER_OP_NE_S64_DOUBLE
):
994 OP(FILTER_OP_GT_S64_DOUBLE
):
995 OP(FILTER_OP_LT_S64_DOUBLE
):
996 OP(FILTER_OP_GE_S64_DOUBLE
):
997 OP(FILTER_OP_LE_S64_DOUBLE
):
1002 OP(FILTER_OP_BIT_RSHIFT
):
1006 /* Catch undefined behavior. */
1007 if (unlikely(estack_ax_v
< 0 || estack_ax_v
>= 64)) {
1011 res
= ((uint64_t) estack_bx_v
>> (uint32_t) estack_ax_v
);
1012 estack_pop(stack
, top
, ax
, bx
);
1014 next_pc
+= sizeof(struct binary_op
);
1017 OP(FILTER_OP_BIT_LSHIFT
):
1021 /* Catch undefined behavior. */
1022 if (unlikely(estack_ax_v
< 0 || estack_ax_v
>= 64)) {
1026 res
= ((uint64_t) estack_bx_v
<< (uint32_t) estack_ax_v
);
1027 estack_pop(stack
, top
, ax
, bx
);
1029 next_pc
+= sizeof(struct binary_op
);
1032 OP(FILTER_OP_BIT_AND
):
1036 res
= ((uint64_t) estack_bx_v
& (uint64_t) estack_ax_v
);
1037 estack_pop(stack
, top
, ax
, bx
);
1039 next_pc
+= sizeof(struct binary_op
);
1042 OP(FILTER_OP_BIT_OR
):
1046 res
= ((uint64_t) estack_bx_v
| (uint64_t) estack_ax_v
);
1047 estack_pop(stack
, top
, ax
, bx
);
1049 next_pc
+= sizeof(struct binary_op
);
1052 OP(FILTER_OP_BIT_XOR
):
1056 res
= ((uint64_t) estack_bx_v
^ (uint64_t) estack_ax_v
);
1057 estack_pop(stack
, top
, ax
, bx
);
1059 next_pc
+= sizeof(struct binary_op
);
1064 OP(FILTER_OP_UNARY_PLUS
):
1065 OP(FILTER_OP_UNARY_MINUS
):
1066 OP(FILTER_OP_UNARY_NOT
):
1067 printk(KERN_WARNING
"unsupported non-specialized bytecode op %u\n",
1068 (unsigned int) *(filter_opcode_t
*) pc
);
1073 OP(FILTER_OP_UNARY_BIT_NOT
):
1075 estack_ax_v
= ~(uint64_t) estack_ax_v
;
1076 next_pc
+= sizeof(struct unary_op
);
1080 OP(FILTER_OP_UNARY_PLUS_S64
):
1082 next_pc
+= sizeof(struct unary_op
);
1085 OP(FILTER_OP_UNARY_MINUS_S64
):
1087 estack_ax_v
= -estack_ax_v
;
1088 next_pc
+= sizeof(struct unary_op
);
1091 OP(FILTER_OP_UNARY_PLUS_DOUBLE
):
1092 OP(FILTER_OP_UNARY_MINUS_DOUBLE
):
1097 OP(FILTER_OP_UNARY_NOT_S64
):
1099 estack_ax_v
= !estack_ax_v
;
1100 next_pc
+= sizeof(struct unary_op
);
1103 OP(FILTER_OP_UNARY_NOT_DOUBLE
):
1112 struct logical_op
*insn
= (struct logical_op
*) pc
;
1114 /* If AX is 0, skip and evaluate to 0 */
1115 if (unlikely(estack_ax_v
== 0)) {
1116 dbg_printk("Jumping to bytecode offset %u\n",
1117 (unsigned int) insn
->skip_offset
);
1118 next_pc
= start_pc
+ insn
->skip_offset
;
1120 /* Pop 1 when jump not taken */
1121 estack_pop(stack
, top
, ax
, bx
);
1122 next_pc
+= sizeof(struct logical_op
);
1128 struct logical_op
*insn
= (struct logical_op
*) pc
;
1130 /* If AX is nonzero, skip and evaluate to 1 */
1132 if (unlikely(estack_ax_v
!= 0)) {
1134 dbg_printk("Jumping to bytecode offset %u\n",
1135 (unsigned int) insn
->skip_offset
);
1136 next_pc
= start_pc
+ insn
->skip_offset
;
1138 /* Pop 1 when jump not taken */
1139 estack_pop(stack
, top
, ax
, bx
);
1140 next_pc
+= sizeof(struct logical_op
);
1146 /* load field ref */
1147 OP(FILTER_OP_LOAD_FIELD_REF_STRING
):
1149 struct load_op
*insn
= (struct load_op
*) pc
;
1150 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
1152 dbg_printk("load field ref offset %u type string\n",
1154 estack_push(stack
, top
, ax
, bx
);
1155 estack_ax(stack
, top
)->u
.s
.str
=
1156 *(const char * const *) &filter_stack_data
[ref
->offset
];
1157 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
1158 dbg_printk("Filter warning: loading a NULL string.\n");
1162 estack_ax(stack
, top
)->u
.s
.seq_len
= LTTNG_SIZE_MAX
;
1163 estack_ax(stack
, top
)->u
.s
.literal_type
=
1164 ESTACK_STRING_LITERAL_TYPE_NONE
;
1165 estack_ax(stack
, top
)->u
.s
.user
= 0;
1166 dbg_printk("ref load string %s\n", estack_ax(stack
, top
)->u
.s
.str
);
1167 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
1171 OP(FILTER_OP_LOAD_FIELD_REF_SEQUENCE
):
1173 struct load_op
*insn
= (struct load_op
*) pc
;
1174 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
1176 dbg_printk("load field ref offset %u type sequence\n",
1178 estack_push(stack
, top
, ax
, bx
);
1179 estack_ax(stack
, top
)->u
.s
.seq_len
=
1180 *(unsigned long *) &filter_stack_data
[ref
->offset
];
1181 estack_ax(stack
, top
)->u
.s
.str
=
1182 *(const char **) (&filter_stack_data
[ref
->offset
1183 + sizeof(unsigned long)]);
1184 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
1185 dbg_printk("Filter warning: loading a NULL sequence.\n");
1189 estack_ax(stack
, top
)->u
.s
.literal_type
=
1190 ESTACK_STRING_LITERAL_TYPE_NONE
;
1191 estack_ax(stack
, top
)->u
.s
.user
= 0;
1192 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
1196 OP(FILTER_OP_LOAD_FIELD_REF_S64
):
1198 struct load_op
*insn
= (struct load_op
*) pc
;
1199 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
1201 dbg_printk("load field ref offset %u type s64\n",
1203 estack_push(stack
, top
, ax
, bx
);
1205 ((struct literal_numeric
*) &filter_stack_data
[ref
->offset
])->v
;
1206 dbg_printk("ref load s64 %lld\n",
1207 (long long) estack_ax_v
);
1208 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
1212 OP(FILTER_OP_LOAD_FIELD_REF_DOUBLE
):
1218 /* load from immediate operand */
1219 OP(FILTER_OP_LOAD_STRING
):
1221 struct load_op
*insn
= (struct load_op
*) pc
;
1223 dbg_printk("load string %s\n", insn
->data
);
1224 estack_push(stack
, top
, ax
, bx
);
1225 estack_ax(stack
, top
)->u
.s
.str
= insn
->data
;
1226 estack_ax(stack
, top
)->u
.s
.seq_len
= LTTNG_SIZE_MAX
;
1227 estack_ax(stack
, top
)->u
.s
.literal_type
=
1228 ESTACK_STRING_LITERAL_TYPE_PLAIN
;
1229 estack_ax(stack
, top
)->u
.s
.user
= 0;
1230 next_pc
+= sizeof(struct load_op
) + strlen(insn
->data
) + 1;
1234 OP(FILTER_OP_LOAD_STAR_GLOB_STRING
):
1236 struct load_op
*insn
= (struct load_op
*) pc
;
1238 dbg_printk("load globbing pattern %s\n", insn
->data
);
1239 estack_push(stack
, top
, ax
, bx
);
1240 estack_ax(stack
, top
)->u
.s
.str
= insn
->data
;
1241 estack_ax(stack
, top
)->u
.s
.seq_len
= LTTNG_SIZE_MAX
;
1242 estack_ax(stack
, top
)->u
.s
.literal_type
=
1243 ESTACK_STRING_LITERAL_TYPE_STAR_GLOB
;
1244 estack_ax(stack
, top
)->u
.s
.user
= 0;
1245 next_pc
+= sizeof(struct load_op
) + strlen(insn
->data
) + 1;
1249 OP(FILTER_OP_LOAD_S64
):
1251 struct load_op
*insn
= (struct load_op
*) pc
;
1253 estack_push(stack
, top
, ax
, bx
);
1254 estack_ax_v
= ((struct literal_numeric
*) insn
->data
)->v
;
1255 dbg_printk("load s64 %lld\n",
1256 (long long) estack_ax_v
);
1257 next_pc
+= sizeof(struct load_op
)
1258 + sizeof(struct literal_numeric
);
1262 OP(FILTER_OP_LOAD_DOUBLE
):
1269 OP(FILTER_OP_CAST_TO_S64
):
1270 printk(KERN_WARNING
"unsupported non-specialized bytecode op %u\n",
1271 (unsigned int) *(filter_opcode_t
*) pc
);
1275 OP(FILTER_OP_CAST_DOUBLE_TO_S64
):
1281 OP(FILTER_OP_CAST_NOP
):
1283 next_pc
+= sizeof(struct cast_op
);
1287 /* get context ref */
1288 OP(FILTER_OP_GET_CONTEXT_REF_STRING
):
1290 struct load_op
*insn
= (struct load_op
*) pc
;
1291 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
1292 struct lttng_ctx_field
*ctx_field
;
1293 union lttng_ctx_value v
;
1295 dbg_printk("get context ref offset %u type string\n",
1297 ctx_field
= <tng_static_ctx
->fields
[ref
->offset
];
1298 ctx_field
->get_value(ctx_field
, lttng_probe_ctx
, &v
);
1299 estack_push(stack
, top
, ax
, bx
);
1300 estack_ax(stack
, top
)->u
.s
.str
= v
.str
;
1301 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
1302 dbg_printk("Filter warning: loading a NULL string.\n");
1306 estack_ax(stack
, top
)->u
.s
.seq_len
= LTTNG_SIZE_MAX
;
1307 estack_ax(stack
, top
)->u
.s
.literal_type
=
1308 ESTACK_STRING_LITERAL_TYPE_NONE
;
1309 estack_ax(stack
, top
)->u
.s
.user
= 0;
1310 dbg_printk("ref get context string %s\n", estack_ax(stack
, top
)->u
.s
.str
);
1311 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
1315 OP(FILTER_OP_GET_CONTEXT_REF_S64
):
1317 struct load_op
*insn
= (struct load_op
*) pc
;
1318 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
1319 struct lttng_ctx_field
*ctx_field
;
1320 union lttng_ctx_value v
;
1322 dbg_printk("get context ref offset %u type s64\n",
1324 ctx_field
= <tng_static_ctx
->fields
[ref
->offset
];
1325 ctx_field
->get_value(ctx_field
, lttng_probe_ctx
, &v
);
1326 estack_push(stack
, top
, ax
, bx
);
1327 estack_ax_v
= v
.s64
;
1328 dbg_printk("ref get context s64 %lld\n",
1329 (long long) estack_ax_v
);
1330 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
1334 OP(FILTER_OP_GET_CONTEXT_REF_DOUBLE
):
1340 /* load userspace field ref */
1341 OP(FILTER_OP_LOAD_FIELD_REF_USER_STRING
):
1343 struct load_op
*insn
= (struct load_op
*) pc
;
1344 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
1346 dbg_printk("load field ref offset %u type user string\n",
1348 estack_push(stack
, top
, ax
, bx
);
1349 estack_ax(stack
, top
)->u
.s
.user_str
=
1350 *(const char * const *) &filter_stack_data
[ref
->offset
];
1351 if (unlikely(!estack_ax(stack
, top
)->u
.s
.user_str
)) {
1352 dbg_printk("Filter warning: loading a NULL string.\n");
1356 estack_ax(stack
, top
)->u
.s
.seq_len
= LTTNG_SIZE_MAX
;
1357 estack_ax(stack
, top
)->u
.s
.literal_type
=
1358 ESTACK_STRING_LITERAL_TYPE_NONE
;
1359 estack_ax(stack
, top
)->u
.s
.user
= 1;
1360 dbg_load_ref_user_str_printk(estack_ax(stack
, top
));
1361 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
1365 OP(FILTER_OP_LOAD_FIELD_REF_USER_SEQUENCE
):
1367 struct load_op
*insn
= (struct load_op
*) pc
;
1368 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
1370 dbg_printk("load field ref offset %u type user sequence\n",
1372 estack_push(stack
, top
, ax
, bx
);
1373 estack_ax(stack
, top
)->u
.s
.seq_len
=
1374 *(unsigned long *) &filter_stack_data
[ref
->offset
];
1375 estack_ax(stack
, top
)->u
.s
.user_str
=
1376 *(const char **) (&filter_stack_data
[ref
->offset
1377 + sizeof(unsigned long)]);
1378 if (unlikely(!estack_ax(stack
, top
)->u
.s
.user_str
)) {
1379 dbg_printk("Filter warning: loading a NULL sequence.\n");
1383 estack_ax(stack
, top
)->u
.s
.literal_type
=
1384 ESTACK_STRING_LITERAL_TYPE_NONE
;
1385 estack_ax(stack
, top
)->u
.s
.user
= 1;
1386 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
1390 OP(FILTER_OP_GET_CONTEXT_ROOT
):
1392 dbg_printk("op get context root\n");
1393 estack_push(stack
, top
, ax
, bx
);
1394 estack_ax(stack
, top
)->u
.ptr
.type
= LOAD_ROOT_CONTEXT
;
1395 /* "field" only needed for variants. */
1396 estack_ax(stack
, top
)->u
.ptr
.field
= NULL
;
1397 next_pc
+= sizeof(struct load_op
);
1401 OP(FILTER_OP_GET_APP_CONTEXT_ROOT
):
1407 OP(FILTER_OP_GET_PAYLOAD_ROOT
):
1409 dbg_printk("op get app payload root\n");
1410 estack_push(stack
, top
, ax
, bx
);
1411 estack_ax(stack
, top
)->u
.ptr
.type
= LOAD_ROOT_PAYLOAD
;
1412 estack_ax(stack
, top
)->u
.ptr
.ptr
= filter_stack_data
;
1413 /* "field" only needed for variants. */
1414 estack_ax(stack
, top
)->u
.ptr
.field
= NULL
;
1415 next_pc
+= sizeof(struct load_op
);
1419 OP(FILTER_OP_GET_SYMBOL
):
1421 dbg_printk("op get symbol\n");
1422 switch (estack_ax(stack
, top
)->u
.ptr
.type
) {
1424 printk(KERN_WARNING
"Nested fields not implemented yet.\n");
1427 case LOAD_ROOT_CONTEXT
:
1428 case LOAD_ROOT_APP_CONTEXT
:
1429 case LOAD_ROOT_PAYLOAD
:
1431 * symbol lookup is performed by
1437 next_pc
+= sizeof(struct load_op
) + sizeof(struct get_symbol
);
1441 OP(FILTER_OP_GET_SYMBOL_FIELD
):
1444 * Used for first variant encountered in a
1445 * traversal. Variants are not implemented yet.
1451 OP(FILTER_OP_GET_INDEX_U16
):
1453 struct load_op
*insn
= (struct load_op
*) pc
;
1454 struct get_index_u16
*index
= (struct get_index_u16
*) insn
->data
;
1456 dbg_printk("op get index u16\n");
1457 ret
= dynamic_get_index(lttng_probe_ctx
, bytecode
, index
->index
, estack_ax(stack
, top
));
1460 estack_ax_v
= estack_ax(stack
, top
)->u
.v
;
1461 next_pc
+= sizeof(struct load_op
) + sizeof(struct get_index_u16
);
1465 OP(FILTER_OP_GET_INDEX_U64
):
1467 struct load_op
*insn
= (struct load_op
*) pc
;
1468 struct get_index_u64
*index
= (struct get_index_u64
*) insn
->data
;
1470 dbg_printk("op get index u64\n");
1471 ret
= dynamic_get_index(lttng_probe_ctx
, bytecode
, index
->index
, estack_ax(stack
, top
));
1474 estack_ax_v
= estack_ax(stack
, top
)->u
.v
;
1475 next_pc
+= sizeof(struct load_op
) + sizeof(struct get_index_u64
);
1479 OP(FILTER_OP_LOAD_FIELD
):
1481 dbg_printk("op load field\n");
1482 ret
= dynamic_load_field(estack_ax(stack
, top
));
1485 estack_ax_v
= estack_ax(stack
, top
)->u
.v
;
1486 next_pc
+= sizeof(struct load_op
);
1490 OP(FILTER_OP_LOAD_FIELD_S8
):
1492 dbg_printk("op load field s8\n");
1494 estack_ax_v
= *(int8_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
1495 next_pc
+= sizeof(struct load_op
);
1498 OP(FILTER_OP_LOAD_FIELD_S16
):
1500 dbg_printk("op load field s16\n");
1502 estack_ax_v
= *(int16_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
1503 next_pc
+= sizeof(struct load_op
);
1506 OP(FILTER_OP_LOAD_FIELD_S32
):
1508 dbg_printk("op load field s32\n");
1510 estack_ax_v
= *(int32_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
1511 next_pc
+= sizeof(struct load_op
);
1514 OP(FILTER_OP_LOAD_FIELD_S64
):
1516 dbg_printk("op load field s64\n");
1518 estack_ax_v
= *(int64_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
1519 next_pc
+= sizeof(struct load_op
);
1522 OP(FILTER_OP_LOAD_FIELD_U8
):
1524 dbg_printk("op load field u8\n");
1526 estack_ax_v
= *(uint8_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
1527 next_pc
+= sizeof(struct load_op
);
1530 OP(FILTER_OP_LOAD_FIELD_U16
):
1532 dbg_printk("op load field u16\n");
1534 estack_ax_v
= *(uint16_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
1535 next_pc
+= sizeof(struct load_op
);
1538 OP(FILTER_OP_LOAD_FIELD_U32
):
1540 dbg_printk("op load field u32\n");
1542 estack_ax_v
= *(uint32_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
1543 next_pc
+= sizeof(struct load_op
);
1546 OP(FILTER_OP_LOAD_FIELD_U64
):
1548 dbg_printk("op load field u64\n");
1550 estack_ax_v
= *(uint64_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
1551 next_pc
+= sizeof(struct load_op
);
1554 OP(FILTER_OP_LOAD_FIELD_DOUBLE
):
1560 OP(FILTER_OP_LOAD_FIELD_STRING
):
1564 dbg_printk("op load field string\n");
1565 str
= (const char *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
1566 estack_ax(stack
, top
)->u
.s
.str
= str
;
1567 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
1568 dbg_printk("Filter warning: loading a NULL string.\n");
1572 estack_ax(stack
, top
)->u
.s
.seq_len
= LTTNG_SIZE_MAX
;
1573 estack_ax(stack
, top
)->u
.s
.literal_type
=
1574 ESTACK_STRING_LITERAL_TYPE_NONE
;
1575 next_pc
+= sizeof(struct load_op
);
1579 OP(FILTER_OP_LOAD_FIELD_SEQUENCE
):
1583 dbg_printk("op load field string sequence\n");
1584 ptr
= estack_ax(stack
, top
)->u
.ptr
.ptr
;
1585 estack_ax(stack
, top
)->u
.s
.seq_len
= *(unsigned long *) ptr
;
1586 estack_ax(stack
, top
)->u
.s
.str
= *(const char **) (ptr
+ sizeof(unsigned long));
1587 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
1588 dbg_printk("Filter warning: loading a NULL sequence.\n");
1592 estack_ax(stack
, top
)->u
.s
.literal_type
=
1593 ESTACK_STRING_LITERAL_TYPE_NONE
;
1594 next_pc
+= sizeof(struct load_op
);
1600 /* return 0 (discard) on error */