2 * lttng-bytecode-interpreter.c
4 * LTTng UST bytecode interpreter.
6 * Copyright (C) 2010-2016 Mathieu Desnoyers <mathieu.desnoyers@efficios.com>
8 * Permission is hereby granted, free of charge, to any person obtaining a copy
9 * of this software and associated documentation files (the "Software"), to deal
10 * in the Software without restriction, including without limitation the rights
11 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
12 * copies of the Software, and to permit persons to whom the Software is
13 * furnished to do so, subject to the following conditions:
15 * The above copyright notice and this permission notice shall be included in
16 * all copies or substantial portions of the Software.
18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
19 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
20 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
21 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
22 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
23 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
30 #include <urcu-pointer.h>
33 #include <lttng/ust-endian.h>
34 #include <lttng/ust-events.h>
36 #include "lttng-bytecode.h"
37 #include "string-utils.h"
42 * -2: unknown escape char.
47 int parse_char(const char **p
)
67 * Returns SIZE_MAX if the string is null-terminated, or the number of
71 size_t get_str_or_seq_len(const struct estack_entry
*entry
)
73 return entry
->u
.s
.seq_len
;
77 int stack_star_glob_match(struct estack
*stack
, int top
, const char *cmp_type
)
80 const char *candidate
;
84 /* Find out which side is the pattern vs. the candidate. */
85 if (estack_ax(stack
, top
)->u
.s
.literal_type
== ESTACK_STRING_LITERAL_TYPE_STAR_GLOB
) {
86 pattern
= estack_ax(stack
, top
)->u
.s
.str
;
87 pattern_len
= get_str_or_seq_len(estack_ax(stack
, top
));
88 candidate
= estack_bx(stack
, top
)->u
.s
.str
;
89 candidate_len
= get_str_or_seq_len(estack_bx(stack
, top
));
91 pattern
= estack_bx(stack
, top
)->u
.s
.str
;
92 pattern_len
= get_str_or_seq_len(estack_bx(stack
, top
));
93 candidate
= estack_ax(stack
, top
)->u
.s
.str
;
94 candidate_len
= get_str_or_seq_len(estack_ax(stack
, top
));
97 /* Perform the match. Returns 0 when the result is true. */
98 return !strutils_star_glob_match(pattern
, pattern_len
, candidate
,
103 int stack_strcmp(struct estack
*stack
, int top
, const char *cmp_type
)
105 const char *p
= estack_bx(stack
, top
)->u
.s
.str
, *q
= estack_ax(stack
, top
)->u
.s
.str
;
112 if (unlikely(p
- estack_bx(stack
, top
)->u
.s
.str
>= estack_bx(stack
, top
)->u
.s
.seq_len
|| *p
== '\0')) {
113 if (q
- estack_ax(stack
, top
)->u
.s
.str
>= estack_ax(stack
, top
)->u
.s
.seq_len
|| *q
== '\0') {
116 if (estack_ax(stack
, top
)->u
.s
.literal_type
==
117 ESTACK_STRING_LITERAL_TYPE_PLAIN
) {
118 ret
= parse_char(&q
);
125 if (unlikely(q
- estack_ax(stack
, top
)->u
.s
.str
>= estack_ax(stack
, top
)->u
.s
.seq_len
|| *q
== '\0')) {
126 if (estack_bx(stack
, top
)->u
.s
.literal_type
==
127 ESTACK_STRING_LITERAL_TYPE_PLAIN
) {
128 ret
= parse_char(&p
);
134 if (estack_bx(stack
, top
)->u
.s
.literal_type
==
135 ESTACK_STRING_LITERAL_TYPE_PLAIN
) {
136 ret
= parse_char(&p
);
139 } else if (ret
== -2) {
142 /* else compare both char */
144 if (estack_ax(stack
, top
)->u
.s
.literal_type
==
145 ESTACK_STRING_LITERAL_TYPE_PLAIN
) {
146 ret
= parse_char(&q
);
149 } else if (ret
== -2) {
169 uint64_t lttng_bytecode_filter_interpret_false(void *filter_data
,
170 const char *filter_stack_data
)
172 return LTTNG_INTERPRETER_DISCARD
;
175 #ifdef INTERPRETER_USE_SWITCH
178 * Fallback for compilers that do not support taking address of labels.
182 start_pc = &bytecode->data[0]; \
183 for (pc = next_pc = start_pc; pc - start_pc < bytecode->len; \
185 dbg_printf("Executing op %s (%u)\n", \
186 print_op((unsigned int) *(bytecode_opcode_t *) pc), \
187 (unsigned int) *(bytecode_opcode_t *) pc); \
188 switch (*(bytecode_opcode_t *) pc) {
190 #define OP(name) jump_target_##name: __attribute__((unused)); \
198 #define JUMP_TO(name) \
199 goto jump_target_##name
204 * Dispatch-table based interpreter.
208 start_pc = &bytecode->code[0]; \
209 pc = next_pc = start_pc; \
210 if (unlikely(pc - start_pc >= bytecode->len)) \
212 goto *dispatch[*(bytecode_opcode_t *) pc];
219 goto *dispatch[*(bytecode_opcode_t *) pc];
223 #define JUMP_TO(name) \
228 #define IS_INTEGER_REGISTER(reg_type) \
229 (reg_type == REG_U64 || reg_type == REG_S64)
231 static int context_get_index(struct lttng_ctx
*ctx
,
232 struct load_ptr
*ptr
,
236 struct lttng_ctx_field
*ctx_field
;
237 struct lttng_event_field
*field
;
238 struct lttng_ctx_value v
;
240 ctx_field
= &ctx
->fields
[idx
];
241 field
= &ctx_field
->event_field
;
242 ptr
->type
= LOAD_OBJECT
;
245 switch (field
->type
.atype
) {
247 ctx_field
->get_value(ctx_field
, &v
);
248 if (field
->type
.u
.integer
.signedness
) {
249 ptr
->object_type
= OBJECT_TYPE_S64
;
250 ptr
->u
.s64
= v
.u
.s64
;
251 ptr
->ptr
= &ptr
->u
.s64
;
253 ptr
->object_type
= OBJECT_TYPE_U64
;
254 ptr
->u
.u64
= v
.u
.s64
; /* Cast. */
255 ptr
->ptr
= &ptr
->u
.u64
;
258 case atype_enum
: /* Fall-through */
259 case atype_enum_nestable
:
261 const struct lttng_integer_type
*itype
;
263 if (field
->type
.atype
== atype_enum
) {
264 itype
= &field
->type
.u
.legacy
.basic
.enumeration
.container_type
;
266 itype
= &field
->type
.u
.enum_nestable
.container_type
->u
.integer
;
268 ctx_field
->get_value(ctx_field
, &v
);
269 if (itype
->signedness
) {
270 ptr
->object_type
= OBJECT_TYPE_SIGNED_ENUM
;
271 ptr
->u
.s64
= v
.u
.s64
;
272 ptr
->ptr
= &ptr
->u
.s64
;
274 ptr
->object_type
= OBJECT_TYPE_UNSIGNED_ENUM
;
275 ptr
->u
.u64
= v
.u
.s64
; /* Cast. */
276 ptr
->ptr
= &ptr
->u
.u64
;
281 if (field
->type
.u
.legacy
.array
.elem_type
.atype
!= atype_integer
) {
282 ERR("Array nesting only supports integer types.");
285 if (field
->type
.u
.legacy
.array
.elem_type
.u
.basic
.integer
.encoding
== lttng_encode_none
) {
286 ERR("Only string arrays are supported for contexts.");
289 ptr
->object_type
= OBJECT_TYPE_STRING
;
290 ctx_field
->get_value(ctx_field
, &v
);
293 case atype_array_nestable
:
294 if (field
->type
.u
.array_nestable
.elem_type
->atype
!= atype_integer
) {
295 ERR("Array nesting only supports integer types.");
298 if (field
->type
.u
.array_nestable
.elem_type
->u
.integer
.encoding
== lttng_encode_none
) {
299 ERR("Only string arrays are supported for contexts.");
302 ptr
->object_type
= OBJECT_TYPE_STRING
;
303 ctx_field
->get_value(ctx_field
, &v
);
307 if (field
->type
.u
.legacy
.sequence
.elem_type
.atype
!= atype_integer
) {
308 ERR("Sequence nesting only supports integer types.");
311 if (field
->type
.u
.legacy
.sequence
.elem_type
.u
.basic
.integer
.encoding
== lttng_encode_none
) {
312 ERR("Only string sequences are supported for contexts.");
315 ptr
->object_type
= OBJECT_TYPE_STRING
;
316 ctx_field
->get_value(ctx_field
, &v
);
319 case atype_sequence_nestable
:
320 if (field
->type
.u
.sequence_nestable
.elem_type
->atype
!= atype_integer
) {
321 ERR("Sequence nesting only supports integer types.");
324 if (field
->type
.u
.sequence_nestable
.elem_type
->u
.integer
.encoding
== lttng_encode_none
) {
325 ERR("Only string sequences are supported for contexts.");
328 ptr
->object_type
= OBJECT_TYPE_STRING
;
329 ctx_field
->get_value(ctx_field
, &v
);
333 ptr
->object_type
= OBJECT_TYPE_STRING
;
334 ctx_field
->get_value(ctx_field
, &v
);
338 ptr
->object_type
= OBJECT_TYPE_DOUBLE
;
339 ctx_field
->get_value(ctx_field
, &v
);
341 ptr
->ptr
= &ptr
->u
.d
;
344 ctx_field
->get_value(ctx_field
, &v
);
346 case LTTNG_UST_DYNAMIC_TYPE_NONE
:
348 case LTTNG_UST_DYNAMIC_TYPE_U8
:
349 case LTTNG_UST_DYNAMIC_TYPE_U16
:
350 case LTTNG_UST_DYNAMIC_TYPE_U32
:
351 case LTTNG_UST_DYNAMIC_TYPE_U64
:
352 ptr
->object_type
= OBJECT_TYPE_U64
;
353 ptr
->u
.u64
= v
.u
.u64
;
354 ptr
->ptr
= &ptr
->u
.u64
;
355 dbg_printf("context get index dynamic u64 %" PRIi64
"\n", ptr
->u
.u64
);
357 case LTTNG_UST_DYNAMIC_TYPE_S8
:
358 case LTTNG_UST_DYNAMIC_TYPE_S16
:
359 case LTTNG_UST_DYNAMIC_TYPE_S32
:
360 case LTTNG_UST_DYNAMIC_TYPE_S64
:
361 ptr
->object_type
= OBJECT_TYPE_S64
;
362 ptr
->u
.s64
= v
.u
.s64
;
363 ptr
->ptr
= &ptr
->u
.s64
;
364 dbg_printf("context get index dynamic s64 %" PRIi64
"\n", ptr
->u
.s64
);
366 case LTTNG_UST_DYNAMIC_TYPE_FLOAT
:
367 case LTTNG_UST_DYNAMIC_TYPE_DOUBLE
:
368 ptr
->object_type
= OBJECT_TYPE_DOUBLE
;
370 ptr
->ptr
= &ptr
->u
.d
;
371 dbg_printf("context get index dynamic double %g\n", ptr
->u
.d
);
373 case LTTNG_UST_DYNAMIC_TYPE_STRING
:
374 ptr
->object_type
= OBJECT_TYPE_STRING
;
376 dbg_printf("context get index dynamic string %s\n", (const char *) ptr
->ptr
);
379 dbg_printf("Interpreter warning: unknown dynamic type (%d).\n", (int) v
.sel
);
384 ERR("Structure type cannot be loaded.");
387 ERR("Unknown type: %d", (int) field
->type
.atype
);
393 static int dynamic_get_index(struct lttng_ctx
*ctx
,
394 struct bytecode_runtime
*runtime
,
395 uint64_t index
, struct estack_entry
*stack_top
)
398 const struct bytecode_get_index_data
*gid
;
400 gid
= (const struct bytecode_get_index_data
*) &runtime
->data
[index
];
401 switch (stack_top
->u
.ptr
.type
) {
403 switch (stack_top
->u
.ptr
.object_type
) {
404 case OBJECT_TYPE_ARRAY
:
408 assert(gid
->offset
< gid
->array_len
);
409 /* Skip count (unsigned long) */
410 ptr
= *(const char **) (stack_top
->u
.ptr
.ptr
+ sizeof(unsigned long));
411 ptr
= ptr
+ gid
->offset
;
412 stack_top
->u
.ptr
.ptr
= ptr
;
413 stack_top
->u
.ptr
.object_type
= gid
->elem
.type
;
414 stack_top
->u
.ptr
.rev_bo
= gid
->elem
.rev_bo
;
415 assert(stack_top
->u
.ptr
.field
->type
.atype
== atype_array
||
416 stack_top
->u
.ptr
.field
->type
.atype
== atype_array_nestable
);
417 stack_top
->u
.ptr
.field
= NULL
;
420 case OBJECT_TYPE_SEQUENCE
:
425 ptr
= *(const char **) (stack_top
->u
.ptr
.ptr
+ sizeof(unsigned long));
426 ptr_seq_len
= *(unsigned long *) stack_top
->u
.ptr
.ptr
;
427 if (gid
->offset
>= gid
->elem
.len
* ptr_seq_len
) {
431 ptr
= ptr
+ gid
->offset
;
432 stack_top
->u
.ptr
.ptr
= ptr
;
433 stack_top
->u
.ptr
.object_type
= gid
->elem
.type
;
434 stack_top
->u
.ptr
.rev_bo
= gid
->elem
.rev_bo
;
435 assert(stack_top
->u
.ptr
.field
->type
.atype
== atype_sequence
||
436 stack_top
->u
.ptr
.field
->type
.atype
== atype_sequence_nestable
);
437 stack_top
->u
.ptr
.field
= NULL
;
440 case OBJECT_TYPE_STRUCT
:
441 ERR("Nested structures are not supported yet.");
444 case OBJECT_TYPE_VARIANT
:
446 ERR("Unexpected get index type %d",
447 (int) stack_top
->u
.ptr
.object_type
);
452 case LOAD_ROOT_CONTEXT
:
453 case LOAD_ROOT_APP_CONTEXT
: /* Fall-through */
455 ret
= context_get_index(ctx
,
463 case LOAD_ROOT_PAYLOAD
:
464 stack_top
->u
.ptr
.ptr
+= gid
->offset
;
465 if (gid
->elem
.type
== OBJECT_TYPE_STRING
)
466 stack_top
->u
.ptr
.ptr
= *(const char * const *) stack_top
->u
.ptr
.ptr
;
467 stack_top
->u
.ptr
.object_type
= gid
->elem
.type
;
468 stack_top
->u
.ptr
.type
= LOAD_OBJECT
;
469 stack_top
->u
.ptr
.field
= gid
->field
;
470 stack_top
->u
.ptr
.rev_bo
= gid
->elem
.rev_bo
;
474 stack_top
->type
= REG_PTR
;
482 static int dynamic_load_field(struct estack_entry
*stack_top
)
486 switch (stack_top
->u
.ptr
.type
) {
489 case LOAD_ROOT_CONTEXT
:
490 case LOAD_ROOT_APP_CONTEXT
:
491 case LOAD_ROOT_PAYLOAD
:
493 dbg_printf("Interpreter warning: cannot load root, missing field name.\n");
497 switch (stack_top
->u
.ptr
.object_type
) {
499 dbg_printf("op load field s8\n");
500 stack_top
->u
.v
= *(int8_t *) stack_top
->u
.ptr
.ptr
;
501 stack_top
->type
= REG_S64
;
503 case OBJECT_TYPE_S16
:
507 dbg_printf("op load field s16\n");
508 tmp
= *(int16_t *) stack_top
->u
.ptr
.ptr
;
509 if (stack_top
->u
.ptr
.rev_bo
)
511 stack_top
->u
.v
= tmp
;
512 stack_top
->type
= REG_S64
;
515 case OBJECT_TYPE_S32
:
519 dbg_printf("op load field s32\n");
520 tmp
= *(int32_t *) stack_top
->u
.ptr
.ptr
;
521 if (stack_top
->u
.ptr
.rev_bo
)
523 stack_top
->u
.v
= tmp
;
524 stack_top
->type
= REG_S64
;
527 case OBJECT_TYPE_S64
:
531 dbg_printf("op load field s64\n");
532 tmp
= *(int64_t *) stack_top
->u
.ptr
.ptr
;
533 if (stack_top
->u
.ptr
.rev_bo
)
535 stack_top
->u
.v
= tmp
;
536 stack_top
->type
= REG_S64
;
539 case OBJECT_TYPE_SIGNED_ENUM
:
543 dbg_printf("op load field signed enumeration\n");
544 tmp
= *(int64_t *) stack_top
->u
.ptr
.ptr
;
545 if (stack_top
->u
.ptr
.rev_bo
)
547 stack_top
->u
.v
= tmp
;
548 stack_top
->type
= REG_S64
;
552 dbg_printf("op load field u8\n");
553 stack_top
->u
.v
= *(uint8_t *) stack_top
->u
.ptr
.ptr
;
554 stack_top
->type
= REG_U64
;
556 case OBJECT_TYPE_U16
:
560 dbg_printf("op load field u16\n");
561 tmp
= *(uint16_t *) stack_top
->u
.ptr
.ptr
;
562 if (stack_top
->u
.ptr
.rev_bo
)
564 stack_top
->u
.v
= tmp
;
565 stack_top
->type
= REG_U64
;
568 case OBJECT_TYPE_U32
:
572 dbg_printf("op load field u32\n");
573 tmp
= *(uint32_t *) stack_top
->u
.ptr
.ptr
;
574 if (stack_top
->u
.ptr
.rev_bo
)
576 stack_top
->u
.v
= tmp
;
577 stack_top
->type
= REG_U64
;
580 case OBJECT_TYPE_U64
:
584 dbg_printf("op load field u64\n");
585 tmp
= *(uint64_t *) stack_top
->u
.ptr
.ptr
;
586 if (stack_top
->u
.ptr
.rev_bo
)
588 stack_top
->u
.v
= tmp
;
589 stack_top
->type
= REG_U64
;
592 case OBJECT_TYPE_UNSIGNED_ENUM
:
596 dbg_printf("op load field unsigned enumeration\n");
597 tmp
= *(uint64_t *) stack_top
->u
.ptr
.ptr
;
598 if (stack_top
->u
.ptr
.rev_bo
)
600 stack_top
->u
.v
= tmp
;
601 stack_top
->type
= REG_U64
;
604 case OBJECT_TYPE_DOUBLE
:
605 memcpy(&stack_top
->u
.d
,
606 stack_top
->u
.ptr
.ptr
,
607 sizeof(struct literal_double
));
608 stack_top
->type
= REG_DOUBLE
;
610 case OBJECT_TYPE_STRING
:
614 dbg_printf("op load field string\n");
615 str
= (const char *) stack_top
->u
.ptr
.ptr
;
616 stack_top
->u
.s
.str
= str
;
617 if (unlikely(!stack_top
->u
.s
.str
)) {
618 dbg_printf("Interpreter warning: loading a NULL string.\n");
622 stack_top
->u
.s
.seq_len
= SIZE_MAX
;
623 stack_top
->u
.s
.literal_type
=
624 ESTACK_STRING_LITERAL_TYPE_NONE
;
625 stack_top
->type
= REG_STRING
;
628 case OBJECT_TYPE_STRING_SEQUENCE
:
632 dbg_printf("op load field string sequence\n");
633 ptr
= stack_top
->u
.ptr
.ptr
;
634 stack_top
->u
.s
.seq_len
= *(unsigned long *) ptr
;
635 stack_top
->u
.s
.str
= *(const char **) (ptr
+ sizeof(unsigned long));
636 stack_top
->type
= REG_STRING
;
637 if (unlikely(!stack_top
->u
.s
.str
)) {
638 dbg_printf("Interpreter warning: loading a NULL sequence.\n");
642 stack_top
->u
.s
.literal_type
=
643 ESTACK_STRING_LITERAL_TYPE_NONE
;
646 case OBJECT_TYPE_DYNAMIC
:
648 * Dynamic types in context are looked up
649 * by context get index.
653 case OBJECT_TYPE_SEQUENCE
:
654 case OBJECT_TYPE_ARRAY
:
655 case OBJECT_TYPE_STRUCT
:
656 case OBJECT_TYPE_VARIANT
:
657 ERR("Sequences, arrays, struct and variant cannot be loaded (nested types).");
668 int lttng_bytecode_interpret_format_output(struct estack_entry
*ax
,
669 struct lttng_interpreter_output
*output
)
676 output
->type
= LTTNG_INTERPRETER_TYPE_S64
;
677 output
->u
.s
= ax
->u
.v
;
680 output
->type
= LTTNG_INTERPRETER_TYPE_U64
;
681 output
->u
.u
= (uint64_t) ax
->u
.v
;
684 output
->type
= LTTNG_INTERPRETER_TYPE_DOUBLE
;
685 output
->u
.d
= ax
->u
.d
;
688 output
->type
= LTTNG_INTERPRETER_TYPE_STRING
;
689 output
->u
.str
.str
= ax
->u
.s
.str
;
690 output
->u
.str
.len
= ax
->u
.s
.seq_len
;
693 switch (ax
->u
.ptr
.object_type
) {
695 case OBJECT_TYPE_S16
:
696 case OBJECT_TYPE_S32
:
697 case OBJECT_TYPE_S64
:
699 case OBJECT_TYPE_U16
:
700 case OBJECT_TYPE_U32
:
701 case OBJECT_TYPE_U64
:
702 case OBJECT_TYPE_DOUBLE
:
703 case OBJECT_TYPE_STRING
:
704 case OBJECT_TYPE_STRING_SEQUENCE
:
705 ret
= dynamic_load_field(ax
);
708 /* Retry after loading ptr into stack top. */
710 case OBJECT_TYPE_SEQUENCE
:
711 output
->type
= LTTNG_INTERPRETER_TYPE_SEQUENCE
;
712 output
->u
.sequence
.ptr
= *(const char **) (ax
->u
.ptr
.ptr
+ sizeof(unsigned long));
713 output
->u
.sequence
.nr_elem
= *(unsigned long *) ax
->u
.ptr
.ptr
;
714 output
->u
.sequence
.nested_type
= ax
->u
.ptr
.field
->type
.u
.sequence_nestable
.elem_type
;
716 case OBJECT_TYPE_ARRAY
:
717 /* Skip count (unsigned long) */
718 output
->type
= LTTNG_INTERPRETER_TYPE_SEQUENCE
;
719 output
->u
.sequence
.ptr
= *(const char **) (ax
->u
.ptr
.ptr
+ sizeof(unsigned long));
720 output
->u
.sequence
.nr_elem
= ax
->u
.ptr
.field
->type
.u
.array_nestable
.length
;
721 output
->u
.sequence
.nested_type
= ax
->u
.ptr
.field
->type
.u
.array_nestable
.elem_type
;
723 case OBJECT_TYPE_SIGNED_ENUM
:
724 ret
= dynamic_load_field(ax
);
727 output
->type
= LTTNG_INTERPRETER_TYPE_SIGNED_ENUM
;
728 output
->u
.s
= ax
->u
.v
;
730 case OBJECT_TYPE_UNSIGNED_ENUM
:
731 ret
= dynamic_load_field(ax
);
734 output
->type
= LTTNG_INTERPRETER_TYPE_UNSIGNED_ENUM
;
735 output
->u
.u
= ax
->u
.v
;
737 case OBJECT_TYPE_STRUCT
:
738 case OBJECT_TYPE_VARIANT
:
744 case REG_STAR_GLOB_STRING
:
750 return LTTNG_INTERPRETER_RECORD_FLAG
;
754 * Return 0 (discard), or raise the 0x1 flag (log event).
755 * Currently, other flags are kept for future extensions and have no
759 uint64_t bytecode_interpret(void *interpreter_data
,
760 const char *interpreter_stack_data
,
761 struct lttng_interpreter_output
*output
)
763 struct bytecode_runtime
*bytecode
= interpreter_data
;
764 struct lttng_ctx
*ctx
= rcu_dereference(*bytecode
->p
.pctx
);
765 void *pc
, *next_pc
, *start_pc
;
768 struct estack _stack
;
769 struct estack
*stack
= &_stack
;
770 register int64_t ax
= 0, bx
= 0;
771 register enum entry_type ax_t
= REG_UNKNOWN
, bx_t
= REG_UNKNOWN
;
772 register int top
= INTERPRETER_STACK_EMPTY
;
773 #ifndef INTERPRETER_USE_SWITCH
774 static void *dispatch
[NR_BYTECODE_OPS
] = {
775 [ BYTECODE_OP_UNKNOWN
] = &&LABEL_BYTECODE_OP_UNKNOWN
,
777 [ BYTECODE_OP_RETURN
] = &&LABEL_BYTECODE_OP_RETURN
,
780 [ BYTECODE_OP_MUL
] = &&LABEL_BYTECODE_OP_MUL
,
781 [ BYTECODE_OP_DIV
] = &&LABEL_BYTECODE_OP_DIV
,
782 [ BYTECODE_OP_MOD
] = &&LABEL_BYTECODE_OP_MOD
,
783 [ BYTECODE_OP_PLUS
] = &&LABEL_BYTECODE_OP_PLUS
,
784 [ BYTECODE_OP_MINUS
] = &&LABEL_BYTECODE_OP_MINUS
,
785 [ BYTECODE_OP_BIT_RSHIFT
] = &&LABEL_BYTECODE_OP_BIT_RSHIFT
,
786 [ BYTECODE_OP_BIT_LSHIFT
] = &&LABEL_BYTECODE_OP_BIT_LSHIFT
,
787 [ BYTECODE_OP_BIT_AND
] = &&LABEL_BYTECODE_OP_BIT_AND
,
788 [ BYTECODE_OP_BIT_OR
] = &&LABEL_BYTECODE_OP_BIT_OR
,
789 [ BYTECODE_OP_BIT_XOR
] = &&LABEL_BYTECODE_OP_BIT_XOR
,
791 /* binary comparators */
792 [ BYTECODE_OP_EQ
] = &&LABEL_BYTECODE_OP_EQ
,
793 [ BYTECODE_OP_NE
] = &&LABEL_BYTECODE_OP_NE
,
794 [ BYTECODE_OP_GT
] = &&LABEL_BYTECODE_OP_GT
,
795 [ BYTECODE_OP_LT
] = &&LABEL_BYTECODE_OP_LT
,
796 [ BYTECODE_OP_GE
] = &&LABEL_BYTECODE_OP_GE
,
797 [ BYTECODE_OP_LE
] = &&LABEL_BYTECODE_OP_LE
,
799 /* string binary comparator */
800 [ BYTECODE_OP_EQ_STRING
] = &&LABEL_BYTECODE_OP_EQ_STRING
,
801 [ BYTECODE_OP_NE_STRING
] = &&LABEL_BYTECODE_OP_NE_STRING
,
802 [ BYTECODE_OP_GT_STRING
] = &&LABEL_BYTECODE_OP_GT_STRING
,
803 [ BYTECODE_OP_LT_STRING
] = &&LABEL_BYTECODE_OP_LT_STRING
,
804 [ BYTECODE_OP_GE_STRING
] = &&LABEL_BYTECODE_OP_GE_STRING
,
805 [ BYTECODE_OP_LE_STRING
] = &&LABEL_BYTECODE_OP_LE_STRING
,
807 /* globbing pattern binary comparator */
808 [ BYTECODE_OP_EQ_STAR_GLOB_STRING
] = &&LABEL_BYTECODE_OP_EQ_STAR_GLOB_STRING
,
809 [ BYTECODE_OP_NE_STAR_GLOB_STRING
] = &&LABEL_BYTECODE_OP_NE_STAR_GLOB_STRING
,
811 /* s64 binary comparator */
812 [ BYTECODE_OP_EQ_S64
] = &&LABEL_BYTECODE_OP_EQ_S64
,
813 [ BYTECODE_OP_NE_S64
] = &&LABEL_BYTECODE_OP_NE_S64
,
814 [ BYTECODE_OP_GT_S64
] = &&LABEL_BYTECODE_OP_GT_S64
,
815 [ BYTECODE_OP_LT_S64
] = &&LABEL_BYTECODE_OP_LT_S64
,
816 [ BYTECODE_OP_GE_S64
] = &&LABEL_BYTECODE_OP_GE_S64
,
817 [ BYTECODE_OP_LE_S64
] = &&LABEL_BYTECODE_OP_LE_S64
,
819 /* double binary comparator */
820 [ BYTECODE_OP_EQ_DOUBLE
] = &&LABEL_BYTECODE_OP_EQ_DOUBLE
,
821 [ BYTECODE_OP_NE_DOUBLE
] = &&LABEL_BYTECODE_OP_NE_DOUBLE
,
822 [ BYTECODE_OP_GT_DOUBLE
] = &&LABEL_BYTECODE_OP_GT_DOUBLE
,
823 [ BYTECODE_OP_LT_DOUBLE
] = &&LABEL_BYTECODE_OP_LT_DOUBLE
,
824 [ BYTECODE_OP_GE_DOUBLE
] = &&LABEL_BYTECODE_OP_GE_DOUBLE
,
825 [ BYTECODE_OP_LE_DOUBLE
] = &&LABEL_BYTECODE_OP_LE_DOUBLE
,
827 /* Mixed S64-double binary comparators */
828 [ BYTECODE_OP_EQ_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_EQ_DOUBLE_S64
,
829 [ BYTECODE_OP_NE_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_NE_DOUBLE_S64
,
830 [ BYTECODE_OP_GT_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_GT_DOUBLE_S64
,
831 [ BYTECODE_OP_LT_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_LT_DOUBLE_S64
,
832 [ BYTECODE_OP_GE_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_GE_DOUBLE_S64
,
833 [ BYTECODE_OP_LE_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_LE_DOUBLE_S64
,
835 [ BYTECODE_OP_EQ_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_EQ_S64_DOUBLE
,
836 [ BYTECODE_OP_NE_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_NE_S64_DOUBLE
,
837 [ BYTECODE_OP_GT_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_GT_S64_DOUBLE
,
838 [ BYTECODE_OP_LT_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_LT_S64_DOUBLE
,
839 [ BYTECODE_OP_GE_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_GE_S64_DOUBLE
,
840 [ BYTECODE_OP_LE_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_LE_S64_DOUBLE
,
843 [ BYTECODE_OP_UNARY_PLUS
] = &&LABEL_BYTECODE_OP_UNARY_PLUS
,
844 [ BYTECODE_OP_UNARY_MINUS
] = &&LABEL_BYTECODE_OP_UNARY_MINUS
,
845 [ BYTECODE_OP_UNARY_NOT
] = &&LABEL_BYTECODE_OP_UNARY_NOT
,
846 [ BYTECODE_OP_UNARY_PLUS_S64
] = &&LABEL_BYTECODE_OP_UNARY_PLUS_S64
,
847 [ BYTECODE_OP_UNARY_MINUS_S64
] = &&LABEL_BYTECODE_OP_UNARY_MINUS_S64
,
848 [ BYTECODE_OP_UNARY_NOT_S64
] = &&LABEL_BYTECODE_OP_UNARY_NOT_S64
,
849 [ BYTECODE_OP_UNARY_PLUS_DOUBLE
] = &&LABEL_BYTECODE_OP_UNARY_PLUS_DOUBLE
,
850 [ BYTECODE_OP_UNARY_MINUS_DOUBLE
] = &&LABEL_BYTECODE_OP_UNARY_MINUS_DOUBLE
,
851 [ BYTECODE_OP_UNARY_NOT_DOUBLE
] = &&LABEL_BYTECODE_OP_UNARY_NOT_DOUBLE
,
854 [ BYTECODE_OP_AND
] = &&LABEL_BYTECODE_OP_AND
,
855 [ BYTECODE_OP_OR
] = &&LABEL_BYTECODE_OP_OR
,
858 [ BYTECODE_OP_LOAD_FIELD_REF
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_REF
,
859 [ BYTECODE_OP_LOAD_FIELD_REF_STRING
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_REF_STRING
,
860 [ BYTECODE_OP_LOAD_FIELD_REF_SEQUENCE
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_REF_SEQUENCE
,
861 [ BYTECODE_OP_LOAD_FIELD_REF_S64
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_REF_S64
,
862 [ BYTECODE_OP_LOAD_FIELD_REF_DOUBLE
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_REF_DOUBLE
,
864 /* load from immediate operand */
865 [ BYTECODE_OP_LOAD_STRING
] = &&LABEL_BYTECODE_OP_LOAD_STRING
,
866 [ BYTECODE_OP_LOAD_STAR_GLOB_STRING
] = &&LABEL_BYTECODE_OP_LOAD_STAR_GLOB_STRING
,
867 [ BYTECODE_OP_LOAD_S64
] = &&LABEL_BYTECODE_OP_LOAD_S64
,
868 [ BYTECODE_OP_LOAD_DOUBLE
] = &&LABEL_BYTECODE_OP_LOAD_DOUBLE
,
871 [ BYTECODE_OP_CAST_TO_S64
] = &&LABEL_BYTECODE_OP_CAST_TO_S64
,
872 [ BYTECODE_OP_CAST_DOUBLE_TO_S64
] = &&LABEL_BYTECODE_OP_CAST_DOUBLE_TO_S64
,
873 [ BYTECODE_OP_CAST_NOP
] = &&LABEL_BYTECODE_OP_CAST_NOP
,
875 /* get context ref */
876 [ BYTECODE_OP_GET_CONTEXT_REF
] = &&LABEL_BYTECODE_OP_GET_CONTEXT_REF
,
877 [ BYTECODE_OP_GET_CONTEXT_REF_STRING
] = &&LABEL_BYTECODE_OP_GET_CONTEXT_REF_STRING
,
878 [ BYTECODE_OP_GET_CONTEXT_REF_S64
] = &&LABEL_BYTECODE_OP_GET_CONTEXT_REF_S64
,
879 [ BYTECODE_OP_GET_CONTEXT_REF_DOUBLE
] = &&LABEL_BYTECODE_OP_GET_CONTEXT_REF_DOUBLE
,
881 /* Instructions for recursive traversal through composed types. */
882 [ BYTECODE_OP_GET_CONTEXT_ROOT
] = &&LABEL_BYTECODE_OP_GET_CONTEXT_ROOT
,
883 [ BYTECODE_OP_GET_APP_CONTEXT_ROOT
] = &&LABEL_BYTECODE_OP_GET_APP_CONTEXT_ROOT
,
884 [ BYTECODE_OP_GET_PAYLOAD_ROOT
] = &&LABEL_BYTECODE_OP_GET_PAYLOAD_ROOT
,
886 [ BYTECODE_OP_GET_SYMBOL
] = &&LABEL_BYTECODE_OP_GET_SYMBOL
,
887 [ BYTECODE_OP_GET_SYMBOL_FIELD
] = &&LABEL_BYTECODE_OP_GET_SYMBOL_FIELD
,
888 [ BYTECODE_OP_GET_INDEX_U16
] = &&LABEL_BYTECODE_OP_GET_INDEX_U16
,
889 [ BYTECODE_OP_GET_INDEX_U64
] = &&LABEL_BYTECODE_OP_GET_INDEX_U64
,
891 [ BYTECODE_OP_LOAD_FIELD
] = &&LABEL_BYTECODE_OP_LOAD_FIELD
,
892 [ BYTECODE_OP_LOAD_FIELD_S8
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_S8
,
893 [ BYTECODE_OP_LOAD_FIELD_S16
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_S16
,
894 [ BYTECODE_OP_LOAD_FIELD_S32
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_S32
,
895 [ BYTECODE_OP_LOAD_FIELD_S64
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_S64
,
896 [ BYTECODE_OP_LOAD_FIELD_U8
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_U8
,
897 [ BYTECODE_OP_LOAD_FIELD_U16
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_U16
,
898 [ BYTECODE_OP_LOAD_FIELD_U32
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_U32
,
899 [ BYTECODE_OP_LOAD_FIELD_U64
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_U64
,
900 [ BYTECODE_OP_LOAD_FIELD_STRING
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_STRING
,
901 [ BYTECODE_OP_LOAD_FIELD_SEQUENCE
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_SEQUENCE
,
902 [ BYTECODE_OP_LOAD_FIELD_DOUBLE
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_DOUBLE
,
904 [ BYTECODE_OP_UNARY_BIT_NOT
] = &&LABEL_BYTECODE_OP_UNARY_BIT_NOT
,
906 [ BYTECODE_OP_RETURN_S64
] = &&LABEL_BYTECODE_OP_RETURN_S64
,
908 #endif /* #ifndef INTERPRETER_USE_SWITCH */
912 OP(BYTECODE_OP_UNKNOWN
):
913 OP(BYTECODE_OP_LOAD_FIELD_REF
):
914 #ifdef INTERPRETER_USE_SWITCH
916 #endif /* INTERPRETER_USE_SWITCH */
917 ERR("unknown bytecode op %u",
918 (unsigned int) *(bytecode_opcode_t
*) pc
);
922 OP(BYTECODE_OP_RETURN
):
923 /* LTTNG_INTERPRETER_DISCARD or LTTNG_INTERPRETER_RECORD_FLAG */
924 /* Handle dynamic typing. */
925 switch (estack_ax_t
) {
928 retval
= !!estack_ax_v
;
939 case REG_STAR_GLOB_STRING
:
948 OP(BYTECODE_OP_RETURN_S64
):
949 /* LTTNG_INTERPRETER_DISCARD or LTTNG_INTERPRETER_RECORD_FLAG */
950 retval
= !!estack_ax_v
;
958 OP(BYTECODE_OP_PLUS
):
959 OP(BYTECODE_OP_MINUS
):
960 ERR("unsupported bytecode op %u",
961 (unsigned int) *(bytecode_opcode_t
*) pc
);
967 /* Dynamic typing. */
968 switch (estack_ax_t
) {
969 case REG_S64
: /* Fall-through */
971 switch (estack_bx_t
) {
972 case REG_S64
: /* Fall-through */
974 JUMP_TO(BYTECODE_OP_EQ_S64
);
976 JUMP_TO(BYTECODE_OP_EQ_DOUBLE_S64
);
977 case REG_STRING
: /* Fall-through */
978 case REG_STAR_GLOB_STRING
:
982 ERR("Unknown interpreter register type (%d)",
989 switch (estack_bx_t
) {
990 case REG_S64
: /* Fall-through */
992 JUMP_TO(BYTECODE_OP_EQ_S64_DOUBLE
);
994 JUMP_TO(BYTECODE_OP_EQ_DOUBLE
);
995 case REG_STRING
: /* Fall-through */
996 case REG_STAR_GLOB_STRING
:
1000 ERR("Unknown interpreter register type (%d)",
1007 switch (estack_bx_t
) {
1008 case REG_S64
: /* Fall-through */
1009 case REG_U64
: /* Fall-through */
1014 JUMP_TO(BYTECODE_OP_EQ_STRING
);
1015 case REG_STAR_GLOB_STRING
:
1016 JUMP_TO(BYTECODE_OP_EQ_STAR_GLOB_STRING
);
1018 ERR("Unknown interpreter register type (%d)",
1024 case REG_STAR_GLOB_STRING
:
1025 switch (estack_bx_t
) {
1026 case REG_S64
: /* Fall-through */
1027 case REG_U64
: /* Fall-through */
1032 JUMP_TO(BYTECODE_OP_EQ_STAR_GLOB_STRING
);
1033 case REG_STAR_GLOB_STRING
:
1037 ERR("Unknown interpreter register type (%d)",
1044 ERR("Unknown interpreter register type (%d)",
1052 /* Dynamic typing. */
1053 switch (estack_ax_t
) {
1054 case REG_S64
: /* Fall-through */
1056 switch (estack_bx_t
) {
1057 case REG_S64
: /* Fall-through */
1059 JUMP_TO(BYTECODE_OP_NE_S64
);
1061 JUMP_TO(BYTECODE_OP_NE_DOUBLE_S64
);
1062 case REG_STRING
: /* Fall-through */
1063 case REG_STAR_GLOB_STRING
:
1067 ERR("Unknown interpreter register type (%d)",
1074 switch (estack_bx_t
) {
1075 case REG_S64
: /* Fall-through */
1077 JUMP_TO(BYTECODE_OP_NE_S64_DOUBLE
);
1079 JUMP_TO(BYTECODE_OP_NE_DOUBLE
);
1080 case REG_STRING
: /* Fall-through */
1081 case REG_STAR_GLOB_STRING
:
1085 ERR("Unknown interpreter register type (%d)",
1092 switch (estack_bx_t
) {
1093 case REG_S64
: /* Fall-through */
1099 JUMP_TO(BYTECODE_OP_NE_STRING
);
1100 case REG_STAR_GLOB_STRING
:
1101 JUMP_TO(BYTECODE_OP_NE_STAR_GLOB_STRING
);
1103 ERR("Unknown interpreter register type (%d)",
1109 case REG_STAR_GLOB_STRING
:
1110 switch (estack_bx_t
) {
1111 case REG_S64
: /* Fall-through */
1117 JUMP_TO(BYTECODE_OP_NE_STAR_GLOB_STRING
);
1118 case REG_STAR_GLOB_STRING
:
1122 ERR("Unknown interpreter register type (%d)",
1129 ERR("Unknown interpreter register type (%d)",
1137 /* Dynamic typing. */
1138 switch (estack_ax_t
) {
1139 case REG_S64
: /* Fall-through */
1141 switch (estack_bx_t
) {
1142 case REG_S64
: /* Fall-through */
1144 JUMP_TO(BYTECODE_OP_GT_S64
);
1146 JUMP_TO(BYTECODE_OP_GT_DOUBLE_S64
);
1147 case REG_STRING
: /* Fall-through */
1148 case REG_STAR_GLOB_STRING
:
1152 ERR("Unknown interpreter register type (%d)",
1159 switch (estack_bx_t
) {
1160 case REG_S64
: /* Fall-through */
1162 JUMP_TO(BYTECODE_OP_GT_S64_DOUBLE
);
1164 JUMP_TO(BYTECODE_OP_GT_DOUBLE
);
1165 case REG_STRING
: /* Fall-through */
1166 case REG_STAR_GLOB_STRING
:
1170 ERR("Unknown interpreter register type (%d)",
1177 switch (estack_bx_t
) {
1178 case REG_S64
: /* Fall-through */
1179 case REG_U64
: /* Fall-through */
1180 case REG_DOUBLE
: /* Fall-through */
1181 case REG_STAR_GLOB_STRING
:
1185 JUMP_TO(BYTECODE_OP_GT_STRING
);
1187 ERR("Unknown interpreter register type (%d)",
1194 ERR("Unknown interpreter register type (%d)",
1202 /* Dynamic typing. */
1203 switch (estack_ax_t
) {
1204 case REG_S64
: /* Fall-through */
1206 switch (estack_bx_t
) {
1207 case REG_S64
: /* Fall-through */
1209 JUMP_TO(BYTECODE_OP_LT_S64
);
1211 JUMP_TO(BYTECODE_OP_LT_DOUBLE_S64
);
1212 case REG_STRING
: /* Fall-through */
1213 case REG_STAR_GLOB_STRING
:
1217 ERR("Unknown interpreter register type (%d)",
1224 switch (estack_bx_t
) {
1225 case REG_S64
: /* Fall-through */
1227 JUMP_TO(BYTECODE_OP_LT_S64_DOUBLE
);
1229 JUMP_TO(BYTECODE_OP_LT_DOUBLE
);
1230 case REG_STRING
: /* Fall-through */
1231 case REG_STAR_GLOB_STRING
:
1235 ERR("Unknown interpreter register type (%d)",
1242 switch (estack_bx_t
) {
1243 case REG_S64
: /* Fall-through */
1244 case REG_U64
: /* Fall-through */
1245 case REG_DOUBLE
: /* Fall-through */
1246 case REG_STAR_GLOB_STRING
:
1250 JUMP_TO(BYTECODE_OP_LT_STRING
);
1252 ERR("Unknown interpreter register type (%d)",
1259 ERR("Unknown interpreter register type (%d)",
1267 /* Dynamic typing. */
1268 switch (estack_ax_t
) {
1269 case REG_S64
: /* Fall-through */
1271 switch (estack_bx_t
) {
1272 case REG_S64
: /* Fall-through */
1274 JUMP_TO(BYTECODE_OP_GE_S64
);
1276 JUMP_TO(BYTECODE_OP_GE_DOUBLE_S64
);
1277 case REG_STRING
: /* Fall-through */
1278 case REG_STAR_GLOB_STRING
:
1282 ERR("Unknown interpreter register type (%d)",
1289 switch (estack_bx_t
) {
1290 case REG_S64
: /* Fall-through */
1292 JUMP_TO(BYTECODE_OP_GE_S64_DOUBLE
);
1294 JUMP_TO(BYTECODE_OP_GE_DOUBLE
);
1295 case REG_STRING
: /* Fall-through */
1296 case REG_STAR_GLOB_STRING
:
1300 ERR("Unknown interpreter register type (%d)",
1307 switch (estack_bx_t
) {
1308 case REG_S64
: /* Fall-through */
1309 case REG_U64
: /* Fall-through */
1310 case REG_DOUBLE
: /* Fall-through */
1311 case REG_STAR_GLOB_STRING
:
1315 JUMP_TO(BYTECODE_OP_GE_STRING
);
1317 ERR("Unknown interpreter register type (%d)",
1324 ERR("Unknown interpreter register type (%d)",
1332 /* Dynamic typing. */
1333 switch (estack_ax_t
) {
1334 case REG_S64
: /* Fall-through */
1336 switch (estack_bx_t
) {
1337 case REG_S64
: /* Fall-through */
1339 JUMP_TO(BYTECODE_OP_LE_S64
);
1341 JUMP_TO(BYTECODE_OP_LE_DOUBLE_S64
);
1342 case REG_STRING
: /* Fall-through */
1343 case REG_STAR_GLOB_STRING
:
1347 ERR("Unknown interpreter register type (%d)",
1354 switch (estack_bx_t
) {
1355 case REG_S64
: /* Fall-through */
1357 JUMP_TO(BYTECODE_OP_LE_S64_DOUBLE
);
1359 JUMP_TO(BYTECODE_OP_LE_DOUBLE
);
1360 case REG_STRING
: /* Fall-through */
1361 case REG_STAR_GLOB_STRING
:
1365 ERR("Unknown interpreter register type (%d)",
1372 switch (estack_bx_t
) {
1373 case REG_S64
: /* Fall-through */
1374 case REG_U64
: /* Fall-through */
1375 case REG_DOUBLE
: /* Fall-through */
1376 case REG_STAR_GLOB_STRING
:
1380 JUMP_TO(BYTECODE_OP_LE_STRING
);
1382 ERR("Unknown interpreter register type (%d)",
1389 ERR("Unknown interpreter register type (%d)",
1396 OP(BYTECODE_OP_EQ_STRING
):
1400 res
= (stack_strcmp(stack
, top
, "==") == 0);
1401 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1403 estack_ax_t
= REG_S64
;
1404 next_pc
+= sizeof(struct binary_op
);
1407 OP(BYTECODE_OP_NE_STRING
):
1411 res
= (stack_strcmp(stack
, top
, "!=") != 0);
1412 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1414 estack_ax_t
= REG_S64
;
1415 next_pc
+= sizeof(struct binary_op
);
1418 OP(BYTECODE_OP_GT_STRING
):
1422 res
= (stack_strcmp(stack
, top
, ">") > 0);
1423 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1425 estack_ax_t
= REG_S64
;
1426 next_pc
+= sizeof(struct binary_op
);
1429 OP(BYTECODE_OP_LT_STRING
):
1433 res
= (stack_strcmp(stack
, top
, "<") < 0);
1434 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1436 estack_ax_t
= REG_S64
;
1437 next_pc
+= sizeof(struct binary_op
);
1440 OP(BYTECODE_OP_GE_STRING
):
1444 res
= (stack_strcmp(stack
, top
, ">=") >= 0);
1445 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1447 estack_ax_t
= REG_S64
;
1448 next_pc
+= sizeof(struct binary_op
);
1451 OP(BYTECODE_OP_LE_STRING
):
1455 res
= (stack_strcmp(stack
, top
, "<=") <= 0);
1456 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1458 estack_ax_t
= REG_S64
;
1459 next_pc
+= sizeof(struct binary_op
);
1463 OP(BYTECODE_OP_EQ_STAR_GLOB_STRING
):
1467 res
= (stack_star_glob_match(stack
, top
, "==") == 0);
1468 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1470 estack_ax_t
= REG_S64
;
1471 next_pc
+= sizeof(struct binary_op
);
1474 OP(BYTECODE_OP_NE_STAR_GLOB_STRING
):
1478 res
= (stack_star_glob_match(stack
, top
, "!=") != 0);
1479 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1481 estack_ax_t
= REG_S64
;
1482 next_pc
+= sizeof(struct binary_op
);
1486 OP(BYTECODE_OP_EQ_S64
):
1490 res
= (estack_bx_v
== estack_ax_v
);
1491 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1493 estack_ax_t
= REG_S64
;
1494 next_pc
+= sizeof(struct binary_op
);
1497 OP(BYTECODE_OP_NE_S64
):
1501 res
= (estack_bx_v
!= estack_ax_v
);
1502 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1504 estack_ax_t
= REG_S64
;
1505 next_pc
+= sizeof(struct binary_op
);
1508 OP(BYTECODE_OP_GT_S64
):
1512 res
= (estack_bx_v
> estack_ax_v
);
1513 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1515 estack_ax_t
= REG_S64
;
1516 next_pc
+= sizeof(struct binary_op
);
1519 OP(BYTECODE_OP_LT_S64
):
1523 res
= (estack_bx_v
< estack_ax_v
);
1524 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1526 estack_ax_t
= REG_S64
;
1527 next_pc
+= sizeof(struct binary_op
);
1530 OP(BYTECODE_OP_GE_S64
):
1534 res
= (estack_bx_v
>= estack_ax_v
);
1535 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1537 estack_ax_t
= REG_S64
;
1538 next_pc
+= sizeof(struct binary_op
);
1541 OP(BYTECODE_OP_LE_S64
):
1545 res
= (estack_bx_v
<= estack_ax_v
);
1546 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1548 estack_ax_t
= REG_S64
;
1549 next_pc
+= sizeof(struct binary_op
);
1553 OP(BYTECODE_OP_EQ_DOUBLE
):
1557 res
= (estack_bx(stack
, top
)->u
.d
== estack_ax(stack
, top
)->u
.d
);
1558 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1560 estack_ax_t
= REG_S64
;
1561 next_pc
+= sizeof(struct binary_op
);
1564 OP(BYTECODE_OP_NE_DOUBLE
):
1568 res
= (estack_bx(stack
, top
)->u
.d
!= estack_ax(stack
, top
)->u
.d
);
1569 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1571 estack_ax_t
= REG_S64
;
1572 next_pc
+= sizeof(struct binary_op
);
1575 OP(BYTECODE_OP_GT_DOUBLE
):
1579 res
= (estack_bx(stack
, top
)->u
.d
> estack_ax(stack
, top
)->u
.d
);
1580 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1582 estack_ax_t
= REG_S64
;
1583 next_pc
+= sizeof(struct binary_op
);
1586 OP(BYTECODE_OP_LT_DOUBLE
):
1590 res
= (estack_bx(stack
, top
)->u
.d
< estack_ax(stack
, top
)->u
.d
);
1591 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1593 estack_ax_t
= REG_S64
;
1594 next_pc
+= sizeof(struct binary_op
);
1597 OP(BYTECODE_OP_GE_DOUBLE
):
1601 res
= (estack_bx(stack
, top
)->u
.d
>= estack_ax(stack
, top
)->u
.d
);
1602 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1604 estack_ax_t
= REG_S64
;
1605 next_pc
+= sizeof(struct binary_op
);
1608 OP(BYTECODE_OP_LE_DOUBLE
):
1612 res
= (estack_bx(stack
, top
)->u
.d
<= estack_ax(stack
, top
)->u
.d
);
1613 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1615 estack_ax_t
= REG_S64
;
1616 next_pc
+= sizeof(struct binary_op
);
1620 /* Mixed S64-double binary comparators */
1621 OP(BYTECODE_OP_EQ_DOUBLE_S64
):
1625 res
= (estack_bx(stack
, top
)->u
.d
== estack_ax_v
);
1626 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1628 estack_ax_t
= REG_S64
;
1629 next_pc
+= sizeof(struct binary_op
);
1632 OP(BYTECODE_OP_NE_DOUBLE_S64
):
1636 res
= (estack_bx(stack
, top
)->u
.d
!= estack_ax_v
);
1637 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1639 estack_ax_t
= REG_S64
;
1640 next_pc
+= sizeof(struct binary_op
);
1643 OP(BYTECODE_OP_GT_DOUBLE_S64
):
1647 res
= (estack_bx(stack
, top
)->u
.d
> estack_ax_v
);
1648 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1650 estack_ax_t
= REG_S64
;
1651 next_pc
+= sizeof(struct binary_op
);
1654 OP(BYTECODE_OP_LT_DOUBLE_S64
):
1658 res
= (estack_bx(stack
, top
)->u
.d
< estack_ax_v
);
1659 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1661 estack_ax_t
= REG_S64
;
1662 next_pc
+= sizeof(struct binary_op
);
1665 OP(BYTECODE_OP_GE_DOUBLE_S64
):
1669 res
= (estack_bx(stack
, top
)->u
.d
>= estack_ax_v
);
1670 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1672 estack_ax_t
= REG_S64
;
1673 next_pc
+= sizeof(struct binary_op
);
1676 OP(BYTECODE_OP_LE_DOUBLE_S64
):
1680 res
= (estack_bx(stack
, top
)->u
.d
<= estack_ax_v
);
1681 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1683 estack_ax_t
= REG_S64
;
1684 next_pc
+= sizeof(struct binary_op
);
1688 OP(BYTECODE_OP_EQ_S64_DOUBLE
):
1692 res
= (estack_bx_v
== estack_ax(stack
, top
)->u
.d
);
1693 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1695 estack_ax_t
= REG_S64
;
1696 next_pc
+= sizeof(struct binary_op
);
1699 OP(BYTECODE_OP_NE_S64_DOUBLE
):
1703 res
= (estack_bx_v
!= estack_ax(stack
, top
)->u
.d
);
1704 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1706 estack_ax_t
= REG_S64
;
1707 next_pc
+= sizeof(struct binary_op
);
1710 OP(BYTECODE_OP_GT_S64_DOUBLE
):
1714 res
= (estack_bx_v
> estack_ax(stack
, top
)->u
.d
);
1715 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1717 estack_ax_t
= REG_S64
;
1718 next_pc
+= sizeof(struct binary_op
);
1721 OP(BYTECODE_OP_LT_S64_DOUBLE
):
1725 res
= (estack_bx_v
< estack_ax(stack
, top
)->u
.d
);
1726 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1728 estack_ax_t
= REG_S64
;
1729 next_pc
+= sizeof(struct binary_op
);
1732 OP(BYTECODE_OP_GE_S64_DOUBLE
):
1736 res
= (estack_bx_v
>= estack_ax(stack
, top
)->u
.d
);
1737 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1739 estack_ax_t
= REG_S64
;
1740 next_pc
+= sizeof(struct binary_op
);
1743 OP(BYTECODE_OP_LE_S64_DOUBLE
):
1747 res
= (estack_bx_v
<= estack_ax(stack
, top
)->u
.d
);
1748 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1750 estack_ax_t
= REG_S64
;
1751 next_pc
+= sizeof(struct binary_op
);
1754 OP(BYTECODE_OP_BIT_RSHIFT
):
1758 if (!IS_INTEGER_REGISTER(estack_ax_t
) || !IS_INTEGER_REGISTER(estack_bx_t
)) {
1763 /* Catch undefined behavior. */
1764 if (caa_unlikely(estack_ax_v
< 0 || estack_ax_v
>= 64)) {
1768 res
= ((uint64_t) estack_bx_v
>> (uint32_t) estack_ax_v
);
1769 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1771 estack_ax_t
= REG_U64
;
1772 next_pc
+= sizeof(struct binary_op
);
1775 OP(BYTECODE_OP_BIT_LSHIFT
):
1779 if (!IS_INTEGER_REGISTER(estack_ax_t
) || !IS_INTEGER_REGISTER(estack_bx_t
)) {
1784 /* Catch undefined behavior. */
1785 if (caa_unlikely(estack_ax_v
< 0 || estack_ax_v
>= 64)) {
1789 res
= ((uint64_t) estack_bx_v
<< (uint32_t) estack_ax_v
);
1790 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1792 estack_ax_t
= REG_U64
;
1793 next_pc
+= sizeof(struct binary_op
);
1796 OP(BYTECODE_OP_BIT_AND
):
1800 if (!IS_INTEGER_REGISTER(estack_ax_t
) || !IS_INTEGER_REGISTER(estack_bx_t
)) {
1805 res
= ((uint64_t) estack_bx_v
& (uint64_t) estack_ax_v
);
1806 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1808 estack_ax_t
= REG_U64
;
1809 next_pc
+= sizeof(struct binary_op
);
1812 OP(BYTECODE_OP_BIT_OR
):
1816 if (!IS_INTEGER_REGISTER(estack_ax_t
) || !IS_INTEGER_REGISTER(estack_bx_t
)) {
1821 res
= ((uint64_t) estack_bx_v
| (uint64_t) estack_ax_v
);
1822 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1824 estack_ax_t
= REG_U64
;
1825 next_pc
+= sizeof(struct binary_op
);
1828 OP(BYTECODE_OP_BIT_XOR
):
1832 if (!IS_INTEGER_REGISTER(estack_ax_t
) || !IS_INTEGER_REGISTER(estack_bx_t
)) {
1837 res
= ((uint64_t) estack_bx_v
^ (uint64_t) estack_ax_v
);
1838 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1840 estack_ax_t
= REG_U64
;
1841 next_pc
+= sizeof(struct binary_op
);
1846 OP(BYTECODE_OP_UNARY_PLUS
):
1848 /* Dynamic typing. */
1849 switch (estack_ax_t
) {
1850 case REG_S64
: /* Fall-through. */
1852 JUMP_TO(BYTECODE_OP_UNARY_PLUS_S64
);
1854 JUMP_TO(BYTECODE_OP_UNARY_PLUS_DOUBLE
);
1855 case REG_STRING
: /* Fall-through */
1856 case REG_STAR_GLOB_STRING
:
1860 ERR("Unknown interpreter register type (%d)",
1866 OP(BYTECODE_OP_UNARY_MINUS
):
1868 /* Dynamic typing. */
1869 switch (estack_ax_t
) {
1870 case REG_S64
: /* Fall-through. */
1872 JUMP_TO(BYTECODE_OP_UNARY_MINUS_S64
);
1874 JUMP_TO(BYTECODE_OP_UNARY_MINUS_DOUBLE
);
1875 case REG_STRING
: /* Fall-through */
1876 case REG_STAR_GLOB_STRING
:
1880 ERR("Unknown interpreter register type (%d)",
1886 OP(BYTECODE_OP_UNARY_NOT
):
1888 /* Dynamic typing. */
1889 switch (estack_ax_t
) {
1890 case REG_S64
: /* Fall-through. */
1892 JUMP_TO(BYTECODE_OP_UNARY_NOT_S64
);
1894 JUMP_TO(BYTECODE_OP_UNARY_NOT_DOUBLE
);
1895 case REG_STRING
: /* Fall-through */
1896 case REG_STAR_GLOB_STRING
:
1900 ERR("Unknown interpreter register type (%d)",
1905 next_pc
+= sizeof(struct unary_op
);
1909 OP(BYTECODE_OP_UNARY_BIT_NOT
):
1911 /* Dynamic typing. */
1912 if (!IS_INTEGER_REGISTER(estack_ax_t
)) {
1917 estack_ax_v
= ~(uint64_t) estack_ax_v
;
1918 estack_ax_t
= REG_U64
;
1919 next_pc
+= sizeof(struct unary_op
);
1923 OP(BYTECODE_OP_UNARY_PLUS_S64
):
1924 OP(BYTECODE_OP_UNARY_PLUS_DOUBLE
):
1926 next_pc
+= sizeof(struct unary_op
);
1929 OP(BYTECODE_OP_UNARY_MINUS_S64
):
1931 estack_ax_v
= -estack_ax_v
;
1932 next_pc
+= sizeof(struct unary_op
);
1935 OP(BYTECODE_OP_UNARY_MINUS_DOUBLE
):
1937 estack_ax(stack
, top
)->u
.d
= -estack_ax(stack
, top
)->u
.d
;
1938 next_pc
+= sizeof(struct unary_op
);
1941 OP(BYTECODE_OP_UNARY_NOT_S64
):
1943 estack_ax_v
= !estack_ax_v
;
1944 estack_ax_t
= REG_S64
;
1945 next_pc
+= sizeof(struct unary_op
);
1948 OP(BYTECODE_OP_UNARY_NOT_DOUBLE
):
1950 estack_ax_v
= !estack_ax(stack
, top
)->u
.d
;
1951 estack_ax_t
= REG_S64
;
1952 next_pc
+= sizeof(struct unary_op
);
1957 OP(BYTECODE_OP_AND
):
1959 struct logical_op
*insn
= (struct logical_op
*) pc
;
1961 if (estack_ax_t
!= REG_S64
&& estack_ax_t
!= REG_U64
) {
1965 /* If AX is 0, skip and evaluate to 0 */
1966 if (unlikely(estack_ax_v
== 0)) {
1967 dbg_printf("Jumping to bytecode offset %u\n",
1968 (unsigned int) insn
->skip_offset
);
1969 next_pc
= start_pc
+ insn
->skip_offset
;
1971 /* Pop 1 when jump not taken */
1972 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1973 next_pc
+= sizeof(struct logical_op
);
1979 struct logical_op
*insn
= (struct logical_op
*) pc
;
1981 if (estack_ax_t
!= REG_S64
&& estack_ax_t
!= REG_U64
) {
1985 /* If AX is nonzero, skip and evaluate to 1 */
1986 if (unlikely(estack_ax_v
!= 0)) {
1988 dbg_printf("Jumping to bytecode offset %u\n",
1989 (unsigned int) insn
->skip_offset
);
1990 next_pc
= start_pc
+ insn
->skip_offset
;
1992 /* Pop 1 when jump not taken */
1993 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1994 next_pc
+= sizeof(struct logical_op
);
2000 /* load field ref */
2001 OP(BYTECODE_OP_LOAD_FIELD_REF_STRING
):
2003 struct load_op
*insn
= (struct load_op
*) pc
;
2004 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2006 dbg_printf("load field ref offset %u type string\n",
2008 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2009 estack_ax(stack
, top
)->u
.s
.str
=
2010 *(const char * const *) &interpreter_stack_data
[ref
->offset
];
2011 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
2012 dbg_printf("Interpreter warning: loading a NULL string.\n");
2016 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
2017 estack_ax(stack
, top
)->u
.s
.literal_type
=
2018 ESTACK_STRING_LITERAL_TYPE_NONE
;
2019 estack_ax_t
= REG_STRING
;
2020 dbg_printf("ref load string %s\n", estack_ax(stack
, top
)->u
.s
.str
);
2021 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2025 OP(BYTECODE_OP_LOAD_FIELD_REF_SEQUENCE
):
2027 struct load_op
*insn
= (struct load_op
*) pc
;
2028 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2030 dbg_printf("load field ref offset %u type sequence\n",
2032 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2033 estack_ax(stack
, top
)->u
.s
.seq_len
=
2034 *(unsigned long *) &interpreter_stack_data
[ref
->offset
];
2035 estack_ax(stack
, top
)->u
.s
.str
=
2036 *(const char **) (&interpreter_stack_data
[ref
->offset
2037 + sizeof(unsigned long)]);
2038 estack_ax_t
= REG_STRING
;
2039 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
2040 dbg_printf("Interpreter warning: loading a NULL sequence.\n");
2044 estack_ax(stack
, top
)->u
.s
.literal_type
=
2045 ESTACK_STRING_LITERAL_TYPE_NONE
;
2046 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2050 OP(BYTECODE_OP_LOAD_FIELD_REF_S64
):
2052 struct load_op
*insn
= (struct load_op
*) pc
;
2053 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2055 dbg_printf("load field ref offset %u type s64\n",
2057 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2059 ((struct literal_numeric
*) &interpreter_stack_data
[ref
->offset
])->v
;
2060 estack_ax_t
= REG_S64
;
2061 dbg_printf("ref load s64 %" PRIi64
"\n", estack_ax_v
);
2062 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2066 OP(BYTECODE_OP_LOAD_FIELD_REF_DOUBLE
):
2068 struct load_op
*insn
= (struct load_op
*) pc
;
2069 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2071 dbg_printf("load field ref offset %u type double\n",
2073 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2074 memcpy(&estack_ax(stack
, top
)->u
.d
, &interpreter_stack_data
[ref
->offset
],
2075 sizeof(struct literal_double
));
2076 estack_ax_t
= REG_DOUBLE
;
2077 dbg_printf("ref load double %g\n", estack_ax(stack
, top
)->u
.d
);
2078 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2082 /* load from immediate operand */
2083 OP(BYTECODE_OP_LOAD_STRING
):
2085 struct load_op
*insn
= (struct load_op
*) pc
;
2087 dbg_printf("load string %s\n", insn
->data
);
2088 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2089 estack_ax(stack
, top
)->u
.s
.str
= insn
->data
;
2090 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
2091 estack_ax(stack
, top
)->u
.s
.literal_type
=
2092 ESTACK_STRING_LITERAL_TYPE_PLAIN
;
2093 estack_ax_t
= REG_STRING
;
2094 next_pc
+= sizeof(struct load_op
) + strlen(insn
->data
) + 1;
2098 OP(BYTECODE_OP_LOAD_STAR_GLOB_STRING
):
2100 struct load_op
*insn
= (struct load_op
*) pc
;
2102 dbg_printf("load globbing pattern %s\n", insn
->data
);
2103 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2104 estack_ax(stack
, top
)->u
.s
.str
= insn
->data
;
2105 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
2106 estack_ax(stack
, top
)->u
.s
.literal_type
=
2107 ESTACK_STRING_LITERAL_TYPE_STAR_GLOB
;
2108 estack_ax_t
= REG_STAR_GLOB_STRING
;
2109 next_pc
+= sizeof(struct load_op
) + strlen(insn
->data
) + 1;
2113 OP(BYTECODE_OP_LOAD_S64
):
2115 struct load_op
*insn
= (struct load_op
*) pc
;
2117 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2118 estack_ax_v
= ((struct literal_numeric
*) insn
->data
)->v
;
2119 estack_ax_t
= REG_S64
;
2120 dbg_printf("load s64 %" PRIi64
"\n", estack_ax_v
);
2121 next_pc
+= sizeof(struct load_op
)
2122 + sizeof(struct literal_numeric
);
2126 OP(BYTECODE_OP_LOAD_DOUBLE
):
2128 struct load_op
*insn
= (struct load_op
*) pc
;
2130 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2131 memcpy(&estack_ax(stack
, top
)->u
.d
, insn
->data
,
2132 sizeof(struct literal_double
));
2133 estack_ax_t
= REG_DOUBLE
;
2134 dbg_printf("load double %g\n", estack_ax(stack
, top
)->u
.d
);
2135 next_pc
+= sizeof(struct load_op
)
2136 + sizeof(struct literal_double
);
2141 OP(BYTECODE_OP_CAST_TO_S64
):
2143 /* Dynamic typing. */
2144 switch (estack_ax_t
) {
2146 JUMP_TO(BYTECODE_OP_CAST_NOP
);
2148 JUMP_TO(BYTECODE_OP_CAST_DOUBLE_TO_S64
);
2150 estack_ax_t
= REG_S64
;
2151 next_pc
+= sizeof(struct cast_op
);
2152 case REG_STRING
: /* Fall-through */
2153 case REG_STAR_GLOB_STRING
:
2157 ERR("Unknown interpreter register type (%d)",
2164 OP(BYTECODE_OP_CAST_DOUBLE_TO_S64
):
2166 estack_ax_v
= (int64_t) estack_ax(stack
, top
)->u
.d
;
2167 estack_ax_t
= REG_S64
;
2168 next_pc
+= sizeof(struct cast_op
);
2172 OP(BYTECODE_OP_CAST_NOP
):
2174 next_pc
+= sizeof(struct cast_op
);
2178 /* get context ref */
2179 OP(BYTECODE_OP_GET_CONTEXT_REF
):
2181 struct load_op
*insn
= (struct load_op
*) pc
;
2182 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2183 struct lttng_ctx_field
*ctx_field
;
2184 struct lttng_ctx_value v
;
2186 dbg_printf("get context ref offset %u type dynamic\n",
2188 ctx_field
= &ctx
->fields
[ref
->offset
];
2189 ctx_field
->get_value(ctx_field
, &v
);
2190 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2192 case LTTNG_UST_DYNAMIC_TYPE_NONE
:
2195 case LTTNG_UST_DYNAMIC_TYPE_S64
:
2196 estack_ax_v
= v
.u
.s64
;
2197 estack_ax_t
= REG_S64
;
2198 dbg_printf("ref get context dynamic s64 %" PRIi64
"\n", estack_ax_v
);
2200 case LTTNG_UST_DYNAMIC_TYPE_DOUBLE
:
2201 estack_ax(stack
, top
)->u
.d
= v
.u
.d
;
2202 estack_ax_t
= REG_DOUBLE
;
2203 dbg_printf("ref get context dynamic double %g\n", estack_ax(stack
, top
)->u
.d
);
2205 case LTTNG_UST_DYNAMIC_TYPE_STRING
:
2206 estack_ax(stack
, top
)->u
.s
.str
= v
.u
.str
;
2207 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
2208 dbg_printf("Interpreter warning: loading a NULL string.\n");
2212 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
2213 estack_ax(stack
, top
)->u
.s
.literal_type
=
2214 ESTACK_STRING_LITERAL_TYPE_NONE
;
2215 dbg_printf("ref get context dynamic string %s\n", estack_ax(stack
, top
)->u
.s
.str
);
2216 estack_ax_t
= REG_STRING
;
2219 dbg_printf("Interpreter warning: unknown dynamic type (%d).\n", (int) v
.sel
);
2223 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2227 OP(BYTECODE_OP_GET_CONTEXT_REF_STRING
):
2229 struct load_op
*insn
= (struct load_op
*) pc
;
2230 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2231 struct lttng_ctx_field
*ctx_field
;
2232 struct lttng_ctx_value v
;
2234 dbg_printf("get context ref offset %u type string\n",
2236 ctx_field
= &ctx
->fields
[ref
->offset
];
2237 ctx_field
->get_value(ctx_field
, &v
);
2238 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2239 estack_ax(stack
, top
)->u
.s
.str
= v
.u
.str
;
2240 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
2241 dbg_printf("Interpreter warning: loading a NULL string.\n");
2245 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
2246 estack_ax(stack
, top
)->u
.s
.literal_type
=
2247 ESTACK_STRING_LITERAL_TYPE_NONE
;
2248 estack_ax_t
= REG_STRING
;
2249 dbg_printf("ref get context string %s\n", estack_ax(stack
, top
)->u
.s
.str
);
2250 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2254 OP(BYTECODE_OP_GET_CONTEXT_REF_S64
):
2256 struct load_op
*insn
= (struct load_op
*) pc
;
2257 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2258 struct lttng_ctx_field
*ctx_field
;
2259 struct lttng_ctx_value v
;
2261 dbg_printf("get context ref offset %u type s64\n",
2263 ctx_field
= &ctx
->fields
[ref
->offset
];
2264 ctx_field
->get_value(ctx_field
, &v
);
2265 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2266 estack_ax_v
= v
.u
.s64
;
2267 estack_ax_t
= REG_S64
;
2268 dbg_printf("ref get context s64 %" PRIi64
"\n", estack_ax_v
);
2269 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2273 OP(BYTECODE_OP_GET_CONTEXT_REF_DOUBLE
):
2275 struct load_op
*insn
= (struct load_op
*) pc
;
2276 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2277 struct lttng_ctx_field
*ctx_field
;
2278 struct lttng_ctx_value v
;
2280 dbg_printf("get context ref offset %u type double\n",
2282 ctx_field
= &ctx
->fields
[ref
->offset
];
2283 ctx_field
->get_value(ctx_field
, &v
);
2284 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2285 memcpy(&estack_ax(stack
, top
)->u
.d
, &v
.u
.d
, sizeof(struct literal_double
));
2286 estack_ax_t
= REG_DOUBLE
;
2287 dbg_printf("ref get context double %g\n", estack_ax(stack
, top
)->u
.d
);
2288 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2292 OP(BYTECODE_OP_GET_CONTEXT_ROOT
):
2294 dbg_printf("op get context root\n");
2295 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2296 estack_ax(stack
, top
)->u
.ptr
.type
= LOAD_ROOT_CONTEXT
;
2297 /* "field" only needed for variants. */
2298 estack_ax(stack
, top
)->u
.ptr
.field
= NULL
;
2299 estack_ax_t
= REG_PTR
;
2300 next_pc
+= sizeof(struct load_op
);
2304 OP(BYTECODE_OP_GET_APP_CONTEXT_ROOT
):
2306 dbg_printf("op get app context root\n");
2307 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2308 estack_ax(stack
, top
)->u
.ptr
.type
= LOAD_ROOT_APP_CONTEXT
;
2309 /* "field" only needed for variants. */
2310 estack_ax(stack
, top
)->u
.ptr
.field
= NULL
;
2311 estack_ax_t
= REG_PTR
;
2312 next_pc
+= sizeof(struct load_op
);
2316 OP(BYTECODE_OP_GET_PAYLOAD_ROOT
):
2318 dbg_printf("op get app payload root\n");
2319 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2320 estack_ax(stack
, top
)->u
.ptr
.type
= LOAD_ROOT_PAYLOAD
;
2321 estack_ax(stack
, top
)->u
.ptr
.ptr
= interpreter_stack_data
;
2322 /* "field" only needed for variants. */
2323 estack_ax(stack
, top
)->u
.ptr
.field
= NULL
;
2324 estack_ax_t
= REG_PTR
;
2325 next_pc
+= sizeof(struct load_op
);
2329 OP(BYTECODE_OP_GET_SYMBOL
):
2331 dbg_printf("op get symbol\n");
2332 switch (estack_ax(stack
, top
)->u
.ptr
.type
) {
2334 ERR("Nested fields not implemented yet.");
2337 case LOAD_ROOT_CONTEXT
:
2338 case LOAD_ROOT_APP_CONTEXT
:
2339 case LOAD_ROOT_PAYLOAD
:
2341 * symbol lookup is performed by
2347 next_pc
+= sizeof(struct load_op
) + sizeof(struct get_symbol
);
2351 OP(BYTECODE_OP_GET_SYMBOL_FIELD
):
2354 * Used for first variant encountered in a
2355 * traversal. Variants are not implemented yet.
2361 OP(BYTECODE_OP_GET_INDEX_U16
):
2363 struct load_op
*insn
= (struct load_op
*) pc
;
2364 struct get_index_u16
*index
= (struct get_index_u16
*) insn
->data
;
2366 dbg_printf("op get index u16\n");
2367 ret
= dynamic_get_index(ctx
, bytecode
, index
->index
, estack_ax(stack
, top
));
2370 estack_ax_v
= estack_ax(stack
, top
)->u
.v
;
2371 estack_ax_t
= estack_ax(stack
, top
)->type
;
2372 next_pc
+= sizeof(struct load_op
) + sizeof(struct get_index_u16
);
2376 OP(BYTECODE_OP_GET_INDEX_U64
):
2378 struct load_op
*insn
= (struct load_op
*) pc
;
2379 struct get_index_u64
*index
= (struct get_index_u64
*) insn
->data
;
2381 dbg_printf("op get index u64\n");
2382 ret
= dynamic_get_index(ctx
, bytecode
, index
->index
, estack_ax(stack
, top
));
2385 estack_ax_v
= estack_ax(stack
, top
)->u
.v
;
2386 estack_ax_t
= estack_ax(stack
, top
)->type
;
2387 next_pc
+= sizeof(struct load_op
) + sizeof(struct get_index_u64
);
2391 OP(BYTECODE_OP_LOAD_FIELD
):
2393 dbg_printf("op load field\n");
2394 ret
= dynamic_load_field(estack_ax(stack
, top
));
2397 estack_ax_v
= estack_ax(stack
, top
)->u
.v
;
2398 estack_ax_t
= estack_ax(stack
, top
)->type
;
2399 next_pc
+= sizeof(struct load_op
);
2403 OP(BYTECODE_OP_LOAD_FIELD_S8
):
2405 dbg_printf("op load field s8\n");
2407 estack_ax_v
= *(int8_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2408 estack_ax_t
= REG_S64
;
2409 next_pc
+= sizeof(struct load_op
);
2412 OP(BYTECODE_OP_LOAD_FIELD_S16
):
2414 dbg_printf("op load field s16\n");
2416 estack_ax_v
= *(int16_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2417 estack_ax_t
= REG_S64
;
2418 next_pc
+= sizeof(struct load_op
);
2421 OP(BYTECODE_OP_LOAD_FIELD_S32
):
2423 dbg_printf("op load field s32\n");
2425 estack_ax_v
= *(int32_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2426 estack_ax_t
= REG_S64
;
2427 next_pc
+= sizeof(struct load_op
);
2430 OP(BYTECODE_OP_LOAD_FIELD_S64
):
2432 dbg_printf("op load field s64\n");
2434 estack_ax_v
= *(int64_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2435 estack_ax_t
= REG_S64
;
2436 next_pc
+= sizeof(struct load_op
);
2439 OP(BYTECODE_OP_LOAD_FIELD_U8
):
2441 dbg_printf("op load field u8\n");
2443 estack_ax_v
= *(uint8_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2444 estack_ax_t
= REG_U64
;
2445 next_pc
+= sizeof(struct load_op
);
2448 OP(BYTECODE_OP_LOAD_FIELD_U16
):
2450 dbg_printf("op load field u16\n");
2452 estack_ax_v
= *(uint16_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2453 estack_ax_t
= REG_U64
;
2454 next_pc
+= sizeof(struct load_op
);
2457 OP(BYTECODE_OP_LOAD_FIELD_U32
):
2459 dbg_printf("op load field u32\n");
2461 estack_ax_v
= *(uint32_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2462 estack_ax_t
= REG_U64
;
2463 next_pc
+= sizeof(struct load_op
);
2466 OP(BYTECODE_OP_LOAD_FIELD_U64
):
2468 dbg_printf("op load field u64\n");
2470 estack_ax_v
= *(uint64_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2471 estack_ax_t
= REG_U64
;
2472 next_pc
+= sizeof(struct load_op
);
2475 OP(BYTECODE_OP_LOAD_FIELD_DOUBLE
):
2477 dbg_printf("op load field double\n");
2479 memcpy(&estack_ax(stack
, top
)->u
.d
,
2480 estack_ax(stack
, top
)->u
.ptr
.ptr
,
2481 sizeof(struct literal_double
));
2482 estack_ax(stack
, top
)->type
= REG_DOUBLE
;
2483 next_pc
+= sizeof(struct load_op
);
2487 OP(BYTECODE_OP_LOAD_FIELD_STRING
):
2491 dbg_printf("op load field string\n");
2492 str
= (const char *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2493 estack_ax(stack
, top
)->u
.s
.str
= str
;
2494 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
2495 dbg_printf("Interpreter warning: loading a NULL string.\n");
2499 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
2500 estack_ax(stack
, top
)->u
.s
.literal_type
=
2501 ESTACK_STRING_LITERAL_TYPE_NONE
;
2502 estack_ax(stack
, top
)->type
= REG_STRING
;
2503 next_pc
+= sizeof(struct load_op
);
2507 OP(BYTECODE_OP_LOAD_FIELD_SEQUENCE
):
2511 dbg_printf("op load field string sequence\n");
2512 ptr
= estack_ax(stack
, top
)->u
.ptr
.ptr
;
2513 estack_ax(stack
, top
)->u
.s
.seq_len
= *(unsigned long *) ptr
;
2514 estack_ax(stack
, top
)->u
.s
.str
= *(const char **) (ptr
+ sizeof(unsigned long));
2515 estack_ax(stack
, top
)->type
= REG_STRING
;
2516 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
2517 dbg_printf("Interpreter warning: loading a NULL sequence.\n");
2521 estack_ax(stack
, top
)->u
.s
.literal_type
=
2522 ESTACK_STRING_LITERAL_TYPE_NONE
;
2523 next_pc
+= sizeof(struct load_op
);
2529 /* Return _DISCARD on error. */
2531 return LTTNG_INTERPRETER_DISCARD
;
2534 return lttng_bytecode_interpret_format_output(estack_ax(stack
, top
),
2541 uint64_t lttng_bytecode_filter_interpret(void *filter_data
,
2542 const char *filter_stack_data
)
2544 return bytecode_interpret(filter_data
, filter_stack_data
, NULL
);