*/
#define _LGPL_SOURCE
+#include <stddef.h>
+#include <stdint.h>
+
#include "lttng-filter.h"
#include <lttng/align.h>
+#include "ust-events-internal.h"
static int lttng_fls(int val)
{
size_t align, size_t len)
{
ssize_t ret;
- size_t padding = offset_align(runtime->data_len, align);
+ size_t padding = lttng_ust_offset_align(runtime->data_len, align);
size_t new_len = runtime->data_len + padding + len;
size_t new_alloc_len = new_len;
size_t old_alloc_len = runtime->data_alloc_len;
break;
case OBJECT_TYPE_U8:
dbg_printf("op load field u8\n");
- stack_top->type = REG_S64;
+ stack_top->type = REG_U64;
insn->op = FILTER_OP_LOAD_FIELD_U8;
break;
case OBJECT_TYPE_U16:
dbg_printf("op load field u16\n");
- stack_top->type = REG_S64;
+ stack_top->type = REG_U64;
if (!stack_top->load.rev_bo)
insn->op = FILTER_OP_LOAD_FIELD_U16;
break;
case OBJECT_TYPE_U32:
dbg_printf("op load field u32\n");
- stack_top->type = REG_S64;
+ stack_top->type = REG_U64;
if (!stack_top->load.rev_bo)
insn->op = FILTER_OP_LOAD_FIELD_U32;
break;
case OBJECT_TYPE_U64:
dbg_printf("op load field u64\n");
- stack_top->type = REG_S64;
+ stack_top->type = REG_U64;
if (!stack_top->load.rev_bo)
insn->op = FILTER_OP_LOAD_FIELD_U64;
break;
switch (stack_top->load.object_type) {
case OBJECT_TYPE_ARRAY:
{
+ const struct lttng_integer_type *integer_type;
const struct lttng_event_field *field;
uint32_t elem_len, num_elems;
int signedness;
field = stack_top->load.field;
- elem_len = field->type.u.array.elem_type.u.basic.integer.size;
- signedness = field->type.u.array.elem_type.u.basic.integer.signedness;
- num_elems = field->type.u.array.length;
+ switch (field->type.atype) {
+ case atype_array:
+ integer_type = &field->type.u.legacy.array.elem_type.u.basic.integer;
+ num_elems = field->type.u.legacy.array.length;
+ break;
+ case atype_array_nestable:
+ if (field->type.u.array_nestable.elem_type->atype != atype_integer) {
+ ret = -EINVAL;
+ goto end;
+ }
+ integer_type = &field->type.u.array_nestable.elem_type->u.integer;
+ num_elems = field->type.u.array_nestable.length;
+ break;
+ default:
+ ret = -EINVAL;
+ goto end;
+ }
+ elem_len = integer_type->size;
+ signedness = integer_type->signedness;
if (index >= num_elems) {
ret = -EINVAL;
goto end;
gid.array_len = num_elems * (elem_len / CHAR_BIT);
gid.elem.type = stack_top->load.object_type;
gid.elem.len = elem_len;
- if (field->type.u.array.elem_type.u.basic.integer.reverse_byte_order)
+ if (integer_type->reverse_byte_order)
gid.elem.rev_bo = true;
stack_top->load.rev_bo = gid.elem.rev_bo;
break;
}
case OBJECT_TYPE_SEQUENCE:
{
+ const struct lttng_integer_type *integer_type;
const struct lttng_event_field *field;
uint32_t elem_len;
int signedness;
field = stack_top->load.field;
- elem_len = field->type.u.sequence.elem_type.u.basic.integer.size;
- signedness = field->type.u.sequence.elem_type.u.basic.integer.signedness;
+ switch (field->type.atype) {
+ case atype_sequence:
+ integer_type = &field->type.u.legacy.sequence.elem_type.u.basic.integer;
+ break;
+ case atype_sequence_nestable:
+ if (field->type.u.sequence_nestable.elem_type->atype != atype_integer) {
+ ret = -EINVAL;
+ goto end;
+ }
+ integer_type = &field->type.u.sequence_nestable.elem_type->u.integer;
+ break;
+ default:
+ ret = -EINVAL;
+ goto end;
+ }
+ elem_len = integer_type->size;
+ signedness = integer_type->signedness;
ret = specialize_get_index_object_type(&stack_top->load.object_type,
signedness, elem_len);
if (ret)
gid.offset = index * (elem_len / CHAR_BIT);
gid.elem.type = stack_top->load.object_type;
gid.elem.len = elem_len;
- if (field->type.u.sequence.elem_type.u.basic.integer.reverse_byte_order)
+ if (integer_type->reverse_byte_order)
gid.elem.rev_bo = true;
stack_top->load.rev_bo = gid.elem.rev_bo;
break;
struct vstack_load *load, bool is_context)
{
load->type = LOAD_OBJECT;
- /*
- * LTTng-UST layout all integer fields as s64 on the stack for the filter.
- */
+
switch (field->type.atype) {
case atype_integer:
- if (field->type.u.basic.integer.signedness)
+ if (field->type.u.integer.signedness)
load->object_type = OBJECT_TYPE_S64;
else
load->object_type = OBJECT_TYPE_U64;
load->rev_bo = false;
break;
case atype_enum:
+ case atype_enum_nestable:
{
- const struct lttng_integer_type *itype =
- &field->type.u.basic.enumeration.container_type;
+ const struct lttng_integer_type *itype;
+ if (field->type.atype == atype_enum) {
+ itype = &field->type.u.legacy.basic.enumeration.container_type;
+ } else {
+ itype = &field->type.u.enum_nestable.container_type->u.integer;
+ }
if (itype->signedness)
load->object_type = OBJECT_TYPE_S64;
else
break;
}
case atype_array:
- if (field->type.u.array.elem_type.atype != atype_integer) {
+ if (field->type.u.legacy.array.elem_type.atype != atype_integer) {
+ ERR("Array nesting only supports integer types.");
+ return -EINVAL;
+ }
+ if (is_context) {
+ load->object_type = OBJECT_TYPE_STRING;
+ } else {
+ if (field->type.u.legacy.array.elem_type.u.basic.integer.encoding == lttng_encode_none) {
+ load->object_type = OBJECT_TYPE_ARRAY;
+ load->field = field;
+ } else {
+ load->object_type = OBJECT_TYPE_STRING_SEQUENCE;
+ }
+ }
+ break;
+ case atype_array_nestable:
+ if (field->type.u.array_nestable.elem_type->atype != atype_integer) {
ERR("Array nesting only supports integer types.");
return -EINVAL;
}
if (is_context) {
load->object_type = OBJECT_TYPE_STRING;
} else {
- if (field->type.u.array.elem_type.u.basic.integer.encoding == lttng_encode_none) {
+ if (field->type.u.array_nestable.elem_type->u.integer.encoding == lttng_encode_none) {
load->object_type = OBJECT_TYPE_ARRAY;
load->field = field;
} else {
}
break;
case atype_sequence:
- if (field->type.u.sequence.elem_type.atype != atype_integer) {
+ if (field->type.u.legacy.sequence.elem_type.atype != atype_integer) {
+ ERR("Sequence nesting only supports integer types.");
+ return -EINVAL;
+ }
+ if (is_context) {
+ load->object_type = OBJECT_TYPE_STRING;
+ } else {
+ if (field->type.u.legacy.sequence.elem_type.u.basic.integer.encoding == lttng_encode_none) {
+ load->object_type = OBJECT_TYPE_SEQUENCE;
+ load->field = field;
+ } else {
+ load->object_type = OBJECT_TYPE_STRING_SEQUENCE;
+ }
+ }
+ break;
+ case atype_sequence_nestable:
+ if (field->type.u.sequence_nestable.elem_type->atype != atype_integer) {
ERR("Sequence nesting only supports integer types.");
return -EINVAL;
}
if (is_context) {
load->object_type = OBJECT_TYPE_STRING;
} else {
- if (field->type.u.sequence.elem_type.u.basic.integer.encoding == lttng_encode_none) {
+ if (field->type.u.sequence_nestable.elem_type->u.integer.encoding == lttng_encode_none) {
load->object_type = OBJECT_TYPE_SEQUENCE;
load->field = field;
} else {
}
}
break;
+
case atype_string:
load->object_type = OBJECT_TYPE_STRING;
break;
break;
case atype_dynamic:
load->object_type = OBJECT_TYPE_DYNAMIC;
- return -EINVAL;
+ break;
case atype_struct:
ERR("Structure type cannot be loaded.");
return -EINVAL;
return 0;
}
-static int specialize_context_lookup(struct lttng_session *session,
+static int specialize_context_lookup(struct lttng_ctx *ctx,
struct bytecode_runtime *runtime,
struct load_op *insn,
struct vstack_load *load)
struct filter_get_index_data gid;
ssize_t data_offset;
- idx = specialize_context_lookup_name(session->ctx, runtime, insn);
+ idx = specialize_context_lookup_name(ctx, runtime, insn);
if (idx < 0) {
return -ENOENT;
}
- ctx_field = &session->ctx->fields[idx];
+ ctx_field = &ctx->fields[idx];
field = &ctx_field->event_field;
ret = specialize_load_object(field, load, true);
if (ret)
return 0;
}
-static int specialize_app_context_lookup(struct lttng_session *session,
+static int specialize_app_context_lookup(struct lttng_ctx **pctx,
struct bytecode_runtime *runtime,
struct load_op *insn,
struct vstack_load *load)
}
strcpy(name, "$app.");
strcat(name, orig_name);
- idx = lttng_get_context_index(session->ctx, name);
+ idx = lttng_get_context_index(*pctx, name);
if (idx < 0) {
assert(lttng_context_is_app(name));
ret = lttng_ust_add_app_context_to_ctx_rcu(name,
- &session->ctx);
+ pctx);
if (ret)
return ret;
- idx = lttng_get_context_index(session->ctx,
- name);
+ idx = lttng_get_context_index(*pctx, name);
if (idx < 0)
return -ENOENT;
}
- ctx_field = &session->ctx->fields[idx];
+ ctx_field = &(*pctx)->fields[idx];
field = &ctx_field->event_field;
ret = specialize_load_object(field, load, true);
if (ret)
return ret;
}
-static int specialize_event_payload_lookup(struct lttng_event *event,
+static int specialize_payload_lookup(const struct lttng_event_desc *event_desc,
struct bytecode_runtime *runtime,
struct load_op *insn,
struct vstack_load *load)
{
const char *name;
uint16_t offset;
- const struct lttng_event_desc *desc = event->desc;
unsigned int i, nr_fields;
bool found = false;
uint32_t field_offset = 0;
struct filter_get_index_data gid;
ssize_t data_offset;
- nr_fields = desc->nr_fields;
+ nr_fields = event_desc->nr_fields;
offset = ((struct get_symbol *) insn->data)->offset;
name = runtime->p.bc->bc.data + runtime->p.bc->bc.reloc_offset + offset;
for (i = 0; i < nr_fields; i++) {
- field = &desc->fields[i];
+ field = &event_desc->fields[i];
+ if (field->u.ext.nofilter) {
+ continue;
+ }
if (!strcmp(field->name, name)) {
found = true;
break;
switch (field->type.atype) {
case atype_integer:
case atype_enum:
+ case atype_enum_nestable:
field_offset += sizeof(int64_t);
break;
case atype_array:
+ case atype_array_nestable:
case atype_sequence:
+ case atype_sequence_nestable:
field_offset += sizeof(unsigned long);
field_offset += sizeof(void *);
break;
return ret;
}
-int lttng_filter_specialize_bytecode(struct lttng_event *event,
+int lttng_filter_specialize_bytecode(const struct lttng_event_desc *event_desc,
struct bytecode_runtime *bytecode)
{
void *pc, *next_pc, *start_pc;
int ret = -EINVAL;
struct vstack _stack;
struct vstack *stack = &_stack;
- struct lttng_session *session = bytecode->p.session;
+ struct lttng_ctx **pctx = bytecode->p.pctx;
vstack_init(stack);
goto end;
case FILTER_OP_RETURN:
- if (vstack_ax(stack)->type == REG_S64)
+ if (vstack_ax(stack)->type == REG_S64 ||
+ vstack_ax(stack)->type == REG_U64)
*(filter_opcode_t *) pc = FILTER_OP_RETURN_S64;
ret = 0;
goto end;
case FILTER_OP_RETURN_S64:
- if (vstack_ax(stack)->type != REG_S64) {
+ if (vstack_ax(stack)->type != REG_S64 &&
+ vstack_ax(stack)->type != REG_U64) {
ERR("Unexpected register type\n");
ret = -EINVAL;
goto end;
insn->op = FILTER_OP_EQ_STAR_GLOB_STRING;
break;
case REG_S64:
+ case REG_U64:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_EQ_S64;
else
insn->op = FILTER_OP_EQ_DOUBLE_S64;
case REG_DOUBLE:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_EQ_S64_DOUBLE;
else
insn->op = FILTER_OP_EQ_DOUBLE;
insn->op = FILTER_OP_NE_STAR_GLOB_STRING;
break;
case REG_S64:
+ case REG_U64:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_NE_S64;
else
insn->op = FILTER_OP_NE_DOUBLE_S64;
case REG_DOUBLE:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_NE_S64_DOUBLE;
else
insn->op = FILTER_OP_NE_DOUBLE;
insn->op = FILTER_OP_GT_STRING;
break;
case REG_S64:
+ case REG_U64:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_GT_S64;
else
insn->op = FILTER_OP_GT_DOUBLE_S64;
case REG_DOUBLE:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_GT_S64_DOUBLE;
else
insn->op = FILTER_OP_GT_DOUBLE;
insn->op = FILTER_OP_LT_STRING;
break;
case REG_S64:
+ case REG_U64:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_LT_S64;
else
insn->op = FILTER_OP_LT_DOUBLE_S64;
case REG_DOUBLE:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_LT_S64_DOUBLE;
else
insn->op = FILTER_OP_LT_DOUBLE;
insn->op = FILTER_OP_GE_STRING;
break;
case REG_S64:
+ case REG_U64:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_GE_S64;
else
insn->op = FILTER_OP_GE_DOUBLE_S64;
case REG_DOUBLE:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_GE_S64_DOUBLE;
else
insn->op = FILTER_OP_GE_DOUBLE;
ret = -EINVAL;
goto end;
}
- vstack_ax(stack)->type = REG_S64;
+ vstack_ax(stack)->type = REG_U64;
next_pc += sizeof(struct binary_op);
break;
}
insn->op = FILTER_OP_LE_STRING;
break;
case REG_S64:
+ case REG_U64:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_LE_S64;
else
insn->op = FILTER_OP_LE_DOUBLE_S64;
case REG_DOUBLE:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_LE_S64_DOUBLE;
else
insn->op = FILTER_OP_LE_DOUBLE;
case FILTER_OP_LT_S64_DOUBLE:
case FILTER_OP_GE_S64_DOUBLE:
case FILTER_OP_LE_S64_DOUBLE:
+ {
+ /* Pop 2, push 1 */
+ if (vstack_pop(stack)) {
+ ret = -EINVAL;
+ goto end;
+ }
+ vstack_ax(stack)->type = REG_S64;
+ next_pc += sizeof(struct binary_op);
+ break;
+ }
+
case FILTER_OP_BIT_RSHIFT:
case FILTER_OP_BIT_LSHIFT:
case FILTER_OP_BIT_AND:
goto end;
case REG_S64:
+ case REG_U64:
insn->op = FILTER_OP_UNARY_PLUS_S64;
break;
case REG_DOUBLE:
goto end;
case REG_S64:
+ case REG_U64:
insn->op = FILTER_OP_UNARY_MINUS_S64;
break;
case REG_DOUBLE:
goto end;
case REG_S64:
+ case REG_U64:
insn->op = FILTER_OP_UNARY_NOT_S64;
break;
case REG_DOUBLE:
insn->op = FILTER_OP_CAST_DOUBLE_TO_S64;
break;
case REG_UNKNOWN:
+ case REG_U64:
break;
}
/* Pop 1, push 1 */
case FILTER_OP_LOAD_FIELD_S16:
case FILTER_OP_LOAD_FIELD_S32:
case FILTER_OP_LOAD_FIELD_S64:
+ {
+ /* Pop 1, push 1 */
+ vstack_ax(stack)->type = REG_S64;
+ next_pc += sizeof(struct load_op);
+ break;
+ }
+
case FILTER_OP_LOAD_FIELD_U8:
case FILTER_OP_LOAD_FIELD_U16:
case FILTER_OP_LOAD_FIELD_U32:
case FILTER_OP_LOAD_FIELD_U64:
{
/* Pop 1, push 1 */
- vstack_ax(stack)->type = REG_S64;
+ vstack_ax(stack)->type = REG_U64;
next_pc += sizeof(struct load_op);
break;
}
goto end;
case LOAD_ROOT_CONTEXT:
/* Lookup context field. */
- ret = specialize_context_lookup(session,
+ ret = specialize_context_lookup(*pctx,
bytecode, insn,
&vstack_ax(stack)->load);
if (ret)
break;
case LOAD_ROOT_APP_CONTEXT:
/* Lookup app context field. */
- ret = specialize_app_context_lookup(session,
+ ret = specialize_app_context_lookup(pctx,
bytecode, insn,
&vstack_ax(stack)->load);
if (ret)
break;
case LOAD_ROOT_PAYLOAD:
/* Lookup event payload field. */
- ret = specialize_event_payload_lookup(event,
+ ret = specialize_payload_lookup(event_desc,
bytecode, insn,
&vstack_ax(stack)->load);
if (ret)