struct field_data *fdata = (struct field_data *) priv;
size_t orig_offset = offset;
int cpu = smp_processor_id();
+ struct irq_ibt_state irq_ibt_state;
/* do not write data if no space is available */
trace = stack_trace_context(fdata, cpu);
switch (fdata->mode) {
case CALLSTACK_KERNEL:
/* do the real work and reserve space */
+ irq_ibt_state = wrapper_irq_ibt_save();
trace->nr_entries = save_func_kernel(trace->entries,
MAX_ENTRIES, 0);
+ wrapper_irq_ibt_restore(irq_ibt_state);
break;
case CALLSTACK_USER:
++per_cpu(callstack_user_nesting, cpu);
/* do the real work and reserve space */
+ irq_ibt_state = wrapper_irq_ibt_save();
trace->nr_entries = save_func_user(trace->entries,
MAX_ENTRIES);
+ wrapper_irq_ibt_restore(irq_ibt_state);
per_cpu(callstack_user_nesting, cpu)--;
break;
default:
static
void lttng_callstack_length_record(void *priv, struct lttng_kernel_probe_ctx *probe_ctx,
struct lttng_kernel_ring_buffer_ctx *ctx,
- struct lttng_channel *chan)
+ struct lttng_kernel_channel_buffer *chan)
{
int cpu = ctx->priv.reserve_cpu;
struct field_data *fdata = (struct field_data *) priv;
struct lttng_stack_trace *trace = stack_trace_context(fdata, cpu);
unsigned int nr_seq_entries;
- lib_ring_buffer_align_ctx(ctx, lttng_alignof(unsigned int));
if (unlikely(!trace)) {
nr_seq_entries = 0;
} else {
if (trace->nr_entries == MAX_ENTRIES)
nr_seq_entries++;
}
- chan->ops->event_write(ctx, &nr_seq_entries, sizeof(unsigned int));
+ chan->ops->event_write(ctx, &nr_seq_entries, sizeof(unsigned int), lttng_alignof(unsigned int));
}
static
void lttng_callstack_sequence_record(void *priv, struct lttng_kernel_probe_ctx *probe_ctx,
struct lttng_kernel_ring_buffer_ctx *ctx,
- struct lttng_channel *chan)
+ struct lttng_kernel_channel_buffer *chan)
{
int cpu = ctx->priv.reserve_cpu;
struct field_data *fdata = (struct field_data *) priv;
struct lttng_stack_trace *trace = stack_trace_context(fdata, cpu);
unsigned int nr_seq_entries;
- lib_ring_buffer_align_ctx(ctx, lttng_alignof(unsigned long));
if (unlikely(!trace)) {
+ /* We need to align even if there are 0 elements. */
+ lib_ring_buffer_align_ctx(ctx, lttng_alignof(unsigned long));
return;
}
nr_seq_entries = trace->nr_entries;
if (trace->nr_entries == MAX_ENTRIES)
nr_seq_entries++;
chan->ops->event_write(ctx, trace->entries,
- sizeof(unsigned long) * trace->nr_entries);
+ sizeof(unsigned long) * trace->nr_entries, lttng_alignof(unsigned long));
/* Add our own ULONG_MAX delimiter to show incomplete stack. */
if (trace->nr_entries == MAX_ENTRIES) {
unsigned long delim = ULONG_MAX;
- chan->ops->event_write(ctx, &delim, sizeof(unsigned long));
+ chan->ops->event_write(ctx, &delim, sizeof(unsigned long), 1);
}
}