buf = shmp(handle, chan->backend.buf[ctx->cpu].shmp);
else
buf = shmp(handle, chan->backend.buf[0].shmp);
+ if (caa_unlikely(!buf))
+ return -EIO;
if (caa_unlikely(uatomic_read(&buf->record_disabled)))
return -EAGAIN;
ctx->buf = buf;
struct commit_counters_hot *cc_hot = shmp_index(handle,
buf->commit_hot, endidx);
+ if (caa_unlikely(!cc_hot))
+ return;
+
/*
* Must count record before incrementing the commit count.
*/
- subbuffer_count_record(config, &buf->backend, endidx, handle);
+ subbuffer_count_record(config, ctx, &buf->backend, endidx, handle);
/*
* Order all writes to buffer before the commit count update that will