struct field_data *fdata = (struct field_data *) priv;
size_t orig_offset = offset;
int cpu = smp_processor_id();
struct field_data *fdata = (struct field_data *) priv;
size_t orig_offset = offset;
int cpu = smp_processor_id();
/* do not write data if no space is available */
trace = stack_trace_context(fdata, cpu);
/* do not write data if no space is available */
trace = stack_trace_context(fdata, cpu);
break;
case CALLSTACK_USER:
++per_cpu(callstack_user_nesting, cpu);
/* do the real work and reserve space */
break;
case CALLSTACK_USER:
++per_cpu(callstack_user_nesting, cpu);
/* do the real work and reserve space */
struct lttng_stack_trace *trace = stack_trace_context(fdata, cpu);
unsigned int nr_seq_entries;
struct lttng_stack_trace *trace = stack_trace_context(fdata, cpu);
unsigned int nr_seq_entries;
- chan->ops->event_write(ctx, &nr_seq_entries, sizeof(unsigned int));
+ chan->ops->event_write(ctx, &nr_seq_entries, sizeof(unsigned int), lttng_alignof(unsigned int));
struct lttng_stack_trace *trace = stack_trace_context(fdata, cpu);
unsigned int nr_seq_entries;
struct lttng_stack_trace *trace = stack_trace_context(fdata, cpu);
unsigned int nr_seq_entries;
return;
}
nr_seq_entries = trace->nr_entries;
if (trace->nr_entries == MAX_ENTRIES)
nr_seq_entries++;
chan->ops->event_write(ctx, trace->entries,
return;
}
nr_seq_entries = trace->nr_entries;
if (trace->nr_entries == MAX_ENTRIES)
nr_seq_entries++;
chan->ops->event_write(ctx, trace->entries,
/* Add our own ULONG_MAX delimiter to show incomplete stack. */
if (trace->nr_entries == MAX_ENTRIES) {
unsigned long delim = ULONG_MAX;
/* Add our own ULONG_MAX delimiter to show incomplete stack. */
if (trace->nr_entries == MAX_ENTRIES) {
unsigned long delim = ULONG_MAX;
- chan->ops->event_write(ctx, &delim, sizeof(unsigned long));
+ chan->ops->event_write(ctx, &delim, sizeof(unsigned long), 1);