From: Francis Giraldeau Date: Wed, 17 Jul 2013 21:05:20 +0000 (-0400) Subject: Add kernel and user callstack contexts X-Git-Tag: v2.11.0-rc1~32 X-Git-Url: http://git.lttng.org/?p=lttng-modules.git;a=commitdiff_plain;h=2fa2d39a42f3407edcb0664bf4d04af28837ea2c Add kernel and user callstack contexts Signed-off-by: Francis Giraldeau Signed-off-by: Mathieu Desnoyers --- diff --git a/Makefile b/Makefile index b08f0bfa..42b7a99e 100644 --- a/Makefile +++ b/Makefile @@ -51,7 +51,8 @@ ifneq ($(KERNELRELEASE),) lttng-context-vtid.o lttng-context-ppid.o \ lttng-context-vppid.o lttng-context-cpu-id.o \ lttng-context-interruptible.o \ - lttng-context-need-reschedule.o lttng-calibrate.o \ + lttng-context-need-reschedule.o \ + lttng-context-callstack.o lttng-calibrate.o \ lttng-context-hostname.o wrapper/random.o \ probes/lttng.o wrapper/trace-clock.o \ wrapper/page_alloc.o \ diff --git a/lttng-abi.c b/lttng-abi.c index ea746c2c..f155c76b 100644 --- a/lttng-abi.c +++ b/lttng-abi.c @@ -247,6 +247,9 @@ long lttng_abi_add_context(struct file *file, return lttng_add_preemptible_to_ctx(ctx); case LTTNG_KERNEL_CONTEXT_MIGRATABLE: return lttng_add_migratable_to_ctx(ctx); + case LTTNG_KERNEL_CONTEXT_CALLSTACK_KERNEL: + case LTTNG_KERNEL_CONTEXT_CALLSTACK_USER: + return lttng_add_callstack_to_ctx(ctx, context_param->ctx); default: return -EINVAL; } diff --git a/lttng-abi.h b/lttng-abi.h index dac86584..c8c4e298 100644 --- a/lttng-abi.h +++ b/lttng-abi.h @@ -146,6 +146,8 @@ enum lttng_kernel_context_type { LTTNG_KERNEL_CONTEXT_PREEMPTIBLE = 13, LTTNG_KERNEL_CONTEXT_NEED_RESCHEDULE = 14, LTTNG_KERNEL_CONTEXT_MIGRATABLE = 15, + LTTNG_KERNEL_CONTEXT_CALLSTACK_KERNEL = 16, + LTTNG_KERNEL_CONTEXT_CALLSTACK_USER = 17, }; struct lttng_kernel_perf_counter_ctx { diff --git a/lttng-context-callstack.c b/lttng-context-callstack.c new file mode 100644 index 00000000..8c3ee3dc --- /dev/null +++ b/lttng-context-callstack.c @@ -0,0 +1,334 @@ +/* + * lttng-context-callstack.c + * + * LTTng callstack event context. + * + * Copyright (C) 2014 Mathieu Desnoyers + * Copyright (C) 2014 Francis Giraldeau + * + * This library is free software; you can redistribute it and/or + * modify it under the terms of the GNU Lesser General Public + * License as published by the Free Software Foundation; only + * version 2.1 of the License. + * + * This library is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Lesser General Public License for more details. + * + * You should have received a copy of the GNU Lesser General Public + * License along with this library; if not, write to the Free Software + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + * + * The callstack context can be added to any kernel + * event. It records either the kernel or the userspace callstack, up to a + * max depth. The context is a CTF sequence, such that it uses only the space + * required for the number of callstack entries. + * + * It allocates callstack buffers per-CPU up to 4 interrupt nesting. This + * nesting limit is the same as defined in the ring buffer. It therefore uses a + * fixed amount of memory, proportional to the number of CPUs: + * + * size = cpus * nest * depth * sizeof(unsigned long) + * + * Which is about 800 bytes per-CPUs on 64-bit host and a depth of 25. The + * allocation is done at the initialization to avoid memory allocation + * overhead while tracing, using a shallow stack. + * + * The kernel callstack is recovered using save_stack_trace(), and the + * userspace callstack uses save_stack_trace_user(). They rely on frame + * pointers. These are usually available for the kernel, but the compiler + * option -fomit-frame-pointer frequently used in popular Linux distributions + * may cause the userspace callstack to be unreliable, and is a known + * limitation of this approach. If frame pointers are not available, it + * produces no error, but the callstack will be empty. We still provide the + * feature, because it works well for runtime environments having frame + * pointers. In the future, unwind support and/or last branch record may + * provide a solution to this problem. + * + * The symbol name resolution is left to the trace reader. + */ + +#include +#include +#include +#include +#include +#include +#include "lttng-events.h" +#include "wrapper/ringbuffer/backend.h" +#include "wrapper/ringbuffer/frontend.h" +#include "wrapper/vmalloc.h" +#include "lttng-tracer.h" + +#define MAX_ENTRIES 25 /* BUG: saving more than 30 entries causes trace corruption */ + +struct lttng_cs { + struct stack_trace items[RING_BUFFER_MAX_NESTING]; +}; + +struct field_data { + int mode; + struct lttng_cs __percpu *cs_percpu; +}; + +struct lttng_cs_type { + const char *name; + const char *save_func_name; + void (*save_func)(struct stack_trace *trace); +}; + +enum lttng_cs_ctx_modes { + CALLSTACK_KERNEL = 0, + CALLSTACK_USER = 1, +}; + +static struct lttng_cs_type cs_types[] = { + { + .name = "callstack_kernel", + .save_func_name = "save_stack_trace", + .save_func = NULL, + }, + { + .name = "callstack_user", + .save_func_name = "save_stack_trace_user", + .save_func = NULL, + }, +}; + +static +int init_type(int mode) +{ + unsigned long func; + + if (cs_types[mode].save_func) + return 0; + func = kallsyms_lookup_funcptr(cs_types[mode].save_func_name); + if (!func) { + printk(KERN_WARNING "LTTng: symbol lookup failed: %s\n", + cs_types[mode].save_func_name); + return -EINVAL; + } + cs_types[mode].save_func = (void *) func; + return 0; +} + +static +struct stack_trace *stack_trace_context(struct lttng_ctx_field *field, + struct lib_ring_buffer_ctx *ctx) +{ + int nesting; + struct lttng_cs *cs; + struct field_data *fdata = field->private; + + /* + * get_cpu() is not required, preemption is already + * disabled while event is written. + * + * max nesting is checked in lib_ring_buffer_get_cpu(). + * Check it again as a safety net. + */ + cs = per_cpu_ptr(fdata->cs_percpu, ctx->cpu); + nesting = per_cpu(lib_ring_buffer_nesting, ctx->cpu) - 1; + if (nesting >= RING_BUFFER_MAX_NESTING) { + return NULL; + } + return &cs->items[nesting]; +} + +/* + * In order to reserve the correct size, the callstack is computed. The + * resulting callstack is saved to be accessed in the record step. + */ +static +size_t lttng_callstack_get_size(size_t offset, struct lttng_ctx_field *field, + struct lib_ring_buffer_ctx *ctx, + struct lttng_channel *chan) +{ + size_t size = 0; + struct stack_trace *trace; + struct field_data *fdata = field->private; + + /* do not write data if no space is available */ + trace = stack_trace_context(field, ctx); + if (!trace) + return 0; + + /* reset stack trace, no need to clear memory */ + trace->nr_entries = 0; + + /* do the real work and reserve space */ + cs_types[fdata->mode].save_func(trace); + size += lib_ring_buffer_align(offset, lttng_alignof(unsigned int)); + size += sizeof(unsigned int); + size += lib_ring_buffer_align(offset, lttng_alignof(unsigned long)); + size += sizeof(unsigned long) * trace->nr_entries; + return size; +} + +static +void lttng_callstack_record(struct lttng_ctx_field *field, + struct lib_ring_buffer_ctx *ctx, + struct lttng_channel *chan) +{ + struct stack_trace *trace = stack_trace_context(field, ctx); + + if (!trace) + return; + lib_ring_buffer_align_ctx(ctx, lttng_alignof(unsigned int)); + chan->ops->event_write(ctx, &trace->nr_entries, sizeof(unsigned int)); + lib_ring_buffer_align_ctx(ctx, lttng_alignof(unsigned long)); + chan->ops->event_write(ctx, trace->entries, + sizeof(unsigned long) * trace->nr_entries); +} + +static +void field_data_free(struct field_data *fdata) +{ + int cpu, i; + struct lttng_cs *cs; + + if (!fdata) + return; + for_each_possible_cpu(cpu) { + cs = per_cpu_ptr(fdata->cs_percpu, cpu); + for (i = 0; i < RING_BUFFER_MAX_NESTING; i++) { + kfree(cs->items[i].entries); + } + } + free_percpu(fdata->cs_percpu); + kfree(fdata); +} + +static +struct field_data __percpu *field_data_create(unsigned int entries, int type) +{ + int cpu, i; + struct stack_trace *item; + struct lttng_cs *cs; + struct lttng_cs __percpu *cs_set; + struct field_data* fdata; + + fdata = kzalloc(sizeof(unsigned long) * entries, GFP_KERNEL); + if (!fdata) + return NULL; + cs_set = alloc_percpu(struct lttng_cs); + if (!cs_set) + goto error_alloc; + + fdata->cs_percpu = cs_set; + for_each_possible_cpu(cpu) { + cs = per_cpu_ptr(cs_set, cpu); + for (i = 0; i < RING_BUFFER_MAX_NESTING; i++) { + item = &cs->items[i]; + item->entries = kzalloc(sizeof(unsigned long) * entries, GFP_KERNEL); + if (!item->entries) { + goto error_alloc; + } + item->max_entries = entries; + } + } + fdata->mode = type; + return fdata; + +error_alloc: + field_data_free(fdata); + return NULL; +} + +static +void lttng_callstack_destroy(struct lttng_ctx_field *field) +{ + struct field_data *fdata = field->private; + + field_data_free(fdata); +} + +static +int __lttng_add_callstack_generic(struct lttng_ctx **ctx, int mode) +{ + const char *ctx_name = cs_types[mode].name; + struct lttng_ctx_field *field; + struct field_data *fdata; + int ret; + + ret = init_type(mode); + if (ret) + return ret; + field = lttng_append_context(ctx); + if (!field) + return -ENOMEM; + if (lttng_find_context(*ctx, ctx_name)) { + printk("%s lttng_find_context failed\n", ctx_name); + ret = -EEXIST; + goto error_find; + } + fdata = field_data_create(MAX_ENTRIES, mode); + if (!fdata) { + ret = -ENOMEM; + goto error_create; + } + + field->event_field.name = ctx_name; + field->event_field.type.atype = atype_sequence; + field->event_field.type.u.sequence.elem_type.atype = atype_integer; + field->event_field.type.u.sequence.elem_type.u.basic.integer.size = sizeof(unsigned long) * CHAR_BIT; + field->event_field.type.u.sequence.elem_type.u.basic.integer.alignment = lttng_alignof(long) * CHAR_BIT; + field->event_field.type.u.sequence.elem_type.u.basic.integer.signedness = lttng_is_signed_type(unsigned long); + field->event_field.type.u.sequence.elem_type.u.basic.integer.reverse_byte_order = 0; + field->event_field.type.u.sequence.elem_type.u.basic.integer.base = 16; + field->event_field.type.u.sequence.elem_type.u.basic.integer.encoding = lttng_encode_none; + + field->event_field.type.u.sequence.length_type.atype = atype_integer; + field->event_field.type.u.sequence.length_type.u.basic.integer.size = sizeof(unsigned int) * CHAR_BIT; + field->event_field.type.u.sequence.length_type.u.basic.integer.alignment = lttng_alignof(unsigned int) * CHAR_BIT; + field->event_field.type.u.sequence.length_type.u.basic.integer.signedness = lttng_is_signed_type(unsigned int); + field->event_field.type.u.sequence.length_type.u.basic.integer.reverse_byte_order = 0; + field->event_field.type.u.sequence.length_type.u.basic.integer.base = 10; + field->event_field.type.u.sequence.length_type.u.basic.integer.encoding = lttng_encode_none; + + field->get_size_arg = lttng_callstack_get_size; + field->record = lttng_callstack_record; + field->private = fdata; + field->destroy = lttng_callstack_destroy; + wrapper_vmalloc_sync_all(); + printk("lttng add-context %s\n", ctx_name); + return 0; + +error_create: + field_data_free(fdata); +error_find: + lttng_remove_context_field(ctx, field); + return ret; +} + +/** + * lttng_add_callstack_to_ctx - add callstack event context + * + * @ctx: the lttng_ctx pointer to initialize + * @type: the context type + * + * Supported callstack type supported: + * LTTNG_KERNEL_CONTEXT_CALLSTACK_KERNEL + * Records the callstack of the kernel + * LTTNG_KERNEL_CONTEXT_CALLSTACK_USER + * Records the callstack of the userspace program (from the kernel) + * + * Return 0 for success, or error code. + */ +int lttng_add_callstack_to_ctx(struct lttng_ctx **ctx, int type) +{ + switch (type) { + case LTTNG_KERNEL_CONTEXT_CALLSTACK_KERNEL: + return __lttng_add_callstack_generic(ctx, CALLSTACK_KERNEL); + case LTTNG_KERNEL_CONTEXT_CALLSTACK_USER: + return __lttng_add_callstack_generic(ctx, CALLSTACK_USER); + default: + return -EINVAL; + } +} +EXPORT_SYMBOL_GPL(lttng_add_callstack_to_ctx); + +MODULE_LICENSE("GPL and additional rights"); +MODULE_AUTHOR("Francis Giraldeau"); +MODULE_DESCRIPTION("Linux Trace Toolkit Callstack Support"); diff --git a/lttng-events.h b/lttng-events.h index 19dd2d30..aca64545 100644 --- a/lttng-events.h +++ b/lttng-events.h @@ -219,6 +219,8 @@ struct lttng_ctx_field { struct lttng_perf_counter_field *perf_counter; } u; void (*destroy)(struct lttng_ctx_field *field); + /* private data to keep state between get_size and record */ + void *private; }; struct lttng_ctx { @@ -708,6 +710,9 @@ int lttng_add_migratable_to_ctx(struct lttng_ctx **ctx) return -ENOSYS; } #endif + +int lttng_add_callstack_to_ctx(struct lttng_ctx **ctx, int type); + #if defined(CONFIG_PERF_EVENTS) && (LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,33)) int lttng_add_perf_counter_to_ctx(uint32_t type, uint64_t config,