#include <linux/string.h>
#include <linux/cpu.h>
#include <lttng/events.h>
+#include <lttng/events-internal.h>
#include <ringbuffer/frontend_types.h>
+#include <wrapper/cpu.h>
#include <wrapper/vmalloc.h>
-#include <wrapper/perf.h>
#include <lttng/tracer.h>
static
-size_t perf_counter_get_size(size_t offset)
+size_t perf_counter_get_size(void *priv, struct lttng_kernel_probe_ctx *probe_ctx, size_t offset)
{
size_t size = 0;
}
static
-void perf_counter_record(struct lttng_kernel_ctx_field *field,
- struct lib_ring_buffer_ctx *ctx,
- struct lttng_channel *chan)
+void perf_counter_record(void *priv, struct lttng_kernel_probe_ctx *probe_ctx,
+ struct lttng_kernel_ring_buffer_ctx *ctx,
+ struct lttng_kernel_channel_buffer *chan)
{
- struct lttng_perf_counter_field *perf_field = field->priv;
+ struct lttng_perf_counter_field *perf_field = (struct lttng_perf_counter_field *) priv;
struct perf_event *event;
uint64_t value;
- event = perf_field->e[ctx->cpu];
+ event = perf_field->e[ctx->priv.reserve_cpu];
if (likely(event)) {
if (unlikely(event->state == PERF_EVENT_STATE_ERROR)) {
value = 0;
*/
value = 0;
}
- lib_ring_buffer_align_ctx(ctx, lttng_alignof(value));
- chan->ops->event_write(ctx, &value, sizeof(value));
+ chan->ops->event_write(ctx, &value, sizeof(value), lttng_alignof(value));
}
#if defined(CONFIG_PERF_EVENTS) && (LTTNG_LINUX_VERSION_CODE >= LTTNG_KERNEL_VERSION(3,1,0))
#endif
static
-void lttng_destroy_perf_counter_ctx_field(struct lttng_kernel_ctx_field *field)
+void lttng_destroy_perf_counter_ctx_field(void *priv)
{
- struct lttng_perf_counter_field *perf_field = field->priv;
+ struct lttng_perf_counter_field *perf_field = priv;
struct perf_event **events = perf_field->e;
#if (LTTNG_LINUX_VERSION_CODE >= LTTNG_KERNEL_VERSION(4,10,0))
{
int cpu;
- get_online_cpus();
+ lttng_cpus_read_lock();
for_each_online_cpu(cpu)
perf_event_release_kernel(events[cpu]);
- put_online_cpus();
+ lttng_cpus_read_unlock();
#ifdef CONFIG_HOTPLUG_CPU
unregister_cpu_notifier(&perf_field->nb);
#endif
struct perf_event_attr *attr = perf_field->attr;
struct perf_event *pevent;
- pevent = wrapper_perf_event_create_kernel_counter(attr,
- cpu, NULL, overflow_callback);
+ pevent = perf_event_create_kernel_counter(attr,
+ cpu, NULL, overflow_callback, NULL);
if (!pevent || IS_ERR(pevent))
return -EINVAL;
if (pevent->state == PERF_EVENT_STATE_ERROR) {
switch (action) {
case CPU_ONLINE:
case CPU_ONLINE_FROZEN:
- pevent = wrapper_perf_event_create_kernel_counter(attr,
- cpu, NULL, overflow_callback);
+ pevent = perf_event_create_kernel_counter(attr,
+ cpu, NULL, overflow_callback, NULL);
if (!pevent || IS_ERR(pevent))
return NOTIFY_BAD;
if (pevent->state == PERF_EVENT_STATE_ERROR) {
perf_field->nb.priority = 0;
register_cpu_notifier(&perf_field->nb);
#endif
- get_online_cpus();
+ lttng_cpus_read_lock();
for_each_online_cpu(cpu) {
- events[cpu] = wrapper_perf_event_create_kernel_counter(attr,
- cpu, NULL, overflow_callback);
+ events[cpu] = perf_event_create_kernel_counter(attr,
+ cpu, NULL, overflow_callback, NULL);
if (!events[cpu] || IS_ERR(events[cpu])) {
ret = -EINVAL;
goto counter_error;
goto counter_busy;
}
}
- put_online_cpus();
+ lttng_cpus_read_unlock();
perf_field->hp_enable = 1;
}
#endif /* #else #if (LTTNG_LINUX_VERSION_CODE >= LTTNG_KERNEL_VERSION(4,10,0)) */
if (events[cpu] && !IS_ERR(events[cpu]))
perf_event_release_kernel(events[cpu]);
}
- put_online_cpus();
+ lttng_cpus_read_unlock();
#ifdef CONFIG_HOTPLUG_CPU
unregister_cpu_notifier(&perf_field->nb);
#endif