X-Git-Url: http://git.lttng.org/?a=blobdiff_plain;f=include%2Fwrapper%2Ftrace-clock.h;h=adb1101f2d0c8c6836fe1e9fc63de9dd7552fec2;hb=03edc768329329821d87ff336923977ec3972d7f;hp=7cef08df3176f60d8a27bfd779384452c75e5da3;hpb=117ab60f69a98dcbc365de8008ab93e6699d153a;p=lttng-modules.git diff --git a/include/wrapper/trace-clock.h b/include/wrapper/trace-clock.h index 7cef08df..adb1101f 100644 --- a/include/wrapper/trace-clock.h +++ b/include/wrapper/trace-clock.h @@ -20,14 +20,14 @@ #include #include #include -#include +#include + +#include #include #include #include #include -#include #include -#include extern struct lttng_trace_clock *lttng_trace_clock; @@ -38,8 +38,7 @@ extern struct lttng_trace_clock *lttng_trace_clock; */ #if (LTTNG_KERNEL_RANGE(4,8,0, 4,8,2) \ || LTTNG_KERNEL_RANGE(4,7,4, 4,7,8) \ - || LTTNG_KERNEL_RANGE(4,4,20, 4,4,25) \ - || LTTNG_KERNEL_RANGE(4,1,32, 4,1,35)) + || LTTNG_KERNEL_RANGE(4,4,20, 4,4,25)) #define LTTNG_CLOCK_NMI_SAFE_BROKEN #endif @@ -51,9 +50,7 @@ extern struct lttng_trace_clock *lttng_trace_clock; * this feature on 64-bit architectures. */ -#if (LINUX_VERSION_CODE >= KERNEL_VERSION(3,17,0) \ - && BITS_PER_LONG == 64 \ - && !defined(LTTNG_CLOCK_NMI_SAFE_BROKEN)) +#if (BITS_PER_LONG == 64 && !defined(LTTNG_CLOCK_NMI_SAFE_BROKEN)) #define LTTNG_USE_NMI_SAFE_CLOCK #endif @@ -71,7 +68,7 @@ static inline u64 trace_clock_monotonic_wrapper(void) /* Use fast nmi-safe monotonic clock provided by the Linux kernel. */ preempt_disable(); - last_tsc_ptr = lttng_this_cpu_ptr(<tng_last_tsc); + last_tsc_ptr = this_cpu_ptr(<tng_last_tsc); last = *last_tsc_ptr; /* * Read "last" before "now". It is not strictly required, but it ensures @@ -160,33 +157,30 @@ static inline void put_trace_clock(void) static inline u64 trace_clock_read64(void) { - struct lttng_trace_clock *ltc = READ_ONCE(lttng_trace_clock); + struct lttng_trace_clock *ltc = LTTNG_READ_ONCE(lttng_trace_clock); if (likely(!ltc)) { return trace_clock_read64_monotonic(); } else { - read_barrier_depends(); /* load ltc before content */ return ltc->read64(); } } static inline u64 trace_clock_freq(void) { - struct lttng_trace_clock *ltc = READ_ONCE(lttng_trace_clock); + struct lttng_trace_clock *ltc = LTTNG_READ_ONCE(lttng_trace_clock); if (!ltc) { return trace_clock_freq_monotonic(); } else { - read_barrier_depends(); /* load ltc before content */ return ltc->freq(); } } static inline int trace_clock_uuid(char *uuid) { - struct lttng_trace_clock *ltc = READ_ONCE(lttng_trace_clock); + struct lttng_trace_clock *ltc = LTTNG_READ_ONCE(lttng_trace_clock); - read_barrier_depends(); /* load ltc before content */ /* Use default UUID cb when NULL */ if (!ltc || !ltc->uuid) { return trace_clock_uuid_monotonic(uuid); @@ -197,24 +191,22 @@ static inline int trace_clock_uuid(char *uuid) static inline const char *trace_clock_name(void) { - struct lttng_trace_clock *ltc = READ_ONCE(lttng_trace_clock); + struct lttng_trace_clock *ltc = LTTNG_READ_ONCE(lttng_trace_clock); if (!ltc) { return trace_clock_name_monotonic(); } else { - read_barrier_depends(); /* load ltc before content */ return ltc->name(); } } static inline const char *trace_clock_description(void) { - struct lttng_trace_clock *ltc = READ_ONCE(lttng_trace_clock); + struct lttng_trace_clock *ltc = LTTNG_READ_ONCE(lttng_trace_clock); if (!ltc) { return trace_clock_description_monotonic(); } else { - read_barrier_depends(); /* load ltc before content */ return ltc->description(); } }