X-Git-Url: https://git.lttng.org/?a=blobdiff_plain;f=include%2Fwrapper%2Fvmalloc.h;h=e6dc8c1ea0f9c39ce4497c959fd9038525aba20c;hb=05355f0bf8b4292ab7b8dc93571a568833b0a3c2;hp=ff279bc1b7042fdb68abccf3844aa1010d2a215c;hpb=0dcc94fab9139ed71921e521f7fc6696808e22c1;p=lttng-modules.git diff --git a/include/wrapper/vmalloc.h b/include/wrapper/vmalloc.h index ff279bc1..e6dc8c1e 100644 --- a/include/wrapper/vmalloc.h +++ b/include/wrapper/vmalloc.h @@ -12,7 +12,7 @@ #ifndef _LTTNG_WRAPPER_VMALLOC_H #define _LTTNG_WRAPPER_VMALLOC_H -#include +#include #include #include @@ -22,7 +22,7 @@ #include #include -#if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,8,0)) +#if (LTTNG_LINUX_VERSION_CODE >= LTTNG_KERNEL_VERSION(5,8,0)) /* * wrapper_vmalloc_sync_mappings was removed in v5.8, the vmalloc mappings @@ -32,17 +32,18 @@ static inline void wrapper_vmalloc_sync_mappings(void) {} -#elif (LINUX_VERSION_CODE >= KERNEL_VERSION(5,6,0) \ +#elif (LTTNG_LINUX_VERSION_CODE >= LTTNG_KERNEL_VERSION(5,6,0) \ || LTTNG_KERNEL_RANGE(5,5,12, 5,6,0) \ || LTTNG_KERNEL_RANGE(5,4,28, 5,5,0) \ || LTTNG_KERNEL_RANGE(5,2,37, 5,3,0) \ || LTTNG_KERNEL_RANGE(4,19,113, 4,20,0) \ || LTTNG_KERNEL_RANGE(4,14,175, 4,15,0) \ || LTTNG_KERNEL_RANGE(4,9,218, 4,10,0) \ - || LTTNG_KERNEL_RANGE(4,4,218, 4,5,0)) \ + || LTTNG_KERNEL_RANGE(4,4,218, 4,5,0) \ || LTTNG_UBUNTU_KERNEL_RANGE(4,15,18,97, 4,16,0,0) \ || LTTNG_UBUNTU_KERNEL_RANGE(5,0,21,48, 5,1,0,0) \ - || LTTNG_UBUNTU_KERNEL_RANGE(5,3,18,52, 5,4,0,0) + || LTTNG_UBUNTU_KERNEL_RANGE(5,3,18,52, 5,4,0,0) \ + || LTTNG_RHEL_KERNEL_RANGE(4,18,0,240,0,0, 4,19,0,0,0,0)) static inline void wrapper_vmalloc_sync_mappings(void) @@ -59,15 +60,28 @@ void wrapper_vmalloc_sync_mappings(void) * trigger recursive page faults. */ printk_once(KERN_WARNING "LTTng: vmalloc_sync_mappings symbol lookup failed.\n"); - printk_once(KERN_WARNING "Page fault handler and NMI tracing might trigger faults.\n"); + printk_once(KERN_WARNING "LTTng: Page fault handler and NMI tracing might trigger faults.\n"); #endif } } -#else /* #if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,6,0)) */ +/* + * Canary function to check for 'vmalloc_sync_mappings()' at compile time. + * + * From 'include/linux/vmalloc.h': + * + * void vmalloc_sync_mappings(void); + */ +static inline +void __canary__vmalloc_sync_mappings(void) +{ + vmalloc_sync_mappings(); +} + +#else /* #if (LTTNG_LINUX_VERSION_CODE >= LTTNG_KERNEL_VERSION(5,6,0)) */ /* - * Map vmalloc_sync_mappings to vmalloc_sync_all() on kernels before 5.7. + * Map vmalloc_sync_mappings to vmalloc_sync_all() on kernels before 5.6. */ static inline void wrapper_vmalloc_sync_mappings(void) @@ -84,16 +98,39 @@ void wrapper_vmalloc_sync_mappings(void) * trigger recursive page faults. */ printk_once(KERN_WARNING "LTTng: vmalloc_sync_all symbol lookup failed.\n"); - printk_once(KERN_WARNING "Page fault handler and NMI tracing might trigger faults.\n"); + printk_once(KERN_WARNING "LTTng: Page fault handler and NMI tracing might trigger faults.\n"); #endif } } -#endif /* #else #if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,6,0)) */ +/* + * Canary function to check for 'vmalloc_sync_all()' at compile time. + * + * From 'include/linux/vmalloc.h': + * + * void vmalloc_sync_all(void); + */ +static inline +void __canary__vmalloc_sync_all(void) +{ + vmalloc_sync_all(); +} + +#endif /* #else #if (LTTNG_LINUX_VERSION_CODE >= LTTNG_KERNEL_VERSION(5,6,0)) */ -#else +#else /* CONFIG_KALLSYMS */ + +#if (LTTNG_LINUX_VERSION_CODE >= LTTNG_KERNEL_VERSION(5,8,0)) + +/* + * wrapper_vmalloc_sync_mappings was removed in v5.8, the vmalloc mappings + * are now synchronized when they are created or torn down. + */ +static inline +void wrapper_vmalloc_sync_mappings(void) +{} -#if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,6,0) \ +#elif (LTTNG_LINUX_VERSION_CODE >= LTTNG_KERNEL_VERSION(5,6,0) \ || LTTNG_KERNEL_RANGE(5,5,12, 5,6,0) \ || LTTNG_KERNEL_RANGE(5,4,28, 5,5,0) \ || LTTNG_KERNEL_RANGE(5,2,37, 5,3,0) \ @@ -111,7 +148,7 @@ void wrapper_vmalloc_sync_mappings(void) return vmalloc_sync_mappings(); } -#else /* #if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,6,0)) */ +#else /* #if (LTTNG_LINUX_VERSION_CODE >= LTTNG_KERNEL_VERSION(5,6,0)) */ static inline void wrapper_vmalloc_sync_mappings(void) @@ -119,11 +156,11 @@ void wrapper_vmalloc_sync_mappings(void) return vmalloc_sync_all(); } -#endif /* #else #if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,6,0)) */ +#endif /* #else #if (LTTNG_LINUX_VERSION_CODE >= LTTNG_KERNEL_VERSION(5,6,0)) */ #endif -#if (LINUX_VERSION_CODE >= KERNEL_VERSION(4,12,0)) +#if (LTTNG_LINUX_VERSION_CODE >= LTTNG_KERNEL_VERSION(4,12,0)) static inline void *lttng_kvmalloc_node(unsigned long size, gfp_t flags, int node) { @@ -172,10 +209,12 @@ static inline void print_vmalloc_node_range_warning(void) { printk_once(KERN_WARNING "LTTng: __vmalloc_node_range symbol lookup failed.\n"); - printk_once(KERN_WARNING "Tracer performance will be degraded on NUMA systems.\n"); - printk_once(KERN_WARNING "Please rebuild your kernel with CONFIG_KALLSYMS enabled.\n"); + printk_once(KERN_WARNING "LTTng: Tracer performance will be degraded on NUMA systems.\n"); + printk_once(KERN_WARNING "LTTng: Please rebuild your kernel with CONFIG_KALLSYMS enabled.\n"); } +#if (LTTNG_LINUX_VERSION_CODE >= LTTNG_KERNEL_VERSION(4,0,0)) + /* * kallsyms wrapper of __vmalloc_node with a fallback to kmalloc_node. */ @@ -204,6 +243,124 @@ void *__lttng_vmalloc_node_range(unsigned long size, unsigned long align, return __vmalloc(size, gfp_mask, prot); } +/* + * Canary function to check for '__vmalloc_node_range()' at compile time. + * + * From 'include/linux/vmalloc.h': + * + * extern void *__vmalloc_node_range(unsigned long size, unsigned long align, + * unsigned long start, unsigned long end, gfp_t gfp_mask, + * pgprot_t prot, unsigned long vm_flags, int node, + * const void *caller); + */ +static inline +void *__canary____lttng_vmalloc_node_range(unsigned long size, unsigned long align, + unsigned long start, unsigned long end, gfp_t gfp_mask, + pgprot_t prot, unsigned long vm_flags, int node, + const void *caller) +{ + return __vmalloc_node_range(size, align, start, end, gfp_mask, prot, + vm_flags, node, caller); +} + +#elif (LTTNG_LINUX_VERSION_CODE >= LTTNG_KERNEL_VERSION(3,6,0)) + +/* + * kallsyms wrapper of __vmalloc_node with a fallback to kmalloc_node. + */ +static inline +void *__lttng_vmalloc_node_range(unsigned long size, unsigned long align, + unsigned long start, unsigned long end, gfp_t gfp_mask, + pgprot_t prot, unsigned long vm_flags, int node, + const void *caller) +{ +#ifdef CONFIG_KALLSYMS + /* + * If we have KALLSYMS, get * __vmalloc_node_range which is not exported. + */ + void *(*lttng__vmalloc_node_range)(unsigned long size, unsigned long align, + unsigned long start, unsigned long end, gfp_t gfp_mask, + pgprot_t prot, int node, const void *caller); + + lttng__vmalloc_node_range = (void *) kallsyms_lookup_funcptr("__vmalloc_node_range"); + if (lttng__vmalloc_node_range) + return lttng__vmalloc_node_range(size, align, start, end, gfp_mask, prot, + node, caller); +#endif + if (node != NUMA_NO_NODE) + print_vmalloc_node_range_warning(); + return __vmalloc(size, gfp_mask, prot); +} + +/* + * Canary function to check for '__vmalloc_node_range()' at compile time. + * + * From 'include/linux/vmalloc.h': + * + * extern void *__vmalloc_node_range(unsigned long size, unsigned long align, + * unsigned long start, unsigned long end, gfp_t gfp_mask, + * pgprot_t prot, unsigned long vm_flags, int node, + * const void *caller); + */ +static inline +void *__canary____lttng_vmalloc_node_range(unsigned long size, unsigned long align, + unsigned long start, unsigned long end, gfp_t gfp_mask, + pgprot_t prot, int node, const void *caller) +{ + return __vmalloc_node_range(size, align, start, end, gfp_mask, prot, + node, caller); +} + +#else /* (LTTNG_LINUX_VERSION_CODE >= LTTNG_KERNEL_VERSION(3,6,0)) */ + +/* + * kallsyms wrapper of __vmalloc_node with a fallback to kmalloc_node. + */ +static inline +void *__lttng_vmalloc_node_range(unsigned long size, unsigned long align, + unsigned long start, unsigned long end, gfp_t gfp_mask, + pgprot_t prot, unsigned long vm_flags, int node, + void *caller) +{ +#ifdef CONFIG_KALLSYMS + /* + * If we have KALLSYMS, get * __vmalloc_node_range which is not exported. + */ + void *(*lttng__vmalloc_node_range)(unsigned long size, unsigned long align, + unsigned long start, unsigned long end, gfp_t gfp_mask, + pgprot_t prot, int node, void *caller); + + lttng__vmalloc_node_range = (void *) kallsyms_lookup_funcptr("__vmalloc_node_range"); + if (lttng__vmalloc_node_range) + return lttng__vmalloc_node_range(size, align, start, end, gfp_mask, prot, + node, caller); +#endif + if (node != NUMA_NO_NODE) + print_vmalloc_node_range_warning(); + return __vmalloc(size, gfp_mask, prot); +} + +/* + * Canary function to check for '__vmalloc_node_range()' at compile time. + * + * From 'include/linux/vmalloc.h': + * + * extern void *__vmalloc_node_range(unsigned long size, unsigned long align, + * unsigned long start, unsigned long end, gfp_t gfp_mask, + * pgprot_t prot, unsigned long vm_flags, int node, + * void *caller); + */ +static inline +void *__canary____lttng_vmalloc_node_range(unsigned long size, unsigned long align, + unsigned long start, unsigned long end, gfp_t gfp_mask, + pgprot_t prot, int node, void *caller) +{ + return __vmalloc_node_range(size, align, start, end, gfp_mask, prot, + node, caller); +} + +#endif + /** * lttng_kvmalloc_node - attempt to allocate physically contiguous memory, but upon * failure, fall back to non-contiguous (vmalloc) allocation.