X-Git-Url: https://git.lttng.org/?a=blobdiff_plain;f=wrapper%2Fvmalloc.h;h=362054b162b9e80eab799ca9be66298a697bb068;hb=60c64030e811165e62a62eb551459dd2416d6ec2;hp=5d0b38987d6b4d0f5cee6a0b6f8db696564b3a7f;hpb=4d1817c129310945bf0b8227ebc0fbb4f6299dc7;p=lttng-modules.git diff --git a/wrapper/vmalloc.h b/wrapper/vmalloc.h index 5d0b3898..362054b1 100644 --- a/wrapper/vmalloc.h +++ b/wrapper/vmalloc.h @@ -20,8 +20,29 @@ #include #include +#include -#if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,7,0)) +#if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,8,0)) + +/* + * wrapper_vmalloc_sync_mappings was removed in v5.8, the vmalloc mappings + * are now synchronized when they are created or torn down. + */ +static inline +void wrapper_vmalloc_sync_mappings(void) +{} + +#elif (LINUX_VERSION_CODE >= KERNEL_VERSION(5,6,0) \ + || LTTNG_KERNEL_RANGE(5,5,12, 5,6,0) \ + || LTTNG_KERNEL_RANGE(5,4,28, 5,5,0) \ + || LTTNG_KERNEL_RANGE(5,2,37, 5,3,0) \ + || LTTNG_KERNEL_RANGE(4,19,113, 4,20,0) \ + || LTTNG_KERNEL_RANGE(4,14,175, 4,15,0) \ + || LTTNG_KERNEL_RANGE(4,9,218, 4,10,0) \ + || LTTNG_KERNEL_RANGE(4,4,218, 4,5,0)) \ + || LTTNG_UBUNTU_KERNEL_RANGE(4,15,18,97, 4,16,0,0) \ + || LTTNG_UBUNTU_KERNEL_RANGE(5,0,21,48, 5,1,0,0) \ + || LTTNG_UBUNTU_KERNEL_RANGE(5,3,18,52, 5,4,0,0) static inline void wrapper_vmalloc_sync_mappings(void) @@ -43,10 +64,23 @@ void wrapper_vmalloc_sync_mappings(void) } } -#else /* #if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,7,0)) */ +/* + * Canary function to check for 'vmalloc_sync_mappings()' at compile time. + * + * From 'include/linux/vmalloc.h': + * + * void vmalloc_sync_mappings(void); + */ +static inline +void __canary__vmalloc_sync_mappings(void) +{ + vmalloc_sync_mappings(); +} + +#else /* #if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,6,0)) */ /* - * Map vmalloc_sync_mappings to vmalloc_sync_all() on kernels before 5.7. + * Map vmalloc_sync_mappings to vmalloc_sync_all() on kernels before 5.6. */ static inline void wrapper_vmalloc_sync_mappings(void) @@ -68,11 +102,34 @@ void wrapper_vmalloc_sync_mappings(void) } } -#endif /* #else #if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,7,0)) */ +/* + * Canary function to check for 'vmalloc_sync_all()' at compile time. + * + * From 'include/linux/vmalloc.h': + * + * void vmalloc_sync_all(void); + */ +static inline +void __canary__vmalloc_sync_all(void) +{ + vmalloc_sync_all(); +} + +#endif /* #else #if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,6,0)) */ #else -#if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,7,0)) +#if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,6,0) \ + || LTTNG_KERNEL_RANGE(5,5,12, 5,6,0) \ + || LTTNG_KERNEL_RANGE(5,4,28, 5,5,0) \ + || LTTNG_KERNEL_RANGE(5,2,37, 5,3,0) \ + || LTTNG_KERNEL_RANGE(4,19,113, 4,20,0) \ + || LTTNG_KERNEL_RANGE(4,14,175, 4,15,0) \ + || LTTNG_KERNEL_RANGE(4,9,218, 4,10,0) \ + || LTTNG_KERNEL_RANGE(4,4,218, 4,5,0)) \ + || LTTNG_UBUNTU_KERNEL_RANGE(4,15,18,97, 4,18,0,0) \ + || LTTNG_UBUNTU_KERNEL_RANGE(5,0,21,48, 5,1,0,0) \ + || LTTNG_UBUNTU_KERNEL_RANGE(5,3,18,52, 5,4,0,0) static inline void wrapper_vmalloc_sync_mappings(void) @@ -80,7 +137,7 @@ void wrapper_vmalloc_sync_mappings(void) return vmalloc_sync_mappings(); } -#else /* #if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,7,0)) */ +#else /* #if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,6,0)) */ static inline void wrapper_vmalloc_sync_mappings(void) @@ -88,7 +145,7 @@ void wrapper_vmalloc_sync_mappings(void) return vmalloc_sync_all(); } -#endif /* #else #if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,7,0)) */ +#endif /* #else #if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,6,0)) */ #endif @@ -173,6 +230,26 @@ void *__lttng_vmalloc_node_range(unsigned long size, unsigned long align, return __vmalloc(size, gfp_mask, prot); } +/* + * Canary function to check for '__vmalloc_node_range()' at compile time. + * + * From 'include/linux/vmalloc.h': + * + * extern void *__vmalloc_node_range(unsigned long size, unsigned long align, + * unsigned long start, unsigned long end, gfp_t gfp_mask, + * pgprot_t prot, unsigned long vm_flags, int node, + * const void *caller); + */ +static inline +void *__canary____lttng_vmalloc_node_range(unsigned long size, unsigned long align, + unsigned long start, unsigned long end, gfp_t gfp_mask, + pgprot_t prot, unsigned long vm_flags, int node, + const void *caller) +{ + return __vmalloc_node_range(size, align, start, end, gfp_mask, prot, + vm_flags, node, caller); +} + /** * lttng_kvmalloc_node - attempt to allocate physically contiguous memory, but upon * failure, fall back to non-contiguous (vmalloc) allocation.