X-Git-Url: http://git.lttng.org/?a=blobdiff_plain;f=liblttng-ust-libc-wrapper%2Flttng-ust-malloc.c;h=e1ff8e807f7a9fb22dc6f783230ce6f1b1693e4f;hb=6ba6fd60507f8e045bdc4f1be14e9d99c6a15f7f;hp=be13f214480bfb09402b25d3328fd36c69d3e2b6;hpb=cd61d9bfa5fd334b086831dcbdec63da6c0d1a16;p=lttng-ust.git diff --git a/liblttng-ust-libc-wrapper/lttng-ust-malloc.c b/liblttng-ust-libc-wrapper/lttng-ust-malloc.c index be13f214..e1ff8e80 100644 --- a/liblttng-ust-libc-wrapper/lttng-ust-malloc.c +++ b/liblttng-ust-libc-wrapper/lttng-ust-malloc.c @@ -14,11 +14,16 @@ #include #include #include +#include + #include #include #include #include #include + +#include + #include #include "ust-compat.h" @@ -56,8 +61,11 @@ void *static_calloc(size_t nmemb, size_t size); */ static int ust_malloc_lock; -static __attribute__((unused)) +static void ust_malloc_spin_lock(pthread_mutex_t *lock) + __attribute__((unused)); +static +void ust_malloc_spin_lock(pthread_mutex_t *lock __attribute__((unused))) { /* * The memory barrier within cmpxchg takes care of ordering @@ -68,8 +76,11 @@ void ust_malloc_spin_lock(pthread_mutex_t *lock) caa_cpu_relax(); } -static __attribute__((unused)) +static void ust_malloc_spin_unlock(pthread_mutex_t *lock) + __attribute__((unused)); +static +void ust_malloc_spin_unlock(pthread_mutex_t *lock __attribute__((unused))) { /* * Ensure memory accesses within the critical section do not @@ -139,7 +150,7 @@ void *static_malloc(size_t size) } static -void static_free(void *ptr) +void static_free(void *ptr __attribute__((unused))) { /* no-op. */ } @@ -414,8 +425,7 @@ void lttng_ust_fixup_malloc_nesting_tls(void) asm volatile ("" : : "m" (URCU_TLS(malloc_nesting))); } -__attribute__((constructor)) -void lttng_ust_malloc_wrapper_init(void) +void lttng_ust_libc_wrapper_malloc_init(void) { /* Initialization already done */ if (cur_alloc.calloc) {