X-Git-Url: http://git.lttng.org/?a=blobdiff_plain;f=share%2Fkernelcompat.h;h=b0c77e0972a66f8d03235494f4e5be72fcbc281a;hb=772030fed323e388da467735cf4b5e8781acb710;hp=9139e75275ddf4c8c3e86d7d30c0916b32233b6d;hpb=98963de4a2dfae12d8aafa0f9a6d97cf4a44e12d;p=ust.git diff --git a/share/kernelcompat.h b/share/kernelcompat.h index 9139e75..b0c77e0 100644 --- a/share/kernelcompat.h +++ b/share/kernelcompat.h @@ -1,9 +1,16 @@ #ifndef KERNELCOMPAT_H #define KERNELCOMPAT_H -#include "compiler.h" +#include +#include #include +#include + +/* FIXME: libkcompat must not define arch-specific local ops, as ust *must* + * fallback to the normal atomic ops. Fix things so we don't add them and + * break things accidentally. + */ #define container_of(ptr, type, member) ({ \ const typeof( ((type *)0)->member ) *__mptr = (ptr); \ @@ -38,14 +45,7 @@ static inline long IS_ERR(const void *ptr) } -/* FIXED SIZE INTEGERS */ - -#include - -typedef uint8_t u8; -typedef uint16_t u16; -typedef uint32_t u32; -typedef uint64_t u64; +/* Min / Max */ #define min_t(type, x, y) ({ \ type __min1 = (x); \ @@ -69,14 +69,6 @@ typedef uint64_t u64; #define mutex_unlock(m) pthread_mutex_unlock(m) -/* SPINLOCKS */ - -typedef int spinlock_t; - -#define spin_lock(a) /* nothing */ -#define spin_unlock(a) /* nothing */ -#define spin_lock_init(a) /* nothing */ - /* MALLOCATION */ @@ -96,142 +88,10 @@ typedef int spinlock_t; #include #define printk(fmt, args...) printf(fmt, ## args) -/* MEMORY BARRIERS */ - -#define smp_rmb() do {} while(0) -#define smp_wmb() do {} while(0) -#define smp_mb() do {} while(0) -#define smp_mb__after_atomic_inc() do {} while(0) - -#define read_barrier_depends() do {} while(0) -#define smp_read_barrier_depends() do {} while(0) - -/* RCU */ - -#define rcu_assign_pointer(a, b) do {} while(0) -#define call_rcu_sched(a,b) do {} while(0) -#define rcu_barrier_sched() do {} while(0) -#define rcu_read_lock_sched_notrace() do{} while (0) -#define rcu_read_unlock_sched_notrace() do{} while (0) - -/* ATOMICITY */ - -#include - -typedef struct { sig_atomic_t counter; } atomic_t; - -static inline int atomic_dec_and_test(atomic_t *p) -{ - (p->counter)--; - return !p->counter; -} - -static inline void atomic_set(atomic_t *p, int v) -{ - p->counter=v; -} - -static inline void atomic_inc(atomic_t *p) -{ - p->counter++; -} - -static int atomic_read(atomic_t *p) -{ - return p->counter; -} - -#define atomic_long_t atomic_t -#define atomic_long_set atomic_set -#define atomic_long_read atomic_read - -#include "asm.h" - -#define __xg(x) ((volatile long *)(x)) - -#define cmpxchg(ptr, o, n) \ - ((__typeof__(*(ptr)))__cmpxchg((ptr), (unsigned long)(o), \ - (unsigned long)(n), sizeof(*(ptr)))) - -static inline unsigned long __cmpxchg(volatile void *ptr, unsigned long old, - unsigned long new, int size) -{ - unsigned long prev; - switch (size) { - case 1: - asm volatile("lock; cmpxchgb %b1,%2" - : "=a"(prev) - : "q"(new), "m"(*__xg(ptr)), "0"(old) - : "memory"); - return prev; - case 2: - asm volatile("lock; cmpxchgw %w1,%2" - : "=a"(prev) - : "r"(new), "m"(*__xg(ptr)), "0"(old) - : "memory"); - return prev; - case 4: - asm volatile("lock; cmpxchgl %k1,%2" - : "=a"(prev) - : "r"(new), "m"(*__xg(ptr)), "0"(old) - : "memory"); - return prev; - case 8: - asm volatile("lock; cmpxchgq %1,%2" - : "=a"(prev) - : "r"(new), "m"(*__xg(ptr)), "0"(old) - : "memory"); - return prev; - } - return old; -} - -//#define local_cmpxchg cmpxchg -#define local_cmpxchg(l, o, n) (cmpxchg(&((l)->a.counter), (o), (n))) - -#define atomic_long_cmpxchg(v, old, new) (cmpxchg(&((v)->counter), (old), (new))) - - -/* LOCAL OPS */ - -//typedef int local_t; -typedef struct -{ - atomic_long_t a; -} local_t; - - -static inline void local_inc(local_t *l) -{ - (l->a.counter)++; -} - -static inline void local_set(local_t *l, int v) -{ - l->a.counter = v; -} - -static inline void local_add(int v, local_t *l) -{ - l->a.counter += v; -} - -static int local_add_return(int v, local_t *l) -{ - return l->a.counter += v; -} - -static inline int local_read(local_t *l) -{ - return l->a.counter; -} - /* ATTRIBUTES */ #define ____cacheline_aligned -#define __init -#define __exit /* MATH */ @@ -279,7 +139,7 @@ static __inline__ int get_count_order(unsigned int count) #define __ALIGN_MASK(x,mask) (((x)+(mask))&~(mask)) #define PAGE_ALIGN(addr) ALIGN(addr, PAGE_SIZE) #define PAGE_SIZE sysconf(_SC_PAGE_SIZE) -#define PAGE_MASK (PAGE_SIZE-1) +#define PAGE_MASK (~(PAGE_SIZE-1)) @@ -290,30 +150,30 @@ static __inline__ int get_count_order(unsigned int count) /* TRACE CLOCK */ -//ust// static inline u64 trace_clock_read64(void) -//ust// { -//ust// uint32_t low; -//ust// uint32_t high; -//ust// uint64_t retval; -//ust// __asm__ volatile ("rdtsc\n" : "=a" (low), "=d" (high)); -//ust// -//ust// retval = high; -//ust// retval <<= 32; -//ust// return retval | low; -//ust// } - static inline u64 trace_clock_read64(void) { - struct timeval tv; - u64 retval; - - gettimeofday(&tv, NULL); - retval = tv.tv_sec; - retval *= 1000000; - retval += tv.tv_usec; - - return retval; -} + uint32_t low; + uint32_t high; + uint64_t retval; + __asm__ volatile ("rdtsc\n" : "=a" (low), "=d" (high)); + + retval = high; + retval <<= 32; + return retval | low; +} + +//static inline u64 trace_clock_read64(void) +//{ +// struct timeval tv; +// u64 retval; +// +// gettimeofday(&tv, NULL); +// retval = tv.tv_sec; +// retval *= 1000000; +// retval += tv.tv_usec; +// +// return retval; +//} static inline u64 trace_clock_frequency(void) { @@ -326,14 +186,4 @@ static inline u32 trace_clock_freq_scale(void) } -/* LISTS */ - -#define list_add_rcu list_add -#define list_for_each_entry_rcu list_for_each_entry - - -#define EXPORT_SYMBOL_GPL(a) /*nothing*/ - -#define smp_processor_id() (-1) - #endif /* KERNELCOMPAT_H */