/* MECHANICALLY GENERATED, DO NOT EDIT!!! */
+#ifndef _INCLUDE_API_H
#define _INCLUDE_API_H
/*
* to redistribute under later versions of GPL might not be available.
*/
+#include <urcu/arch.h>
+
#ifndef __always_inline
#define __always_inline inline
#endif
#define CONFIG_PPC64
-#define CACHE_LINE_SIZE 128
+/*#define CAA_CACHE_LINE_SIZE 128 */
#define ____cacheline_internodealigned_in_smp \
__attribute__((__aligned__(1 << 7)))
+#if 0 /* duplicate with arch_atomic.h */
+
/*
* Atomic data structure, initialization, and access.
*/
# define ISYNC_ON_SMP
#endif
-
/*
* Atomic exchange
*
#define atomic_dec_return(v) (atomic_sub_return(1,v))
/* Atomic operations are already serializing on x86 */
-#define smp_mb__before_atomic_dec() smp_mb()
-#define smp_mb__after_atomic_dec() smp_mb()
-#define smp_mb__before_atomic_inc() smp_mb()
-#define smp_mb__after_atomic_inc() smp_mb()
+#define smp_mb__before_atomic_dec() cmm_smp_mb()
+#define smp_mb__after_atomic_dec() cmm_smp_mb()
+#define smp_mb__before_atomic_inc() cmm_smp_mb()
+#define smp_mb__after_atomic_inc() cmm_smp_mb()
+
+#endif //0 /* duplicate with arch_atomic.h */
/*
* api_pthreads.h: API mapping to pthreads environment.
#include <sys/param.h>
/* #include "atomic.h" */
-/*
- * Compiler magic.
- */
-#define offsetof(TYPE, MEMBER) ((size_t) &((TYPE *)0)->MEMBER)
-#define container_of(ptr, type, member) ({ \
- const typeof( ((type *)0)->member ) *__mptr = (ptr); \
- (type *)( (char *)__mptr - offsetof(type,member) );})
-
/*
* Default machine parameters.
*/
-#ifndef CACHE_LINE_SIZE
-#define CACHE_LINE_SIZE 128
-#endif /* #ifndef CACHE_LINE_SIZE */
+#ifndef CAA_CACHE_LINE_SIZE
+/* #define CAA_CACHE_LINE_SIZE 128 */
+#endif /* #ifndef CAA_CACHE_LINE_SIZE */
/*
* Exclusive locking primitives.
#define DEFINE_PER_THREAD(type, name) \
struct { \
__typeof__(type) v \
- __attribute__((__aligned__(CACHE_LINE_SIZE))); \
+ __attribute__((__aligned__(CAA_CACHE_LINE_SIZE))); \
} __per_thread_##name[NR_THREADS];
#define DECLARE_PER_THREAD(type, name) extern DEFINE_PER_THREAD(type, name)
#define DEFINE_PER_CPU(type, name) \
struct { \
__typeof__(type) v \
- __attribute__((__aligned__(CACHE_LINE_SIZE))); \
+ __attribute__((__aligned__(CAA_CACHE_LINE_SIZE))); \
} __per_cpu_##name[NR_CPUS]
#define DECLARE_PER_CPU(type, name) extern DEFINE_PER_CPU(type, name)
#define LIST_POISON1 ((void *) 0x00100100)
#define LIST_POISON2 ((void *) 0x00200200)
-#define offsetof(TYPE, MEMBER) ((size_t) &((TYPE *)0)->MEMBER)
-#define container_of(ptr, type, member) ({ \
- const typeof( ((type *)0)->member ) *__mptr = (ptr); \
- (type *)( (char *)__mptr - offsetof(type,member) );})
+#if 0
/*
* Simple doubly linked list implementation.
* @member: the name of the list_struct within the struct.
*/
#define list_entry(ptr, type, member) \
- container_of(ptr, type, member)
+ caa_container_of(ptr, type, member)
/**
* list_first_entry - get the first element from a list
&pos->member != (head); \
pos = n, n = list_entry(n->member.prev, typeof(*n), member))
+#endif //0
+
/*
* Double linked lists with a single pointer list head.
* Mostly useful for hash tables where the two pointer list head is
old->first = NULL;
}
-#define hlist_entry(ptr, type, member) container_of(ptr,type,member)
+#define hlist_entry(ptr, type, member) caa_container_of(ptr,type,member)
#define hlist_for_each(pos, head) \
for (pos = (head)->first; pos && ({ prefetch(pos->next); 1; }); \
pos = n)
#endif
+
+#endif