projects
/
urcu.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Rename all arch primitives with prefix caa_
[urcu.git]
/
tests
/
api_gcc.h
diff --git
a/tests/api_gcc.h
b/tests/api_gcc.h
index 872c873f9760da2b812c54aa19a2dc4d0c7894f0..632bdd5b4ce132132a08592a5b6c2a64dffc9aa5 100644
(file)
--- a/
tests/api_gcc.h
+++ b/
tests/api_gcc.h
@@
-2,6
+2,8
@@
#ifndef _INCLUDE_API_H
#define _INCLUDE_API_H
#ifndef _INCLUDE_API_H
#define _INCLUDE_API_H
+#include "../config.h"
+
/*
* common.h: Common Linux kernel-isms.
*
/*
* common.h: Common Linux kernel-isms.
*
@@
-71,7
+73,7
@@
* Machine parameters.
*/
* Machine parameters.
*/
-/* #define CACHE_LINE_SIZE 64 */
+/* #define CA
A_CA
CHE_LINE_SIZE 64 */
#define ____cacheline_internodealigned_in_smp \
__attribute__((__aligned__(1 << 6)))
#define ____cacheline_internodealigned_in_smp \
__attribute__((__aligned__(1 << 6)))
@@
-249,10
+251,10
@@
cmpxchg(volatile long *ptr, long oldval, long newval)
#define atomic_dec_return(v) (atomic_sub_return(1,v))
/* Atomic operations are already serializing on x86 */
#define atomic_dec_return(v) (atomic_sub_return(1,v))
/* Atomic operations are already serializing on x86 */
-#define smp_mb__before_atomic_dec() barrier()
-#define smp_mb__after_atomic_dec() barrier()
-#define smp_mb__before_atomic_inc() barrier()
-#define smp_mb__after_atomic_inc() barrier()
+#define smp_mb__before_atomic_dec()
cmm_
barrier()
+#define smp_mb__after_atomic_dec()
cmm_
barrier()
+#define smp_mb__before_atomic_inc()
cmm_
barrier()
+#define smp_mb__after_atomic_inc()
cmm_
barrier()
#endif //0 /* duplicate with arch_atomic.h */
#endif //0 /* duplicate with arch_atomic.h */
@@
-289,20
+291,13
@@
cmpxchg(volatile long *ptr, long oldval, long newval)
#include <sys/param.h>
/* #include "atomic.h" */
#include <sys/param.h>
/* #include "atomic.h" */
-/*
- * Compiler magic.
- */
-#define container_of(ptr, type, member) ({ \
- const typeof( ((type *)0)->member ) *__mptr = (ptr); \
- (type *)( (char *)__mptr - offsetof(type,member) );})
-
/*
* Default machine parameters.
*/
/*
* Default machine parameters.
*/
-#ifndef CACHE_LINE_SIZE
-/* #define CACHE_LINE_SIZE 128 */
-#endif /* #ifndef CACHE_LINE_SIZE */
+#ifndef CA
A_CA
CHE_LINE_SIZE
+/* #define CA
A_CA
CHE_LINE_SIZE 128 */
+#endif /* #ifndef CA
A_CA
CHE_LINE_SIZE */
/*
* Exclusive locking primitives.
/*
* Exclusive locking primitives.
@@
-461,13
+456,25
@@
static void wait_all_threads(void)
}
}
}
}
+#ifndef HAVE_CPU_SET_T
+typedef unsigned long cpu_set_t;
+# define CPU_ZERO(cpuset) do { *(cpuset) = 0; } while(0)
+# define CPU_SET(cpu, cpuset) do { *(cpuset) |= (1UL << (cpu)); } while(0)
+#endif
+
static void run_on(int cpu)
{
static void run_on(int cpu)
{
+#if HAVE_SCHED_SETAFFINITY
cpu_set_t mask;
CPU_ZERO(&mask);
CPU_SET(cpu, &mask);
cpu_set_t mask;
CPU_ZERO(&mask);
CPU_SET(cpu, &mask);
+#if SCHED_SETAFFINITY_ARGS == 2
+ sched_setaffinity(0, &mask);
+#else
sched_setaffinity(0, sizeof(mask), &mask);
sched_setaffinity(0, sizeof(mask), &mask);
+#endif
+#endif /* HAVE_SCHED_SETAFFINITY */
}
/*
}
/*
@@
-490,7
+497,7
@@
long long get_microseconds(void)
#define DEFINE_PER_THREAD(type, name) \
struct { \
__typeof__(type) v \
#define DEFINE_PER_THREAD(type, name) \
struct { \
__typeof__(type) v \
- __attribute__((__aligned__(CACHE_LINE_SIZE))); \
+ __attribute__((__aligned__(CA
A_CA
CHE_LINE_SIZE))); \
} __per_thread_##name[NR_THREADS];
#define DECLARE_PER_THREAD(type, name) extern DEFINE_PER_THREAD(type, name)
} __per_thread_##name[NR_THREADS];
#define DECLARE_PER_THREAD(type, name) extern DEFINE_PER_THREAD(type, name)
@@
-524,7
+531,7
@@
long long get_microseconds(void)
#define DEFINE_PER_CPU(type, name) \
struct { \
__typeof__(type) v \
#define DEFINE_PER_CPU(type, name) \
struct { \
__typeof__(type) v \
- __attribute__((__aligned__(CACHE_LINE_SIZE))); \
+ __attribute__((__aligned__(CA
A_CA
CHE_LINE_SIZE))); \
} __per_cpu_##name[NR_CPUS]
#define DECLARE_PER_CPU(type, name) extern DEFINE_PER_CPU(type, name)
} __per_cpu_##name[NR_CPUS]
#define DECLARE_PER_CPU(type, name) extern DEFINE_PER_CPU(type, name)
@@
-638,10
+645,6
@@
static void smp_init(void)
#define LIST_POISON1 ((void *) 0x00100100)
#define LIST_POISON2 ((void *) 0x00200200)
#define LIST_POISON1 ((void *) 0x00100100)
#define LIST_POISON2 ((void *) 0x00200200)
-#define container_of(ptr, type, member) ({ \
- const typeof( ((type *)0)->member ) *__mptr = (ptr); \
- (type *)( (char *)__mptr - offsetof(type,member) );})
-
#if 0
/*
* Simple doubly linked list implementation.
#if 0
/*
* Simple doubly linked list implementation.
@@
-969,7
+972,7
@@
static inline void list_splice_tail_init(struct list_head *list,
* @member: the name of the list_struct within the struct.
*/
#define list_entry(ptr, type, member) \
* @member: the name of the list_struct within the struct.
*/
#define list_entry(ptr, type, member) \
- container_of(ptr, type, member)
+ c
aa_c
ontainer_of(ptr, type, member)
/**
* list_first_entry - get the first element from a list
/**
* list_first_entry - get the first element from a list
@@
-1271,7
+1274,7
@@
static inline void hlist_move_list(struct hlist_head *old,
old->first = NULL;
}
old->first = NULL;
}
-#define hlist_entry(ptr, type, member) container_of(ptr,type,member)
+#define hlist_entry(ptr, type, member) c
aa_c
ontainer_of(ptr,type,member)
#define hlist_for_each(pos, head) \
for (pos = (head)->first; pos && ({ prefetch(pos->next); 1; }); \
#define hlist_for_each(pos, head) \
for (pos = (head)->first; pos && ({ prefetch(pos->next); 1; }); \
This page took
0.023879 seconds
and
4
git commands to generate.