X-Git-Url: https://git.lttng.org/?p=urcu.git;a=blobdiff_plain;f=tests%2Fapi_x86.h;h=fe00a3599ab9a9e657e9daa6bb91cad8bc4ffac7;hp=d64eb8d4a6a2d89a17c4f4636697d303d778811e;hb=5481ddb381061bda64aebc039900d21cac6a6caf;hpb=3d7be7ca4e2451381a8fc8a00bd6d7dcbf3b678f diff --git a/tests/api_x86.h b/tests/api_x86.h index d64eb8d..fe00a35 100644 --- a/tests/api_x86.h +++ b/tests/api_x86.h @@ -1,7 +1,10 @@ /* MECHANICALLY GENERATED, DO NOT EDIT!!! */ +#ifndef _INCLUDE_API_H #define _INCLUDE_API_H +#include "../config.h" + /* * common.h: Common Linux kernel-isms. * @@ -25,6 +28,8 @@ * to redistribute under later versions of GPL might not be available. */ +#include + #ifndef __always_inline #define __always_inline inline #endif @@ -71,12 +76,14 @@ * Machine parameters. */ -#define CACHE_LINE_SIZE 64 +/* #define CACHE_LINE_SIZE 64 */ #define ____cacheline_internodealigned_in_smp \ __attribute__((__aligned__(1 << 6))) #define LOCK_PREFIX "lock ; " +#if 0 /* duplicate with arch_atomic.h */ + /* * Atomic data structure, initialization, and access. */ @@ -305,10 +312,12 @@ __asm__ __volatile__(LOCK_PREFIX "orl %0,%1" \ : : "r" (mask),"m" (*(addr)) : "memory") /* Atomic operations are already serializing on x86 */ -#define smp_mb__before_atomic_dec() barrier() -#define smp_mb__after_atomic_dec() barrier() -#define smp_mb__before_atomic_inc() barrier() -#define smp_mb__after_atomic_inc() barrier() +#define smp_mb__before_atomic_dec() cmm_barrier() +#define smp_mb__after_atomic_dec() cmm_barrier() +#define smp_mb__before_atomic_inc() cmm_barrier() +#define smp_mb__after_atomic_inc() cmm_barrier() + +#endif //0 /* * api_pthreads.h: API mapping to pthreads environment. @@ -343,19 +352,12 @@ __asm__ __volatile__(LOCK_PREFIX "orl %0,%1" \ #include /* #include "atomic.h" */ -/* - * Compiler magic. - */ -#define container_of(ptr, type, member) ({ \ - const typeof( ((type *)0)->member ) *__mptr = (ptr); \ - (type *)( (char *)__mptr - offsetof(type,member) );}) - /* * Default machine parameters. */ #ifndef CACHE_LINE_SIZE -#define CACHE_LINE_SIZE 128 +/* #define CACHE_LINE_SIZE 128 */ #endif /* #ifndef CACHE_LINE_SIZE */ /* @@ -515,13 +517,25 @@ static void wait_all_threads(void) } } +#ifndef HAVE_CPU_SET_T +typedef unsigned long cpu_set_t; +# define CPU_ZERO(cpuset) do { *(cpuset) = 0; } while(0) +# define CPU_SET(cpu, cpuset) do { *(cpuset) |= (1UL << (cpu)); } while(0) +#endif + static void run_on(int cpu) { +#if HAVE_SCHED_SETAFFINITY cpu_set_t mask; CPU_ZERO(&mask); CPU_SET(cpu, &mask); +#if SCHED_SETAFFINITY_ARGS == 2 + sched_setaffinity(0, &mask); +#else sched_setaffinity(0, sizeof(mask), &mask); +#endif +#endif /* HAVE_SCHED_SETAFFINITY */ } /* @@ -692,9 +706,7 @@ static void smp_init(void) #define LIST_POISON1 ((void *) 0x00100100) #define LIST_POISON2 ((void *) 0x00200200) -#define container_of(ptr, type, member) ({ \ - const typeof( ((type *)0)->member ) *__mptr = (ptr); \ - (type *)( (char *)__mptr - offsetof(type,member) );}) +#if 0 /* * Simple doubly linked list implementation. @@ -1220,6 +1232,8 @@ static inline void list_splice_tail_init(struct list_head *list, &pos->member != (head); \ pos = n, n = list_entry(n->member.prev, typeof(*n), member)) +#endif //0 + /* * Double linked lists with a single pointer list head. * Mostly useful for hash tables where the two pointer list head is @@ -1383,3 +1397,5 @@ static inline void hlist_move_list(struct hlist_head *old, pos = n) #endif + +#endif