X-Git-Url: https://git.lttng.org/?p=urcu.git;a=blobdiff_plain;f=test_urcu_gc.c;h=12ea1811173f3520d36b420de05c09da17e280a2;hp=328b570fa4d2e5219e1adb5afb644c4fb3a827ad;hb=dd052bd346ddd497ed79f771ea1d5b3406bbcf07;hpb=a813abf8d937798b4a1b6e3246e844d75eef5756 diff --git a/test_urcu_gc.c b/test_urcu_gc.c index 328b570..12ea181 100644 --- a/test_urcu_gc.c +++ b/test_urcu_gc.c @@ -38,6 +38,9 @@ /* Make this big enough to include the POWER5+ L3 cacheline size of 256B */ #define CACHE_LINE_SIZE 4096 +/* hardcoded number of CPUs */ +#define NR_CPUS 16384 + #if defined(_syscall0) _syscall0(pid_t, gettid) #elif defined(__NR_gettid) @@ -70,7 +73,7 @@ static unsigned long wdelay; static struct test_array *test_rcu_pointer; -static int reclaim_batch = 1; +static unsigned int reclaim_batch = 1; struct reclaim_queue { void **queue; /* Beginning of queue */ @@ -98,6 +101,37 @@ static int verbose_mode; printf(fmt, args); \ } while (0) +static unsigned int cpu_affinities[NR_CPUS]; +static unsigned int next_aff = 0; +static int use_affinity = 0; + +pthread_mutex_t affinity_mutex = PTHREAD_MUTEX_INITIALIZER; + +static void set_affinity(void) +{ + cpu_set_t mask; + int cpu; + int ret; + + if (!use_affinity) + return; + + ret = pthread_mutex_lock(&affinity_mutex); + if (ret) { + perror("Error in pthread mutex lock"); + exit(-1); + } + cpu = cpu_affinities[next_aff++]; + ret = pthread_mutex_unlock(&affinity_mutex); + if (ret) { + perror("Error in pthread mutex unlock"); + exit(-1); + } + CPU_ZERO(&mask); + CPU_SET(cpu, &mask); + sched_setaffinity(0, sizeof(mask), &mask); +} + /* * returns 0 if test should end. */ @@ -151,6 +185,8 @@ void *thr_reader(void *_count) printf_verbose("thread_begin %s, thread id : %lx, tid %lu\n", "reader", pthread_self(), (unsigned long)gettid()); + set_affinity(); + rcu_register_thread(); while (!test_go) @@ -181,20 +217,11 @@ void *thr_reader(void *_count) } -/* Using per-thread queue */ -static void rcu_gc_reclaim(unsigned long wtidx, void *old) +static void rcu_gc_clear_queue(unsigned long wtidx) { void **p; - /* Queue pointer */ - *pending_reclaims[wtidx].head = old; - pending_reclaims[wtidx].head++; - - if (likely(pending_reclaims[wtidx].head - pending_reclaims[wtidx].queue - < reclaim_batch)) - return; - - /* If queue is full, wait for Q.S and empty queue */ + /* Wait for Q.S and empty queue */ synchronize_rcu(); for (p = pending_reclaims[wtidx].queue; @@ -207,45 +234,45 @@ static void rcu_gc_reclaim(unsigned long wtidx, void *old) pending_reclaims[wtidx].head = pending_reclaims[wtidx].queue; } -static void rcu_gc_cleanup(unsigned long wtidx) +/* Using per-thread queue */ +static void rcu_gc_reclaim(unsigned long wtidx, void *old) { - void **p; + /* Queue pointer */ + *pending_reclaims[wtidx].head = old; + pending_reclaims[wtidx].head++; - /* Wait for Q.S and empty queue */ - synchronize_rcu(); + if (likely(pending_reclaims[wtidx].head - pending_reclaims[wtidx].queue + < reclaim_batch)) + return; - for (p = pending_reclaims[wtidx].queue; - p < pending_reclaims[wtidx].head; p++) { - /* poison */ - if (*p) - ((struct test_array *)*p)->a = 0; - free(*p); - } - pending_reclaims[wtidx].head = pending_reclaims[wtidx].queue; + rcu_gc_clear_queue(wtidx); } void *thr_writer(void *data) { unsigned long wtidx = (unsigned long)data; +#ifdef TEST_LOCAL_GC + struct test_array *old = NULL; +#else struct test_array *new, *old; +#endif printf_verbose("thread_begin %s, thread id : %lx, tid %lu\n", "writer", pthread_self(), (unsigned long)gettid()); + set_affinity(); + while (!test_go) { } smp_mb(); for (;;) { +#ifndef TEST_LOCAL_GC new = malloc(sizeof(*new)); - rcu_copy_mutex_lock(); - old = test_rcu_pointer; - if (old) - assert(old->a == 8); new->a = 8; old = rcu_xchg_pointer(&test_rcu_pointer, new); - rcu_copy_mutex_unlock(); +#endif rcu_gc_reclaim(wtidx, old); nr_writes++; if (unlikely(!test_duration_write())) @@ -273,8 +300,6 @@ void show_usage(int argc, char **argv) printf("\n"); } -cpu_set_t affinity; - int main(int argc, char **argv) { int err; @@ -283,7 +308,6 @@ int main(int argc, char **argv) unsigned long long *count_reader; unsigned long long tot_reads = 0, tot_writes = 0; int i, a; - int use_affinity = 0; if (argc < 4) { show_usage(argc, argv); @@ -308,8 +332,6 @@ int main(int argc, char **argv) return -1; } - CPU_ZERO(&affinity); - for (i = 4; i < argc; i++) { if (argv[i][0] != '-') continue; @@ -328,7 +350,7 @@ int main(int argc, char **argv) return -1; } a = atoi(argv[++i]); - CPU_SET(a, &affinity); + cpu_affinities[next_aff++] = a; use_affinity = 1; printf_verbose("Adding CPU %d affinity\n", a); break; @@ -366,12 +388,6 @@ int main(int argc, char **argv) printf_verbose("thread %-6s, thread id : %lx, tid %lu\n", "main", pthread_self(), (unsigned long)gettid()); - if (use_affinity - && sched_setaffinity(0, sizeof(affinity), &affinity) < 0) { - perror("sched_setaffinity"); - exit(-1); - } - tid_reader = malloc(sizeof(*tid_reader) * nr_readers); tid_writer = malloc(sizeof(*tid_writer) * nr_writers); count_reader = malloc(sizeof(*count_reader) * nr_readers); @@ -388,6 +404,8 @@ int main(int argc, char **argv) for (i = 0; i < nr_writers; i++) pending_reclaims[i].head = pending_reclaims[i].queue; + next_aff = 0; + for (i = 0; i < nr_readers; i++) { err = pthread_create(&tid_reader[i], NULL, thr_reader, &count_reader[i]); @@ -420,17 +438,18 @@ int main(int argc, char **argv) if (err != 0) exit(1); tot_writes += tot_nr_writes[i]; - rcu_gc_cleanup(i); + rcu_gc_clear_queue(i); } printf_verbose("total number of reads : %llu, writes %llu\n", tot_reads, tot_writes); printf("SUMMARY %-25s testdur %4lu nr_readers %3u rdur %6lu " "nr_writers %3u " - "wdelay %6lu nr_reads %12llu nr_writes %12llu nr_ops %12llu\n", + "wdelay %6lu nr_reads %12llu nr_writes %12llu nr_ops %12llu " + "batch %u\n", argv[0], duration, nr_readers, rduration, nr_writers, wdelay, tot_reads, tot_writes, - tot_reads + tot_writes); + tot_reads + tot_writes, reclaim_batch); free(tid_reader); free(tid_writer); free(count_reader);