Fix: x86 and s390: uatomic __hp() macro clang support
[urcu.git] / include / urcu / uatomic / s390.h
... / ...
CommitLineData
1#ifndef _URCU_UATOMIC_ARCH_S390_H
2#define _URCU_UATOMIC_ARCH_S390_H
3
4/*
5 * Atomic exchange operations for the S390 architecture. Based on information
6 * taken from the Principles of Operation Appendix A "Conditional Swapping
7 * Instructions (CS, CDS)".
8 *
9 * Copyright (c) 2009 Novell, Inc.
10 * Author: Jan Blunck <jblunck@suse.de>
11 * Copyright (c) 2009 Mathieu Desnoyers <mathieu.desnoyers@efficios.com>
12 *
13 * Permission is hereby granted, free of charge, to any person obtaining a copy
14 * of this software and associated documentation files (the "Software"), to
15 * deal in the Software without restriction, including without limitation the
16 * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
17 * sell copies of the Software, and to permit persons to whom the Software is
18 * furnished to do so, subject to the following conditions:
19 *
20 * The above copyright notice and this permission notice shall be included in
21 * all copies or substantial portions of the Software.
22 *
23 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
24 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
25 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
26 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
27 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
28 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
29 * IN THE SOFTWARE.
30 */
31
32#include <urcu/compiler.h>
33#include <urcu/system.h>
34
35#ifdef __cplusplus
36extern "C" {
37#endif
38
39#if __GNUC__ > 3 || (__GNUC__ == 3 && __GNUC_MINOR__ > 2)
40#define COMPILER_HAVE_SHORT_MEM_OPERAND
41#endif
42
43/*
44 * MEMOP assembler operand rules:
45 * - op refer to MEMOP_IN operand
46 * - MEMOP_IN can expand to more than a single operand. Use it at the end of
47 * operand list only.
48 */
49
50#ifdef COMPILER_HAVE_SHORT_MEM_OPERAND
51
52#define MEMOP_OUT(addr) "=Q" (*(addr))
53#define MEMOP_IN(addr) "Q" (*(addr))
54#define MEMOP_REF(op) #op /* op refer to MEMOP_IN operand */
55
56#else /* !COMPILER_HAVE_SHORT_MEM_OPERAND */
57
58#define MEMOP_OUT(addr) "=m" (*(addr))
59#define MEMOP_IN(addr) "a" (addr), "m" (*(addr))
60#define MEMOP_REF(op) "0(" #op ")" /* op refer to MEMOP_IN operand */
61
62#endif /* !COMPILER_HAVE_SHORT_MEM_OPERAND */
63
64/*
65 * The __hp() macro casts the void pointer @x to a pointer to a structure
66 * containing an array of char of the specified size. This allows passing the
67 * @addr arguments of the following inline functions as "m" and "+m" operands
68 * to the assembly. The @size parameter should be a constant to support
69 * compilers such as clang which do not support VLA.
70 */
71
72#define __hp(size, x) ((struct { char v[size]; } *)(x))
73
74/* xchg */
75
76static inline __attribute__((always_inline))
77unsigned long _uatomic_exchange(volatile void *addr, unsigned long val, int len)
78{
79 switch (len) {
80 case 4:
81 {
82 unsigned int old_val;
83
84 __asm__ __volatile__(
85 "0: cs %0,%2," MEMOP_REF(%3) "\n"
86 " brc 4,0b\n"
87 : "=&r" (old_val), MEMOP_OUT (__hp(4, addr))
88 : "r" (val), MEMOP_IN (__hp(4, addr))
89 : "memory", "cc");
90 return old_val;
91 }
92#if (CAA_BITS_PER_LONG == 64)
93 case 8:
94 {
95 unsigned long old_val;
96
97 __asm__ __volatile__(
98 "0: csg %0,%2," MEMOP_REF(%3) "\n"
99 " brc 4,0b\n"
100 : "=&r" (old_val), MEMOP_OUT (__hp(8, addr))
101 : "r" (val), MEMOP_IN (__hp(8, addr))
102 : "memory", "cc");
103 return old_val;
104 }
105#endif
106 default:
107 __asm__ __volatile__(".long 0xd00d00");
108 }
109
110 return 0;
111}
112
113#define uatomic_xchg(addr, v) \
114 (__typeof__(*(addr))) _uatomic_exchange((addr), \
115 caa_cast_long_keep_sign(v), \
116 sizeof(*(addr)))
117
118/* cmpxchg */
119
120static inline __attribute__((always_inline))
121unsigned long _uatomic_cmpxchg(void *addr, unsigned long old,
122 unsigned long _new, int len)
123{
124 switch (len) {
125 case 4:
126 {
127 unsigned int old_val = (unsigned int)old;
128
129 __asm__ __volatile__(
130 " cs %0,%2," MEMOP_REF(%3) "\n"
131 : "+r" (old_val), MEMOP_OUT (__hp(4, addr))
132 : "r" (_new), MEMOP_IN (__hp(4, addr))
133 : "memory", "cc");
134 return old_val;
135 }
136#if (CAA_BITS_PER_LONG == 64)
137 case 8:
138 {
139 __asm__ __volatile__(
140 " csg %0,%2," MEMOP_REF(%3) "\n"
141 : "+r" (old), MEMOP_OUT (__hp(8, addr))
142 : "r" (_new), MEMOP_IN (__hp(8, addr))
143 : "memory", "cc");
144 return old;
145 }
146#endif
147 default:
148 __asm__ __volatile__(".long 0xd00d00");
149 }
150
151 return 0;
152}
153
154#define uatomic_cmpxchg(addr, old, _new) \
155 (__typeof__(*(addr))) _uatomic_cmpxchg((addr), \
156 caa_cast_long_keep_sign(old), \
157 caa_cast_long_keep_sign(_new),\
158 sizeof(*(addr)))
159
160#ifdef __cplusplus
161}
162#endif
163
164#include <urcu/uatomic/generic.h>
165
166#endif /* _URCU_UATOMIC_ARCH_S390_H */
This page took 0.022737 seconds and 4 git commands to generate.