Update return value of "set" operations
[urcu.git] / urcu / uatomic / generic.h
index 337fe4098bf8337649f9a6bb52af4f9359453091..9e2e78086c220b7286c508403069d5aa3b27e246 100644 (file)
@@ -29,7 +29,7 @@ extern "C" {
 #endif
 
 #ifndef uatomic_set
-#define uatomic_set(addr, v)   CMM_STORE_SHARED(*(addr), (v))
+#define uatomic_set(addr, v)   ((void) CMM_STORE_SHARED(*(addr), (v)))
 #endif
 
 #ifndef uatomic_read
@@ -81,9 +81,10 @@ unsigned long _uatomic_cmpxchg(void *addr, unsigned long old,
 }
 
 
-#define uatomic_cmpxchg(addr, old, _new)                                   \
-       ((__typeof__(*(addr))) _uatomic_cmpxchg((addr), (unsigned long)(old),\
-                                               (unsigned long)(_new),      \
+#define uatomic_cmpxchg(addr, old, _new)                                     \
+       ((__typeof__(*(addr))) _uatomic_cmpxchg((addr),                       \
+                                               caa_cast_long_keep_sign(old), \
+                                               caa_cast_long_keep_sign(_new),\
                                                sizeof(*(addr))))
 
 
@@ -98,26 +99,32 @@ void _uatomic_and(void *addr, unsigned long val,
 #ifdef UATOMIC_HAS_ATOMIC_BYTE
        case 1:
                __sync_and_and_fetch_1(addr, val);
+               return;
 #endif
 #ifdef UATOMIC_HAS_ATOMIC_SHORT
        case 2:
                __sync_and_and_fetch_2(addr, val);
+               return;
 #endif
        case 4:
                __sync_and_and_fetch_4(addr, val);
+               return;
 #if (CAA_BITS_PER_LONG == 64)
        case 8:
                __sync_and_and_fetch_8(addr, val);
+               return;
 #endif
        }
        _uatomic_link_error();
-       return 0;
 }
 
 #define uatomic_and(addr, v)                   \
        (_uatomic_and((addr),                   \
-                     (unsigned long)(v),       \
-                     sizeof(*(addr))))
+               caa_cast_long_keep_sign(v),     \
+               sizeof(*(addr))))
+#define cmm_smp_mb__before_uatomic_and()       cmm_barrier()
+#define cmm_smp_mb__after_uatomic_and()                cmm_barrier()
+
 #endif
 
 /* uatomic_or */
@@ -131,28 +138,36 @@ void _uatomic_or(void *addr, unsigned long val,
 #ifdef UATOMIC_HAS_ATOMIC_BYTE
        case 1:
                __sync_or_and_fetch_1(addr, val);
+               return;
 #endif
 #ifdef UATOMIC_HAS_ATOMIC_SHORT
        case 2:
                __sync_or_and_fetch_2(addr, val);
+               return;
 #endif
        case 4:
                __sync_or_and_fetch_4(addr, val);
+               return;
 #if (CAA_BITS_PER_LONG == 64)
        case 8:
                __sync_or_and_fetch_8(addr, val);
+               return;
 #endif
        }
        _uatomic_link_error();
-       return 0;
+       return;
 }
 
 #define uatomic_or(addr, v)                    \
        (_uatomic_or((addr),                    \
-                    (unsigned long)(v),        \
-                    sizeof(*(addr))))
+               caa_cast_long_keep_sign(v),     \
+               sizeof(*(addr))))
+#define cmm_smp_mb__before_uatomic_or()                cmm_barrier()
+#define cmm_smp_mb__after_uatomic_or()         cmm_barrier()
+
 #endif
 
+
 /* uatomic_add_return */
 
 #ifndef uatomic_add_return
@@ -181,10 +196,10 @@ unsigned long _uatomic_add_return(void *addr, unsigned long val,
 }
 
 
-#define uatomic_add_return(addr, v)                                    \
-       ((__typeof__(*(addr))) _uatomic_add_return((addr),              \
-                                                 (unsigned long)(v),   \
-                                                 sizeof(*(addr))))
+#define uatomic_add_return(addr, v)                                        \
+       ((__typeof__(*(addr))) _uatomic_add_return((addr),                  \
+                                               caa_cast_long_keep_sign(v), \
+                                               sizeof(*(addr))))
 #endif /* #ifndef uatomic_add_return */
 
 #ifndef uatomic_xchg
@@ -246,7 +261,8 @@ unsigned long _uatomic_exchange(void *addr, unsigned long val, int len)
 }
 
 #define uatomic_xchg(addr, v)                                              \
-       ((__typeof__(*(addr))) _uatomic_exchange((addr), (unsigned long)(v), \
+       ((__typeof__(*(addr))) _uatomic_exchange((addr),                    \
+                                               caa_cast_long_keep_sign(v), \
                                                sizeof(*(addr))))
 #endif /* #ifndef uatomic_xchg */
 
@@ -269,6 +285,8 @@ void _uatomic_and(void *addr, unsigned long val, int len)
                        old = oldt;
                        oldt = _uatomic_cmpxchg(addr, old, old & val, 1);
                } while (oldt != old);
+
+               return;
        }
 #endif
 #ifdef UATOMIC_HAS_ATOMIC_SHORT
@@ -292,6 +310,8 @@ void _uatomic_and(void *addr, unsigned long val, int len)
                        old = oldt;
                        oldt = _uatomic_cmpxchg(addr, old, old & val, 4);
                } while (oldt != old);
+
+               return;
        }
 #if (CAA_BITS_PER_LONG == 64)
        case 8:
@@ -303,17 +323,21 @@ void _uatomic_and(void *addr, unsigned long val, int len)
                        old = oldt;
                        oldt = _uatomic_cmpxchg(addr, old, old & val, 8);
                } while (oldt != old);
+
+               return;
        }
 #endif
        }
        _uatomic_link_error();
-       return 0;
 }
 
-#define uatomic_and(addr, v)           \
-       (uatomic_and((addr),            \
-                   (unsigned long)(v), \
-                   sizeof(*(addr))))
+#define uatomic_and(addr, v)                   \
+       (_uatomic_and((addr),                   \
+               caa_cast_long_keep_sign(v),     \
+               sizeof(*(addr))))
+#define cmm_smp_mb__before_uatomic_and()       cmm_barrier()
+#define cmm_smp_mb__after_uatomic_and()                cmm_barrier()
+
 #endif /* #ifndef uatomic_and */
 
 #ifndef uatomic_or
@@ -333,6 +357,8 @@ void _uatomic_or(void *addr, unsigned long val, int len)
                        old = oldt;
                        oldt = _uatomic_cmpxchg(addr, old, old | val, 1);
                } while (oldt != old);
+
+               return;
        }
 #endif
 #ifdef UATOMIC_HAS_ATOMIC_SHORT
@@ -345,6 +371,8 @@ void _uatomic_or(void *addr, unsigned long val, int len)
                        old = oldt;
                        oldt = _uatomic_cmpxchg(addr, old, old | val, 2);
                } while (oldt != old);
+
+               return;
        }
 #endif
        case 4:
@@ -356,6 +384,8 @@ void _uatomic_or(void *addr, unsigned long val, int len)
                        old = oldt;
                        oldt = _uatomic_cmpxchg(addr, old, old | val, 4);
                } while (oldt != old);
+
+               return;
        }
 #if (CAA_BITS_PER_LONG == 64)
        case 8:
@@ -367,17 +397,21 @@ void _uatomic_or(void *addr, unsigned long val, int len)
                        old = oldt;
                        oldt = _uatomic_cmpxchg(addr, old, old | val, 8);
                } while (oldt != old);
+
+               return;
        }
 #endif
        }
        _uatomic_link_error();
-       return 0;
 }
 
-#define uatomic_or(addr, v)            \
-       (uatomic_or((addr),             \
-                   (unsigned long)(v), \
-                   sizeof(*(addr))))
+#define uatomic_or(addr, v)                    \
+       (_uatomic_or((addr),                    \
+               caa_cast_long_keep_sign(v),     \
+               sizeof(*(addr))))
+#define cmm_smp_mb__before_uatomic_or()                cmm_barrier()
+#define cmm_smp_mb__after_uatomic_or()         cmm_barrier()
+
 #endif /* #ifndef uatomic_or */
 
 #ifndef uatomic_add_return
@@ -450,10 +484,10 @@ unsigned long _uatomic_add_return(void *addr, unsigned long val, int len)
        return 0;
 }
 
-#define uatomic_add_return(addr, v)                                    \
-       ((__typeof__(*(addr))) _uatomic_add_return((addr),              \
-                                                 (unsigned long)(v),   \
-                                                 sizeof(*(addr))))
+#define uatomic_add_return(addr, v)                                        \
+       ((__typeof__(*(addr))) _uatomic_add_return((addr),                  \
+                                               caa_cast_long_keep_sign(v), \
+                                               sizeof(*(addr))))
 #endif /* #ifndef uatomic_add_return */
 
 #ifndef uatomic_xchg
@@ -527,7 +561,8 @@ unsigned long _uatomic_exchange(void *addr, unsigned long val, int len)
 }
 
 #define uatomic_xchg(addr, v)                                              \
-       ((__typeof__(*(addr))) _uatomic_exchange((addr), (unsigned long)(v), \
+       ((__typeof__(*(addr))) _uatomic_exchange((addr),                    \
+                                               caa_cast_long_keep_sign(v), \
                                                sizeof(*(addr))))
 #endif /* #ifndef uatomic_xchg */
 
@@ -537,17 +572,27 @@ unsigned long _uatomic_exchange(void *addr, unsigned long val, int len)
 
 #ifndef uatomic_add
 #define uatomic_add(addr, v)           (void)uatomic_add_return((addr), (v))
+#define cmm_smp_mb__before_uatomic_add()       cmm_barrier()
+#define cmm_smp_mb__after_uatomic_add()                cmm_barrier()
 #endif
 
-#define uatomic_sub_return(addr, v)    uatomic_add_return((addr), -(v))
-#define uatomic_sub(addr, v)           uatomic_add((addr), -(v))
+#define uatomic_sub_return(addr, v)    \
+       uatomic_add_return((addr), -(caa_cast_long_keep_sign(v)))
+#define uatomic_sub(addr, v)           \
+       uatomic_add((addr), -(caa_cast_long_keep_sign(v)))
+#define cmm_smp_mb__before_uatomic_sub()       cmm_smp_mb__before_uatomic_add()
+#define cmm_smp_mb__after_uatomic_sub()                cmm_smp_mb__after_uatomic_add()
 
 #ifndef uatomic_inc
 #define uatomic_inc(addr)              uatomic_add((addr), 1)
+#define cmm_smp_mb__before_uatomic_inc()       cmm_smp_mb__before_uatomic_add()
+#define cmm_smp_mb__after_uatomic_inc()                cmm_smp_mb__after_uatomic_add()
 #endif
 
 #ifndef uatomic_dec
 #define uatomic_dec(addr)              uatomic_add((addr), -1)
+#define cmm_smp_mb__before_uatomic_dec()       cmm_smp_mb__before_uatomic_add()
+#define cmm_smp_mb__after_uatomic_dec()                cmm_smp_mb__after_uatomic_add()
 #endif
 
 #ifdef __cplusplus
This page took 0.028087 seconds and 4 git commands to generate.