Add CMM memory model
authorOlivier Dion <odion@efficios.com>
Wed, 29 Mar 2023 18:44:43 +0000 (14:44 -0400)
committerMathieu Desnoyers <mathieu.desnoyers@efficios.com>
Mon, 14 Aug 2023 19:39:04 +0000 (15:39 -0400)
Introducing the CMM memory model with the following new primitives:

  - uatomic_load(addr, memory_order)

  - uatomic_store(addr, value, memory_order)
  - uatomic_and_mo(addr, mask, memory_order)
  - uatomic_or_mo(addr, mask, memory_order)
  - uatomic_add_mo(addr, value, memory_order)
  - uatomic_sub_mo(addr, value, memory_order)
  - uatomic_inc_mo(addr, memory_order)
  - uatomic_dec_mo(addr, memory_order)

  - uatomic_add_return_mo(addr, value, memory_order)
  - uatomic_sub_return_mo(addr, value, memory_order)

  - uatomic_xchg_mo(addr, value, memory_order)

  - uatomic_cmpxchg_mo(addr, old, new,
                       memory_order_success,
                       memory_order_failure)

The CMM memory model reflects the C11 memory model with an additional
CMM_SEQ_CST_FENCE memory order. The memory order can be selected through
the enum cmm_memorder.

* With Atomic Builtins

If configured with atomic builtins, the correspondence between the CMM
memory model and the C11 memory model is a one to one at the exception
of the CMM_SEQ_CST_FENCE memory order which implies the memory order
CMM_SEQ_CST and a thread fence after the operation.

* Without Atomic Builtins

However, if not configured with atomic builtins, the following stipulate
the memory model.

For load operations with uatomic_load(), the memory orders CMM_RELAXED,
CMM_CONSUME, CMM_ACQUIRE, CMM_SEQ_CST and CMM_SEQ_CST_FENCE are
allowed. A barrier may be inserted before and after the load from memory
depending on the memory order:

  - CMM_RELAXED: No barrier
  - CMM_CONSUME: Memory barrier after read
  - CMM_ACQUIRE: Memory barrier after read
  - CMM_SEQ_CST: Memory barriers before and after read
  - CMM_SEQ_CST_FENCE: Memory barriers before and after read

For store operations with uatomic_store(), the memory orders
CMM_RELAXED, CMM_RELEASE, CMM_SEQ_CST and CMM_SEQ_CST_FENCE are
allowed. A barrier may be inserted before and after the store to memory
depending on the memory order:

  - CMM_RELAXED: No barrier
  - CMM_RELEASE: Memory barrier before operation
  - CMM_SEQ_CST: Memory barriers before and after operation
  - CMM_SEQ_CST_FENCE: Memory barriers before and after operation

For load/store operations with uatomic_and_mo(), uatomic_or_mo(),
uatomic_add_mo(), uatomic_sub_mo(), uatomic_inc_mo(), uatomic_dec_mo(),
uatomic_add_return_mo() and uatomic_sub_return_mo(), all memory orders
are allowed. A barrier may be inserted before and after the operation
depending on the memory order:

  - CMM_RELAXED: No barrier
  - CMM_ACQUIRE: Memory barrier after operation
  - CMM_CONSUME: Memory barrier after operation
  - CMM_RELEASE: Memory barrier before operation
  - CMM_ACQ_REL: Memory barriers before and after operation
  - CMM_SEQ_CST: Memory barriers before and after operation
  - CMM_SEQ_CST_FENCE: Memory barriers before and after operation

For the exchange operation uatomic_xchg_mo(), any memory order is
valid. A barrier may be inserted before and after the exchange to memory
depending on the memory order:

  - CMM_RELAXED: No barrier
  - CMM_ACQUIRE: Memory barrier after operation
  - CMM_CONSUME: Memory barrier after operation
  - CMM_RELEASE: Memory barrier before operation
  - CMM_ACQ_REL: Memory barriers before and after operation
  - CMM_SEQ_CST: Memory barriers before and after operation
  - CMM_SEQ_CST_FENCE: Memory barriers before and after operation

For the compare exchange operation uatomic_cmpxchg_mo(), the success
memory order can be anything while the failure memory order cannot be
CMM_RELEASE nor CMM_ACQ_REL and cannot be stronger than the success
memory order. A barrier may be inserted before and after the store to
memory depending on the memory orders:

 Success memory order:

  - CMM_RELAXED: No barrier
  - CMM_ACQUIRE: Memory barrier after operation
  - CMM_CONSUME: Memory barrier after operation
  - CMM_RELEASE: Memory barrier before operation
  - CMM_ACQ_REL: Memory barriers before and after operation
  - CMM_SEQ_CST: Memory barriers before and after operation
  - CMM_SEQ_CST_FENCE: Memory barriers before and after operation

  Barriers after the operations are only emitted if the compare exchange
  succeed.

 Failure memory order:
  - CMM_RELAXED: No barrier
  - CMM_ACQUIRE: Memory barrier after operation
  - CMM_CONSUME: Memory barrier after operation
  - CMM_SEQ_CST: Memory barriers before and after operation
  - CMM_SEQ_CST_FENCE: Memory barriers before and after operation

  Barriers after the operations are only emitted if the compare exchange
  failed.  Barriers before the operation are never emitted by this
  memory order.

Change-Id: I213ba19c84e82a63083f00143a3142ffbdab1d52
Co-authored-by: Mathieu Desnoyers <mathieu.desnoyers@efficios.com>
Signed-off-by: Olivier Dion <odion@efficios.com>
Signed-off-by: Mathieu Desnoyers <mathieu.desnoyers@efficios.com>
doc/uatomic-api.md
include/Makefile.am
include/urcu/compiler.h
include/urcu/static/pointer.h
include/urcu/system.h
include/urcu/uatomic.h
include/urcu/uatomic/builtins-generic.h [new file with mode: 0644]
include/urcu/uatomic/builtins.h [new file with mode: 0644]
include/urcu/uatomic/generic.h
src/urcu-pointer.c

index 84b97166cef4c2ec81cdf61734eb6f2d1d5fe6b1..642170857a1e2a4d260267c10b5dcf6892ebb5b7 100644 (file)
@@ -58,7 +58,8 @@ An atomic read-modify-write operation that performs this
 sequence of operations atomically: check if `addr` contains `old`.
 If true, then replace the content of `addr` by `new`. Return the
 value previously contained by `addr`. This function implies a full
-memory barrier before and after the atomic operation.
+memory barrier before and after the atomic operation on success.
+On failure, no memory order is guaranteed.
 
 
 ```c
index 859bccd6d6f3eb3e63bb59a954c44ecf1cfd7182..58aa73668d51b83af38e8cfbec5e37208c7896c4 100644 (file)
@@ -67,6 +67,8 @@ nobase_include_HEADERS = \
        urcu/uatomic/alpha.h \
        urcu/uatomic_arch.h \
        urcu/uatomic/arm.h \
+       urcu/uatomic/builtins.h \
+       urcu/uatomic/builtins-generic.h \
        urcu/uatomic/gcc.h \
        urcu/uatomic/generic.h \
        urcu/uatomic.h \
index 48211293417131c6ff73d0d107de84c4284d9589..887d9be1d3f1e8ee492d225412ba884a1c79428b 100644 (file)
                                + __GNUC_PATCHLEVEL__)
 #endif
 
+#ifdef __cplusplus
+#define caa_unqual_scalar_typeof(x)                                    \
+       std::remove_cv<std::remove_reference<decltype(x)>::type>::type
+#else
+#define caa_scalar_type_to_expr(type)                                  \
+       unsigned type: (unsigned type)0,                                \
+       signed type: (signed type)0
+
+/*
+ * Use C11 _Generic to express unqualified type from expression. This removes
+ * volatile qualifier from expression type.
+ */
+#define caa_unqual_scalar_typeof(x)                                    \
+       __typeof__(                                                     \
+               _Generic((x),                                           \
+                       char: (char)0,                                  \
+                       caa_scalar_type_to_expr(char),                  \
+                       caa_scalar_type_to_expr(short),         \
+                       caa_scalar_type_to_expr(int),                   \
+                       caa_scalar_type_to_expr(long),                  \
+                       caa_scalar_type_to_expr(long long),             \
+                       default: (x)                                    \
+               )                                                       \
+       )
+#endif
+
+/*
+ * Allow user to manually define CMM_SANITIZE_THREAD if their toolchain is not
+ * supported by this check.
+ */
+#ifndef CMM_SANITIZE_THREAD
+# if defined(__GNUC__) && defined(__SANITIZE_THREAD__)
+#  define CMM_SANITIZE_THREAD
+# elif defined(__clang__) && defined(__has_feature)
+#  if __has_feature(thread_sanitizer)
+#   define CMM_SANITIZE_THREAD
+#  endif
+# endif
+#endif /* !CMM_SANITIZE_THREAD */
+
+/*
+ * Helper to add the volatile qualifier to a pointer.
+ */
+#if defined __cplusplus
+template <typename T>
+volatile T cmm_cast_volatile(T t)
+{
+    return static_cast<volatile T>(t);
+}
+#else
+#  define cmm_cast_volatile(ptr)                       \
+       __extension__                                   \
+       ({                                              \
+               (volatile __typeof__(ptr))(ptr);        \
+       })
+#endif
+
 #endif /* _URCU_COMPILER_H */
index 055a9b884524581a58610a7f9f40dbe0df780a76..9dc0d3eaa5f4c6177fc503449153cd08f78267ec 100644 (file)
@@ -82,23 +82,8 @@ extern "C" {
  * -Wincompatible-pointer-types errors.  Using the statement expression
  * makes it an rvalue and gets rid of the const-ness.
  */
-#ifdef __URCU_DEREFERENCE_USE_ATOMIC_CONSUME
-# define _rcu_dereference(p) __extension__ ({                                          \
-                               __typeof__(__extension__ ({                             \
-                                       __typeof__(p) __attribute__((unused)) _________p0 = { 0 }; \
-                                       _________p0;                                    \
-                               })) _________p1;                                        \
-                               __atomic_load(&(p), &_________p1, __ATOMIC_CONSUME);    \
-                               (_________p1);                                          \
-                       })
-#else
-# define _rcu_dereference(p) __extension__ ({                                          \
-                               __typeof__(p) _________p1 = CMM_LOAD_SHARED(p);         \
-                               cmm_smp_read_barrier_depends();                         \
-                               (_________p1);                                          \
-                       })
-#endif
-
+# define _rcu_dereference(p)                   \
+       uatomic_load(&(p), CMM_CONSUME)
 /**
  * _rcu_cmpxchg_pointer - same as rcu_assign_pointer, but tests if the pointer
  * is as expected by "old". If succeeds, returns the previous pointer to the
@@ -106,7 +91,7 @@ extern "C" {
  * using synchronize_rcu(). If fails (unexpected value), returns old (which
  * should not be freed !).
  *
- * uatomic_cmpxchg() acts as both release and acquire barriers.
+ * uatomic_cmpxchg() acts as both release and acquire barriers on success.
  *
  * This macro is less than 10 lines long.  The intent is that this macro
  * meets the 10-line criterion in LGPL, allowing this function to be
@@ -117,8 +102,9 @@ extern "C" {
        ({                                                              \
                __typeof__(*p) _________pold = (old);                   \
                __typeof__(*p) _________pnew = (_new);                  \
-               uatomic_cmpxchg(p, _________pold, _________pnew);       \
-       })
+               uatomic_cmpxchg_mo(p, _________pold, _________pnew,     \
+                                  CMM_SEQ_CST, CMM_RELAXED);           \
+       });
 
 /**
  * _rcu_xchg_pointer - same as rcu_assign_pointer, but returns the previous
@@ -135,17 +121,17 @@ extern "C" {
        __extension__                                   \
        ({                                              \
                __typeof__(*p) _________pv = (v);       \
-               uatomic_xchg(p, _________pv);           \
+               uatomic_xchg_mo(p, _________pv,         \
+                               CMM_SEQ_CST);           \
        })
 
 
-#define _rcu_set_pointer(p, v)                         \
-       do {                                            \
-               __typeof__(*p) _________pv = (v);       \
-               if (!__builtin_constant_p(v) ||         \
-                   ((v) != NULL))                      \
-                       cmm_wmb();                              \
-               uatomic_set(p, _________pv);            \
+#define _rcu_set_pointer(p, v)                                         \
+       do {                                                            \
+               __typeof__(*p) _________pv = (v);                       \
+               uatomic_store(p, _________pv,                           \
+                       __builtin_constant_p(v) && (v) == NULL ?        \
+                       CMM_RELAXED : CMM_RELEASE);                     \
        } while (0)
 
 /**
index 8816c50c924e07293aabf72af9a615046c85ef77..139c37f6d474a646141bb0650b6f8c936061a9a2 100644 (file)
@@ -9,9 +9,29 @@
  * System definitions.
  */
 
+#include <urcu/config.h>
 #include <urcu/compiler.h>
 #include <urcu/arch.h>
 
+#ifdef CONFIG_RCU_USE_ATOMIC_BUILTINS
+
+#define CMM_LOAD_SHARED(x)                                             \
+       __atomic_load_n(cmm_cast_volatile(&(x)), __ATOMIC_RELAXED)
+
+#define _CMM_LOAD_SHARED(x) CMM_LOAD_SHARED(x)
+
+#define CMM_STORE_SHARED(x, v)                                 \
+       __extension__                                           \
+       ({                                                      \
+               __typeof__(v) _v = (v);                         \
+               __atomic_store_n(cmm_cast_volatile(&(x)), _v,   \
+                                __ATOMIC_RELAXED);             \
+               _v;                                             \
+       })
+
+#define _CMM_STORE_SHARED(x, v) CMM_STORE_SHARED(x, v)
+
+#else
 /*
  * Identify a shared load. A cmm_smp_rmc() or cmm_smp_mc() should come
  * before the load.
@@ -46,4 +66,6 @@
                _v = _v;        /* Work around clang "unused result" */ \
        })
 
+#endif /* CONFIG_RCU_USE_ATOMIC_BUILTINS */
+
 #endif /* _URCU_SYSTEM_H */
index b7b1d89d0ca1c6e66209ccfcfc4f2b76f9ebaf36..5365f1fb0fdac0159edb06835f639b3045cc79eb 100644 (file)
@@ -5,9 +5,70 @@
 #ifndef _URCU_UATOMIC_H
 #define _URCU_UATOMIC_H
 
+#include <assert.h>
+
 #include <urcu/arch.h>
+#include <urcu/config.h>
+
+enum cmm_memorder {
+       CMM_RELAXED = 0,
+       CMM_CONSUME = 1,
+       CMM_ACQUIRE = 2,
+       CMM_RELEASE = 3,
+       CMM_ACQ_REL = 4,
+       CMM_SEQ_CST = 5,
+       CMM_SEQ_CST_FENCE = 6,
+};
+
+#ifdef CONFIG_RCU_USE_ATOMIC_BUILTINS
+
+/*
+ * Make sure that CMM_SEQ_CST_FENCE is not equivalent to other memory orders.
+ */
+# ifdef static_assert
+static_assert(CMM_RELAXED == __ATOMIC_RELAXED, "");
+static_assert(CMM_CONSUME == __ATOMIC_CONSUME, "");
+static_assert(CMM_ACQUIRE == __ATOMIC_ACQUIRE, "");
+static_assert(CMM_RELEASE == __ATOMIC_RELEASE, "");
+static_assert(CMM_ACQ_REL == __ATOMIC_ACQ_REL, "");
+static_assert(CMM_SEQ_CST == __ATOMIC_SEQ_CST, "");
+# endif
+
+/*
+ * This is not part of the public API. It it used internally to implement the
+ * CMM_SEQ_CST_FENCE memory order.
+ *
+ * NOTE: Using switch here instead of if statement to avoid -Wduplicated-cond
+ * warning when memory order is conditionally determined.
+ */
+static inline void cmm_seq_cst_fence_after_atomic(enum cmm_memorder mo)
+{
+       switch (mo) {
+       case CMM_SEQ_CST_FENCE:
+               cmm_smp_mb();
+               break;
+       default:
+               break;
+       }
+}
+
+#endif
+
+/*
+ * This is not part of the public API. It is used internally to convert from the
+ * CMM memory model to the C11 memory model.
+ */
+static inline int cmm_to_c11(int mo)
+{
+       if (mo == CMM_SEQ_CST_FENCE) {
+               return CMM_SEQ_CST;
+       }
+       return mo;
+}
 
-#if defined(URCU_ARCH_X86)
+#if defined(CONFIG_RCU_USE_ATOMIC_BUILTINS)
+#include <urcu/uatomic/builtins.h>
+#elif defined(URCU_ARCH_X86)
 #include <urcu/uatomic/x86.h>
 #elif defined(URCU_ARCH_PPC)
 #include <urcu/uatomic/ppc.h>
diff --git a/include/urcu/uatomic/builtins-generic.h b/include/urcu/uatomic/builtins-generic.h
new file mode 100644 (file)
index 0000000..a641bc9
--- /dev/null
@@ -0,0 +1,174 @@
+/*
+ * urcu/uatomic/builtins-generic.h
+ *
+ * Copyright (c) 2023 Olivier Dion <odion@efficios.com>
+ *
+ * This library is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Lesser General Public
+ * License as published by the Free Software Foundation; either
+ * version 2.1 of the License, or (at your option) any later version.
+ *
+ * This library is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+ * Lesser General Public License for more details.
+ *
+ * You should have received a copy of the GNU Lesser General Public
+ * License along with this library; if not, write to the Free Software
+ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
+ */
+
+#ifndef _URCU_UATOMIC_BUILTINS_GENERIC_H
+#define _URCU_UATOMIC_BUILTINS_GENERIC_H
+
+#include <urcu/compiler.h>
+#include <urcu/system.h>
+
+#define uatomic_store(addr, v, mo)                             \
+       do {                                                    \
+               __atomic_store_n(cmm_cast_volatile(addr), v,    \
+                               cmm_to_c11(mo));                \
+               cmm_seq_cst_fence_after_atomic(mo);             \
+       } while (0)
+
+#define uatomic_set(addr, v)                           \
+       do {                                            \
+               uatomic_store(addr, v, CMM_RELAXED);    \
+       } while (0)
+
+#define uatomic_load(addr, mo)                                         \
+       __extension__                                                   \
+       ({                                                              \
+               __typeof__(*(addr)) _value =                            \
+                       __atomic_load_n(cmm_cast_volatile(addr),        \
+                                       cmm_to_c11(mo));                \
+               cmm_seq_cst_fence_after_atomic(mo);                     \
+                                                                       \
+               _value;                                                 \
+       })
+
+#define uatomic_read(addr)                     \
+       uatomic_load(addr, CMM_RELAXED)
+
+#define uatomic_cmpxchg_mo(addr, old, new, mos, mof)                   \
+       __extension__                                                   \
+       ({                                                              \
+               __typeof__(*(addr)) _old = (__typeof__(*(addr)))old;    \
+                                                                       \
+               if (__atomic_compare_exchange_n(cmm_cast_volatile(addr), \
+                                                       &_old, new, 0,  \
+                                                       cmm_to_c11(mos), \
+                                                       cmm_to_c11(mof))) { \
+                       cmm_seq_cst_fence_after_atomic(mos);            \
+               } else {                                                \
+                       cmm_seq_cst_fence_after_atomic(mof);            \
+               }                                                       \
+               _old;                                                   \
+       })
+
+#define uatomic_cmpxchg(addr, old, new)                                        \
+       uatomic_cmpxchg_mo(addr, old, new, CMM_SEQ_CST_FENCE, CMM_RELAXED)
+
+#define uatomic_xchg_mo(addr, v, mo)                                   \
+       __extension__                                                   \
+       ({                                                              \
+               __typeof__((*addr)) _old =                              \
+                       __atomic_exchange_n(cmm_cast_volatile(addr), v, \
+                                       cmm_to_c11(mo));                \
+               cmm_seq_cst_fence_after_atomic(mo);                     \
+               _old;                                                   \
+       })
+
+#define uatomic_xchg(addr, v)                          \
+       uatomic_xchg_mo(addr, v, CMM_SEQ_CST_FENCE)
+
+#define uatomic_add_return_mo(addr, v, mo)                             \
+       __extension__                                                   \
+       ({                                                              \
+               __typeof__(*(addr)) _old =                              \
+                       __atomic_add_fetch(cmm_cast_volatile(addr), v,  \
+                                       cmm_to_c11(mo));                \
+               cmm_seq_cst_fence_after_atomic(mo);                     \
+               _old;                                                   \
+       })
+
+#define uatomic_add_return(addr, v)                            \
+       uatomic_add_return_mo(addr, v, CMM_SEQ_CST_FENCE)
+
+#define uatomic_sub_return_mo(addr, v, mo)                             \
+       __extension__                                                   \
+       ({                                                              \
+               __typeof__(*(addr)) _old =                              \
+                       __atomic_sub_fetch(cmm_cast_volatile(addr), v,  \
+                                       cmm_to_c11(mo));                \
+               cmm_seq_cst_fence_after_atomic(mo);                     \
+               _old;                                                   \
+       })
+
+#define uatomic_sub_return(addr, v)                            \
+       uatomic_sub_return_mo(addr, v, CMM_SEQ_CST_FENCE)
+
+#define uatomic_and_mo(addr, mask, mo)                                 \
+       do {                                                            \
+               (void) __atomic_and_fetch(cmm_cast_volatile(addr), mask, \
+                                       cmm_to_c11(mo));                \
+               cmm_seq_cst_fence_after_atomic(mo);                     \
+       } while (0)
+
+#define uatomic_and(addr, mask)                        \
+       uatomic_and_mo(addr, mask, CMM_SEQ_CST)
+
+#define uatomic_or_mo(addr, mask, mo)                                  \
+       do {                                                            \
+               (void) __atomic_or_fetch(cmm_cast_volatile(addr), mask, \
+                                       cmm_to_c11(mo));                \
+               cmm_seq_cst_fence_after_atomic(mo);                     \
+       } while (0)
+
+
+#define uatomic_or(addr, mask)                 \
+       uatomic_or_mo(addr, mask, CMM_RELAXED)
+
+#define uatomic_add_mo(addr, v, mo)                    \
+       (void) uatomic_add_return_mo(addr, v, mo)
+
+#define uatomic_add(addr, v)                   \
+       uatomic_add_mo(addr, v, CMM_RELAXED)
+
+#define uatomic_sub_mo(addr, v, mo)                    \
+       (void) uatomic_sub_return_mo(addr, v, mo)
+
+#define uatomic_sub(addr, v)                   \
+       uatomic_sub_mo(addr, v, CMM_RELAXED)
+
+#define uatomic_inc_mo(addr, mo)               \
+       uatomic_add_mo(addr, 1, mo)
+
+#define uatomic_inc(addr)                      \
+       uatomic_inc_mo(addr, CMM_RELAXED)
+
+#define uatomic_dec_mo(addr, mo)               \
+       uatomic_sub_mo(addr, 1, mo)
+
+#define uatomic_dec(addr)                      \
+       uatomic_dec_mo(addr, CMM_RELAXED)
+
+#define cmm_smp_mb__before_uatomic_and() cmm_smp_mb()
+#define cmm_smp_mb__after_uatomic_and()  cmm_smp_mb()
+
+#define cmm_smp_mb__before_uatomic_or() cmm_smp_mb()
+#define cmm_smp_mb__after_uatomic_or()  cmm_smp_mb()
+
+#define cmm_smp_mb__before_uatomic_add() cmm_smp_mb()
+#define cmm_smp_mb__after_uatomic_add()  cmm_smp_mb()
+
+#define cmm_smp_mb__before_uatomic_sub() cmm_smp_mb()
+#define cmm_smp_mb__after_uatomic_sub()  cmm_smp_mb()
+
+#define cmm_smp_mb__before_uatomic_inc() cmm_smp_mb()
+#define cmm_smp_mb__after_uatomic_inc() cmm_smp_mb()
+
+#define cmm_smp_mb__before_uatomic_dec() cmm_smp_mb()
+#define cmm_smp_mb__after_uatomic_dec() cmm_smp_mb()
+
+#endif /* _URCU_UATOMIC_BUILTINS_X86_H */
diff --git a/include/urcu/uatomic/builtins.h b/include/urcu/uatomic/builtins.h
new file mode 100644 (file)
index 0000000..82e98f8
--- /dev/null
@@ -0,0 +1,79 @@
+/*
+ * urcu/uatomic/builtins.h
+ *
+ * Copyright (c) 2023 Olivier Dion <odion@efficios.com>
+ *
+ * This library is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Lesser General Public
+ * License as published by the Free Software Foundation; either
+ * version 2.1 of the License, or (at your option) any later version.
+ *
+ * This library is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+ * Lesser General Public License for more details.
+ *
+ * You should have received a copy of the GNU Lesser General Public
+ * License along with this library; if not, write to the Free Software
+ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
+ */
+
+#ifndef _URCU_UATOMIC_BUILTINS_H
+#define _URCU_UATOMIC_BUILTINS_H
+
+#include <urcu/arch.h>
+
+#if defined(__has_builtin)
+# if !__has_builtin(__atomic_store_n)
+#  error "Toolchain does not support __atomic_store_n."
+# endif
+# if !__has_builtin(__atomic_load_n)
+#  error "Toolchain does not support __atomic_load_n."
+# endif
+# if !__has_builtin(__atomic_exchange_n)
+#  error "Toolchain does not support __atomic_exchange_n."
+# endif
+# if !__has_builtin(__atomic_compare_exchange_n)
+#  error "Toolchain does not support __atomic_compare_exchange_n."
+# endif
+# if !__has_builtin(__atomic_add_fetch)
+#  error "Toolchain does not support __atomic_add_fetch."
+# endif
+# if !__has_builtin(__atomic_sub_fetch)
+#  error "Toolchain does not support __atomic_sub_fetch."
+# endif
+# if !__has_builtin(__atomic_or_fetch)
+#  error "Toolchain does not support __atomic_or_fetch."
+# endif
+# if !__has_builtin(__atomic_thread_fence)
+#  error "Toolchain does not support __atomic_thread_fence."
+# endif
+# if !__has_builtin(__atomic_signal_fence)
+#  error "Toolchain does not support __atomic_signal_fence."
+# endif
+#elif defined(__GNUC__)
+# define GCC_VERSION (__GNUC__       * 10000 + \
+                      __GNUC_MINOR__ * 100   + \
+                      __GNUC_PATCHLEVEL__)
+# if  GCC_VERSION < 40700
+#  error "GCC version is too old. Version must be 4.7 or greater"
+# endif
+# undef  GCC_VERSION
+#else
+# error "Toolchain is not supported."
+#endif
+
+#if defined(__GNUC__)
+# define UATOMIC_HAS_ATOMIC_BYTE  __GCC_ATOMIC_CHAR_LOCK_FREE
+# define UATOMIC_HAS_ATOMIC_SHORT __GCC_ATOMIC_SHORT_LOCK_FREE
+#elif defined(__clang__)
+# define UATOMIC_HAS_ATOMIC_BYTE  __CLANG_ATOMIC_CHAR_LOCK_FREE
+# define UATOMIC_HAS_ATOMIC_SHORT __CLANG_ATOMIC_SHORT_LOCK_FREE
+#else
+/* #  define UATOMIC_HAS_ATOMIC_BYTE  */
+/* #  define UATOMIC_HAS_ATOMIC_SHORT */
+#endif
+
+#include <urcu/uatomic/builtins-generic.h>
+
+#endif /* _URCU_UATOMIC_BUILTINS_H */
index a03015897baa6fe980c36b03da95d181b4d93fc0..afc5185e18aa07a080e2d05ae2f086d4fc0040c8 100644 (file)
@@ -26,10 +26,244 @@ extern "C" {
 #define uatomic_set(addr, v)   ((void) CMM_STORE_SHARED(*(addr), (v)))
 #endif
 
+extern void abort(void);
+
+#define uatomic_load_store_return_op(op, addr, v, mo)                  \
+       __extension__                                                   \
+       ({                                                              \
+                                                                       \
+               switch (mo) {                                           \
+               case CMM_ACQUIRE:                                       \
+               case CMM_CONSUME:                                       \
+               case CMM_RELAXED:                                       \
+                       break;                                          \
+               case CMM_RELEASE:                                       \
+               case CMM_ACQ_REL:                                       \
+               case CMM_SEQ_CST:                                       \
+               case CMM_SEQ_CST_FENCE:                                 \
+                       cmm_smp_mb();                                   \
+                       break;                                          \
+               default:                                                \
+                       abort();                                        \
+               }                                                       \
+                                                                       \
+               __typeof__((*addr)) _value = op(addr, v);               \
+                                                                       \
+               switch (mo) {                                           \
+               case CMM_CONSUME:                                       \
+                       cmm_smp_read_barrier_depends();                 \
+                       break;                                          \
+               case CMM_ACQUIRE:                                       \
+               case CMM_ACQ_REL:                                       \
+               case CMM_SEQ_CST:                                       \
+               case CMM_SEQ_CST_FENCE:                                 \
+                       cmm_smp_mb();                                   \
+                       break;                                          \
+               case CMM_RELAXED:                                       \
+               case CMM_RELEASE:                                       \
+                       break;                                          \
+               default:                                                \
+                       abort();                                        \
+               }                                                       \
+               _value;                                                 \
+       })
+
+#define uatomic_load_store_op(op, addr, v, mo)                         \
+       do {                                                            \
+               switch (mo) {                                           \
+               case CMM_ACQUIRE:                                       \
+               case CMM_CONSUME:                                       \
+               case CMM_RELAXED:                                       \
+                       break;                                          \
+               case CMM_RELEASE:                                       \
+               case CMM_ACQ_REL:                                       \
+               case CMM_SEQ_CST:                                       \
+               case CMM_SEQ_CST_FENCE:                                 \
+                       cmm_smp_mb();                                   \
+                       break;                                          \
+               default:                                                \
+                       abort();                                        \
+               }                                                       \
+                                                                       \
+               op(addr, v);                                            \
+                                                                       \
+               switch (mo) {                                           \
+               case CMM_CONSUME:                                       \
+                       cmm_smp_read_barrier_depends();                 \
+                       break;                                          \
+               case CMM_ACQUIRE:                                       \
+               case CMM_ACQ_REL:                                       \
+               case CMM_SEQ_CST:                                       \
+               case CMM_SEQ_CST_FENCE:                                 \
+                       cmm_smp_mb();                                   \
+                       break;                                          \
+               case CMM_RELAXED:                                       \
+               case CMM_RELEASE:                                       \
+                       break;                                          \
+               default:                                                \
+                       abort();                                        \
+               }                                                       \
+       } while (0)
+
+#define uatomic_store(addr, v, mo)                     \
+       do {                                            \
+               switch (mo) {                           \
+               case CMM_RELAXED:                       \
+                       break;                          \
+               case CMM_RELEASE:                       \
+               case CMM_SEQ_CST:                       \
+               case CMM_SEQ_CST_FENCE:                 \
+                       cmm_smp_mb();                   \
+                       break;                          \
+               default:                                \
+                       abort();                        \
+               }                                       \
+                                                       \
+               uatomic_set(addr, v);                   \
+                                                       \
+               switch (mo) {                           \
+               case CMM_RELAXED:                       \
+               case CMM_RELEASE:                       \
+                       break;                          \
+               case CMM_SEQ_CST:                       \
+               case CMM_SEQ_CST_FENCE:                 \
+                       cmm_smp_mb();                   \
+                       break;                          \
+               default:                                \
+                       abort();                        \
+               }                                       \
+       } while (0)
+
+#define uatomic_and_mo(addr, v, mo)                            \
+       uatomic_load_store_op(uatomic_and, addr, v, mo)
+
+#define uatomic_or_mo(addr, v, mo)                             \
+       uatomic_load_store_op(uatomic_or, addr, v, mo)
+
+#define uatomic_add_mo(addr, v, mo)                            \
+       uatomic_load_store_op(uatomic_add, addr, v, mo)
+
+#define uatomic_sub_mo(addr, v, mo)                            \
+       uatomic_load_store_op(uatomic_sub, addr, v, mo)
+
+#define uatomic_inc_mo(addr, mo)                               \
+       uatomic_load_store_op(uatomic_add, addr, 1, mo)
+
+#define uatomic_dec_mo(addr, mo)                               \
+       uatomic_load_store_op(uatomic_add, addr, -1, mo)
+/*
+ * NOTE: We can not just do switch (_value == (old) ? mos : mof) otherwise the
+ * compiler emit a -Wduplicated-cond warning.
+ */
+#define uatomic_cmpxchg_mo(addr, old, new, mos, mof)                   \
+       __extension__                                                   \
+       ({                                                              \
+               switch (mos) {                                          \
+               case CMM_ACQUIRE:                                       \
+               case CMM_CONSUME:                                       \
+               case CMM_RELAXED:                                       \
+                       break;                                          \
+               case CMM_RELEASE:                                       \
+               case CMM_ACQ_REL:                                       \
+               case CMM_SEQ_CST:                                       \
+               case CMM_SEQ_CST_FENCE:                                 \
+                       cmm_smp_mb();                                   \
+                       break;                                          \
+               default:                                                \
+                       abort();                                        \
+               }                                                       \
+                                                                       \
+               __typeof__(*(addr)) _value = uatomic_cmpxchg(addr, old, \
+                                                       new);           \
+                                                                       \
+               if (_value == (old)) {                                  \
+                       switch (mos) {                                  \
+                       case CMM_CONSUME:                               \
+                               cmm_smp_read_barrier_depends();         \
+                               break;                                  \
+                       case CMM_ACQUIRE:                               \
+                       case CMM_ACQ_REL:                               \
+                       case CMM_SEQ_CST:                               \
+                       case CMM_SEQ_CST_FENCE:                         \
+                               cmm_smp_mb();                           \
+                               break;                                  \
+                       case CMM_RELAXED:                               \
+                       case CMM_RELEASE:                               \
+                               break;                                  \
+                       default:                                        \
+                               abort();                                \
+                       }                                               \
+               } else {                                                \
+                       switch (mof) {                                  \
+                       case CMM_CONSUME:                               \
+                               cmm_smp_read_barrier_depends();         \
+                               break;                                  \
+                       case CMM_ACQUIRE:                               \
+                       case CMM_ACQ_REL:                               \
+                       case CMM_SEQ_CST:                               \
+                       case CMM_SEQ_CST_FENCE:                         \
+                               cmm_smp_mb();                           \
+                               break;                                  \
+                       case CMM_RELAXED:                               \
+                       case CMM_RELEASE:                               \
+                               break;                                  \
+                       default:                                        \
+                               abort();                                \
+                       }                                               \
+               }                                                       \
+               _value;                                                 \
+       })
+
+#define uatomic_xchg_mo(addr, v, mo)                           \
+       uatomic_load_store_return_op(uatomic_xchg, addr, v, mo)
+
+#define uatomic_add_return_mo(addr, v, mo)                             \
+       uatomic_load_store_return_op(uatomic_add_return, addr, v)
+
+#define uatomic_sub_return_mo(addr, v, mo)                             \
+       uatomic_load_store_return_op(uatomic_sub_return, addr, v)
+
+
 #ifndef uatomic_read
 #define uatomic_read(addr)     CMM_LOAD_SHARED(*(addr))
 #endif
 
+#define uatomic_load(addr, mo)                                         \
+       __extension__                                                   \
+       ({                                                              \
+               switch (mo) {                                           \
+               case CMM_ACQUIRE:                                       \
+               case CMM_CONSUME:                                       \
+               case CMM_RELAXED:                                       \
+                       break;                                          \
+               case CMM_SEQ_CST:                                       \
+               case CMM_SEQ_CST_FENCE:                                 \
+                       cmm_smp_mb();                                   \
+                       break;                                          \
+               default:                                                \
+                       abort();                                        \
+               }                                                       \
+                                                                       \
+               __typeof__(*(addr)) _rcu_value = uatomic_read(addr);    \
+                                                                       \
+               switch (mo) {                                           \
+               case CMM_RELAXED:                                       \
+                       break;                                          \
+               case CMM_CONSUME:                                       \
+                       cmm_smp_read_barrier_depends();                 \
+                       break;                                          \
+               case CMM_ACQUIRE:                                       \
+               case CMM_SEQ_CST:                                       \
+               case CMM_SEQ_CST_FENCE:                                 \
+                       cmm_smp_mb();                                   \
+                       break;                                          \
+               default:                                                \
+                       abort();                                        \
+               }                                                       \
+                                                                       \
+               _rcu_value;                                             \
+       })
+
 #if !defined __OPTIMIZE__  || defined UATOMIC_NO_LINK_ERROR
 #ifdef ILLEGAL_INSTR
 static inline __attribute__((always_inline))
index bc78c10bd255b12780eb36844c7e6383ef01ed51..a42fe87be47c0feea1549ccc91ba9db685342538 100644 (file)
@@ -24,19 +24,16 @@ void *rcu_dereference_sym(void *p)
 
 void *rcu_set_pointer_sym(void **p, void *v)
 {
-       cmm_wmb();
-       uatomic_set(p, v);
+       uatomic_store(p, v, CMM_RELEASE);
        return v;
 }
 
 void *rcu_xchg_pointer_sym(void **p, void *v)
 {
-       cmm_wmb();
-       return uatomic_xchg(p, v);
+       return uatomic_xchg_mo(p, v, CMM_SEQ_CST);
 }
 
 void *rcu_cmpxchg_pointer_sym(void **p, void *old, void *_new)
 {
-       cmm_wmb();
-       return uatomic_cmpxchg(p, old, _new);
+       return uatomic_cmpxchg_mo(p, old, _new, CMM_SEQ_CST, CMM_RELAXED);
 }
This page took 0.036991 seconds and 4 git commands to generate.