1 #ifndef _URCU_STATIC_WFSTACK_H
2 #define _URCU_STATIC_WFSTACK_H
5 * urcu/static/wfstack.h
7 * Userspace RCU library - Stack with with wait-free push, blocking traversal.
9 * TO BE INCLUDED ONLY IN LGPL-COMPATIBLE CODE. See urcu/wfstack.h for
10 * linking dynamically with the userspace rcu library.
12 * Copyright 2010-2012 - Mathieu Desnoyers <mathieu.desnoyers@efficios.com>
14 * This library is free software; you can redistribute it and/or
15 * modify it under the terms of the GNU Lesser General Public
16 * License as published by the Free Software Foundation; either
17 * version 2.1 of the License, or (at your option) any later version.
19 * This library is distributed in the hope that it will be useful,
20 * but WITHOUT ANY WARRANTY; without even the implied warranty of
21 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
22 * Lesser General Public License for more details.
24 * You should have received a copy of the GNU Lesser General Public
25 * License along with this library; if not, write to the Free Software
26 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
33 #include <urcu/compiler.h>
34 #include <urcu/uatomic.h>
40 #define CDS_WFS_END ((void *) 0x1UL)
41 #define CDS_WFS_ADAPT_ATTEMPTS 10 /* Retry if being set */
42 #define CDS_WFS_WAIT 10 /* Wait 10 ms if being set */
45 * Stack with wait-free push, blocking traversal.
47 * Stack implementing push, pop, pop_all operations, as well as iterator
48 * on the stack head returned by pop_all.
50 * Wait-free operations: cds_wfs_push, __cds_wfs_pop_all, cds_wfs_empty,
52 * Blocking operations: cds_wfs_pop, cds_wfs_pop_all, cds_wfs_next,
53 * iteration on stack head returned by pop_all.
55 * Synchronization table:
57 * External synchronization techniques described in the API below is
58 * required between pairs marked with "X". No external synchronization
59 * required between pairs marked with "-".
61 * cds_wfs_push __cds_wfs_pop __cds_wfs_pop_all
64 * __cds_wfs_pop_all - X -
66 * cds_wfs_pop and cds_wfs_pop_all use an internal mutex to provide
71 * cds_wfs_node_init: initialize wait-free stack node.
74 void _cds_wfs_node_init(struct cds_wfs_node
*node
)
80 * __cds_wfs_init: initialize wait-free stack.
82 static inline void ___cds_wfs_init(struct __cds_wfs_stack
*s
)
84 s
->head
= CDS_WFS_END
;
88 * cds_wfs_init: initialize wait-free stack.
91 void _cds_wfs_init(struct cds_wfs_stack
*s
)
95 s
->head
= CDS_WFS_END
;
96 ret
= pthread_mutex_init(&s
->lock
, NULL
);
100 static inline bool ___cds_wfs_end(void *node
)
102 return node
== CDS_WFS_END
;
106 * cds_wfs_empty: return whether wait-free stack is empty.
108 * No memory barrier is issued. No mutual exclusion is required.
110 static inline bool _cds_wfs_empty(cds_wfs_stack_ptr_t u_stack
)
112 struct __cds_wfs_stack
*s
= u_stack
._s
;
114 return ___cds_wfs_end(CMM_LOAD_SHARED(s
->head
));
118 * cds_wfs_push: push a node into the stack.
120 * Issues a full memory barrier before push. No mutual exclusion is
123 * Returns 0 if the stack was empty prior to adding the node.
124 * Returns non-zero otherwise.
127 int _cds_wfs_push(cds_wfs_stack_ptr_t u_stack
, struct cds_wfs_node
*node
)
129 struct __cds_wfs_stack
*s
= u_stack
._s
;
130 struct cds_wfs_head
*old_head
, *new_head
;
132 assert(node
->next
== NULL
);
133 new_head
= caa_container_of(node
, struct cds_wfs_head
, node
);
135 * uatomic_xchg() implicit memory barrier orders earlier stores
136 * to node (setting it to NULL) before publication.
138 old_head
= uatomic_xchg(&s
->head
, new_head
);
140 * At this point, dequeuers see a NULL node->next, they should
141 * busy-wait until node->next is set to old_head.
143 CMM_STORE_SHARED(node
->next
, &old_head
->node
);
144 return !___cds_wfs_end(old_head
);
148 * Waiting for push to complete enqueue and return the next node.
150 static inline struct cds_wfs_node
*
151 ___cds_wfs_node_sync_next(struct cds_wfs_node
*node
, int blocking
)
153 struct cds_wfs_node
*next
;
157 * Adaptative busy-looping waiting for push to complete.
159 while ((next
= CMM_LOAD_SHARED(node
->next
)) == NULL
) {
161 return CDS_WFS_WOULDBLOCK
;
162 if (++attempt
>= CDS_WFS_ADAPT_ATTEMPTS
) {
163 poll(NULL
, 0, CDS_WFS_WAIT
); /* Wait for 10ms */
174 struct cds_wfs_node
*
175 ___cds_wfs_pop(cds_wfs_stack_ptr_t u_stack
, int *state
, int blocking
)
177 struct cds_wfs_head
*head
, *new_head
;
178 struct cds_wfs_node
*next
;
179 struct __cds_wfs_stack
*s
= u_stack
._s
;
184 head
= CMM_LOAD_SHARED(s
->head
);
185 if (___cds_wfs_end(head
)) {
188 next
= ___cds_wfs_node_sync_next(&head
->node
, blocking
);
189 if (!blocking
&& next
== CDS_WFS_WOULDBLOCK
) {
190 return CDS_WFS_WOULDBLOCK
;
192 new_head
= caa_container_of(next
, struct cds_wfs_head
, node
);
193 if (uatomic_cmpxchg(&s
->head
, head
, new_head
) == head
) {
194 if (state
&& ___cds_wfs_end(new_head
))
195 *state
|= CDS_WFS_STATE_LAST
;
199 return CDS_WFS_WOULDBLOCK
;
201 /* busy-loop if head changed under us */
206 * __cds_wfs_pop_with_state_blocking: pop a node from the stack, with state.
208 * Returns NULL if stack is empty.
210 * __cds_wfs_pop_blocking needs to be synchronized using one of the
211 * following techniques:
213 * 1) Calling __cds_wfs_pop_blocking under rcu read lock critical
214 * section. The caller must wait for a grace period to pass before
215 * freeing the returned node or modifying the cds_wfs_node structure.
216 * 2) Using mutual exclusion (e.g. mutexes) to protect
217 * __cds_wfs_pop_blocking and __cds_wfs_pop_all callers.
218 * 3) Ensuring that only ONE thread can call __cds_wfs_pop_blocking()
219 * and __cds_wfs_pop_all(). (multi-provider/single-consumer scheme).
221 * "state" saves state flags atomically sampled with pop operation.
224 struct cds_wfs_node
*
225 ___cds_wfs_pop_with_state_blocking(cds_wfs_stack_ptr_t u_stack
, int *state
)
227 return ___cds_wfs_pop(u_stack
, state
, 1);
231 struct cds_wfs_node
*
232 ___cds_wfs_pop_blocking(cds_wfs_stack_ptr_t u_stack
)
234 return ___cds_wfs_pop_with_state_blocking(u_stack
, NULL
);
238 * __cds_wfs_pop_with_state_nonblocking: pop a node from the stack.
240 * Same as __cds_wfs_pop_with_state_blocking, but returns
241 * CDS_WFS_WOULDBLOCK if it needs to block.
243 * "state" saves state flags atomically sampled with pop operation.
246 struct cds_wfs_node
*
247 ___cds_wfs_pop_with_state_nonblocking(cds_wfs_stack_ptr_t u_stack
, int *state
)
249 return ___cds_wfs_pop(u_stack
, state
, 0);
253 * __cds_wfs_pop_nonblocking: pop a node from the stack.
255 * Same as __cds_wfs_pop_blocking, but returns CDS_WFS_WOULDBLOCK if
259 struct cds_wfs_node
*
260 ___cds_wfs_pop_nonblocking(cds_wfs_stack_ptr_t u_stack
)
262 return ___cds_wfs_pop_with_state_nonblocking(u_stack
, NULL
);
266 * __cds_wfs_pop_all: pop all nodes from a stack.
268 * __cds_wfs_pop_all does not require any synchronization with other
269 * push, nor with other __cds_wfs_pop_all, but requires synchronization
270 * matching the technique used to synchronize __cds_wfs_pop_blocking:
272 * 1) If __cds_wfs_pop_blocking is called under rcu read lock critical
273 * section, both __cds_wfs_pop_blocking and cds_wfs_pop_all callers
274 * must wait for a grace period to pass before freeing the returned
275 * node or modifying the cds_wfs_node structure. However, no RCU
276 * read-side critical section is needed around __cds_wfs_pop_all.
277 * 2) Using mutual exclusion (e.g. mutexes) to protect
278 * __cds_wfs_pop_blocking and __cds_wfs_pop_all callers.
279 * 3) Ensuring that only ONE thread can call __cds_wfs_pop_blocking()
280 * and __cds_wfs_pop_all(). (multi-provider/single-consumer scheme).
283 struct cds_wfs_head
*
284 ___cds_wfs_pop_all(cds_wfs_stack_ptr_t u_stack
)
286 struct __cds_wfs_stack
*s
= u_stack
._s
;
287 struct cds_wfs_head
*head
;
290 * Implicit memory barrier after uatomic_xchg() matches implicit
291 * memory barrier before uatomic_xchg() in cds_wfs_push. It
292 * ensures that all nodes of the returned list are consistent.
293 * There is no need to issue memory barriers when iterating on
294 * the returned list, because the full memory barrier issued
295 * prior to each uatomic_cmpxchg, which each write to head, are
296 * taking care to order writes to each node prior to the full
297 * memory barrier after this uatomic_xchg().
299 head
= uatomic_xchg(&s
->head
, CDS_WFS_END
);
300 if (___cds_wfs_end(head
))
306 * cds_wfs_pop_lock: lock stack pop-protection mutex.
308 static inline void _cds_wfs_pop_lock(struct cds_wfs_stack
*s
)
312 ret
= pthread_mutex_lock(&s
->lock
);
317 * cds_wfs_pop_unlock: unlock stack pop-protection mutex.
319 static inline void _cds_wfs_pop_unlock(struct cds_wfs_stack
*s
)
323 ret
= pthread_mutex_unlock(&s
->lock
);
328 * Call __cds_wfs_pop_with_state_blocking with an internal pop mutex held.
331 struct cds_wfs_node
*
332 _cds_wfs_pop_with_state_blocking(struct cds_wfs_stack
*s
, int *state
)
334 struct cds_wfs_node
*retnode
;
336 _cds_wfs_pop_lock(s
);
337 retnode
= ___cds_wfs_pop_with_state_blocking(s
, state
);
338 _cds_wfs_pop_unlock(s
);
343 * Call _cds_wfs_pop_with_state_blocking without saving any state.
346 struct cds_wfs_node
*
347 _cds_wfs_pop_blocking(struct cds_wfs_stack
*s
)
349 return _cds_wfs_pop_with_state_blocking(s
, NULL
);
353 * Call __cds_wfs_pop_all with an internal pop mutex held.
356 struct cds_wfs_head
*
357 _cds_wfs_pop_all_blocking(struct cds_wfs_stack
*s
)
359 struct cds_wfs_head
*rethead
;
361 _cds_wfs_pop_lock(s
);
362 rethead
= ___cds_wfs_pop_all(s
);
363 _cds_wfs_pop_unlock(s
);
368 * cds_wfs_first: get first node of a popped stack.
370 * Content written into the node before enqueue is guaranteed to be
371 * consistent, but no other memory ordering is ensured.
373 * Used by for-like iteration macros in urcu/wfstack.h:
374 * cds_wfs_for_each_blocking()
375 * cds_wfs_for_each_blocking_safe()
377 * Returns NULL if popped stack is empty, top stack node otherwise.
379 static inline struct cds_wfs_node
*
380 _cds_wfs_first(struct cds_wfs_head
*head
)
382 if (___cds_wfs_end(head
))
387 static inline struct cds_wfs_node
*
388 ___cds_wfs_next(struct cds_wfs_node
*node
, int blocking
)
390 struct cds_wfs_node
*next
;
392 next
= ___cds_wfs_node_sync_next(node
, blocking
);
394 * CDS_WFS_WOULDBLOCK != CSD_WFS_END, so we can check for end
395 * even if ___cds_wfs_node_sync_next returns CDS_WFS_WOULDBLOCK,
396 * and still return CDS_WFS_WOULDBLOCK.
398 if (___cds_wfs_end(next
))
404 * cds_wfs_next_blocking: get next node of a popped stack.
406 * Content written into the node before enqueue is guaranteed to be
407 * consistent, but no other memory ordering is ensured.
409 * Used by for-like iteration macros in urcu/wfstack.h:
410 * cds_wfs_for_each_blocking()
411 * cds_wfs_for_each_blocking_safe()
413 * Returns NULL if reached end of popped stack, non-NULL next stack
416 static inline struct cds_wfs_node
*
417 _cds_wfs_next_blocking(struct cds_wfs_node
*node
)
419 return ___cds_wfs_next(node
, 1);
424 * cds_wfs_next_nonblocking: get next node of a popped stack.
426 * Same as cds_wfs_next_blocking, but returns CDS_WFS_WOULDBLOCK if it
429 static inline struct cds_wfs_node
*
430 _cds_wfs_next_nonblocking(struct cds_wfs_node
*node
)
432 return ___cds_wfs_next(node
, 0);
439 #endif /* _URCU_STATIC_WFSTACK_H */