cmm_smp_mb__before_uatomic_or();
uatomic_or(&crdp->flags, URCU_CALL_RCU_PAUSED);
while ((uatomic_read(&crdp->flags) & URCU_CALL_RCU_PAUSE) != 0)
cmm_smp_mb__before_uatomic_or();
uatomic_or(&crdp->flags, URCU_CALL_RCU_PAUSED);
while ((uatomic_read(&crdp->flags) & URCU_CALL_RCU_PAUSE) != 0)
uatomic_and(&crdp->flags, ~URCU_CALL_RCU_PAUSED);
cmm_smp_mb__after_uatomic_and();
rcu_register_thread();
uatomic_and(&crdp->flags, ~URCU_CALL_RCU_PAUSED);
cmm_smp_mb__after_uatomic_and();
rcu_register_thread();
crdp = get_call_rcu_data();
_call_rcu(head, func, crdp);
crdp = get_call_rcu_data();
_call_rcu(head, func, crdp);
uatomic_or(&crdp->flags, URCU_CALL_RCU_STOP);
wake_call_rcu_thread(crdp);
while ((uatomic_read(&crdp->flags) & URCU_CALL_RCU_STOPPED) == 0)
uatomic_or(&crdp->flags, URCU_CALL_RCU_STOP);
wake_call_rcu_thread(crdp);
while ((uatomic_read(&crdp->flags) & URCU_CALL_RCU_STOPPED) == 0)
}
if (!cds_wfcq_empty(&crdp->cbs_head, &crdp->cbs_tail)) {
/* Create default call rcu data if need be */
}
if (!cds_wfcq_empty(&crdp->cbs_head, &crdp->cbs_tail)) {
/* Create default call rcu data if need be */
work = caa_container_of(head, struct call_rcu_completion_work, head);
completion = work->completion;
work = caa_container_of(head, struct call_rcu_completion_work, head);
completion = work->completion;
- uatomic_dec(&completion->barrier_count);
- call_rcu_completion_wake_up(completion);
+ if (!uatomic_sub_return(&completion->barrier_count, 1))
+ call_rcu_completion_wake_up(completion);
+ urcu_ref_put(&completion->ref, free_completion);
call_rcu_lock(&call_rcu_mutex);
cds_list_for_each_entry(crdp, &call_rcu_data_list, list)
count++;
call_rcu_lock(&call_rcu_mutex);
cds_list_for_each_entry(crdp, &call_rcu_data_list, list)
count++;
- completion.barrier_count = count;
+ /* Referenced by rcu_barrier() and each call_rcu thread. */
+ urcu_ref_set(&completion->ref, count + 1);
+ completion->barrier_count = count;
cds_list_for_each_entry(crdp, &call_rcu_data_list, list) {
struct call_rcu_completion_work *work;
cds_list_for_each_entry(crdp, &call_rcu_data_list, list) {
struct call_rcu_completion_work *work;
_call_rcu(&work->head, _rcu_barrier_complete, crdp);
}
call_rcu_unlock(&call_rcu_mutex);
/* Wait for them */
for (;;) {
_call_rcu(&work->head, _rcu_barrier_complete, crdp);
}
call_rcu_unlock(&call_rcu_mutex);
/* Wait for them */
for (;;) {
}
cds_list_for_each_entry(crdp, &call_rcu_data_list, list) {
while ((uatomic_read(&crdp->flags) & URCU_CALL_RCU_PAUSED) == 0)
}
cds_list_for_each_entry(crdp, &call_rcu_data_list, list) {
while ((uatomic_read(&crdp->flags) & URCU_CALL_RCU_PAUSED) == 0)
uatomic_and(&crdp->flags, ~URCU_CALL_RCU_PAUSE);
cds_list_for_each_entry(crdp, &call_rcu_data_list, list) {
while ((uatomic_read(&crdp->flags) & URCU_CALL_RCU_PAUSED) != 0)
uatomic_and(&crdp->flags, ~URCU_CALL_RCU_PAUSE);
cds_list_for_each_entry(crdp, &call_rcu_data_list, list) {
while ((uatomic_read(&crdp->flags) & URCU_CALL_RCU_PAUSED) != 0)