cmm_smp_mb__before_uatomic_or();
uatomic_or(&crdp->flags, URCU_CALL_RCU_PAUSED);
while ((uatomic_read(&crdp->flags) & URCU_CALL_RCU_PAUSE) != 0)
- poll(NULL, 0, 1);
+ (void) poll(NULL, 0, 1);
uatomic_and(&crdp->flags, ~URCU_CALL_RCU_PAUSED);
cmm_smp_mb__after_uatomic_and();
rcu_register_thread();
if (cds_wfcq_empty(&crdp->cbs_head,
&crdp->cbs_tail)) {
call_rcu_wait(crdp);
- poll(NULL, 0, 10);
+ (void) poll(NULL, 0, 10);
uatomic_dec(&crdp->futex);
/*
* Decrement futex before reading
*/
cmm_smp_mb();
} else {
- poll(NULL, 0, 10);
+ (void) poll(NULL, 0, 10);
}
} else {
- poll(NULL, 0, 10);
+ (void) poll(NULL, 0, 10);
}
rcu_thread_online();
}
struct call_rcu_data *crdp;
/* Holding rcu read-side lock across use of per-cpu crdp */
- rcu_read_lock();
+ _rcu_read_lock();
crdp = get_call_rcu_data();
_call_rcu(head, func, crdp);
- rcu_read_unlock();
+ _rcu_read_unlock();
}
/*
uatomic_or(&crdp->flags, URCU_CALL_RCU_STOP);
wake_call_rcu_thread(crdp);
while ((uatomic_read(&crdp->flags) & URCU_CALL_RCU_STOPPED) == 0)
- poll(NULL, 0, 1);
+ (void) poll(NULL, 0, 1);
}
if (!cds_wfcq_empty(&crdp->cbs_head, &crdp->cbs_tail)) {
/* Create default call rcu data if need be */
int was_online;
/* Put in offline state in QSBR. */
- was_online = rcu_read_ongoing();
+ was_online = _rcu_read_ongoing();
if (was_online)
rcu_thread_offline();
/*
* Calling a rcu_barrier() within a RCU read-side critical
* section is an error.
*/
- if (rcu_read_ongoing()) {
+ if (_rcu_read_ongoing()) {
static int warned = 0;
if (!warned) {
}
cds_list_for_each_entry(crdp, &call_rcu_data_list, list) {
while ((uatomic_read(&crdp->flags) & URCU_CALL_RCU_PAUSED) == 0)
- poll(NULL, 0, 1);
+ (void) poll(NULL, 0, 1);
}
}
uatomic_and(&crdp->flags, ~URCU_CALL_RCU_PAUSE);
cds_list_for_each_entry(crdp, &call_rcu_data_list, list) {
while ((uatomic_read(&crdp->flags) & URCU_CALL_RCU_PAUSED) != 0)
- poll(NULL, 0, 1);
+ (void) poll(NULL, 0, 1);
}
call_rcu_unlock(&call_rcu_mutex);
}