|
@@ -2347,7 +2347,7 @@ static void rcu_do_batch(struct rcu_state *rsp, struct rcu_data *rdp)
|
|
|
}
|
|
|
smp_mb(); /* List handling before counting for rcu_barrier(). */
|
|
|
rdp->qlen_lazy -= count_lazy;
|
|
|
- ACCESS_ONCE(rdp->qlen) -= count;
|
|
|
+ ACCESS_ONCE(rdp->qlen) = rdp->qlen - count;
|
|
|
rdp->n_cbs_invoked += count;
|
|
|
|
|
|
/* Reinstate batch limit if we have worked down the excess. */
|
|
@@ -2492,7 +2492,7 @@ static void force_quiescent_state(struct rcu_state *rsp)
|
|
|
if (rnp_old != NULL)
|
|
|
raw_spin_unlock(&rnp_old->fqslock);
|
|
|
if (ret) {
|
|
|
- ACCESS_ONCE(rsp->n_force_qs_lh)++;
|
|
|
+ rsp->n_force_qs_lh++;
|
|
|
return;
|
|
|
}
|
|
|
rnp_old = rnp;
|
|
@@ -2504,7 +2504,7 @@ static void force_quiescent_state(struct rcu_state *rsp)
|
|
|
smp_mb__after_unlock_lock();
|
|
|
raw_spin_unlock(&rnp_old->fqslock);
|
|
|
if (ACCESS_ONCE(rsp->gp_flags) & RCU_GP_FLAG_FQS) {
|
|
|
- ACCESS_ONCE(rsp->n_force_qs_lh)++;
|
|
|
+ rsp->n_force_qs_lh++;
|
|
|
raw_spin_unlock_irqrestore(&rnp_old->lock, flags);
|
|
|
return; /* Someone beat us to it. */
|
|
|
}
|
|
@@ -2693,7 +2693,7 @@ __call_rcu(struct rcu_head *head, void (*func)(struct rcu_head *rcu),
|
|
|
local_irq_restore(flags);
|
|
|
return;
|
|
|
}
|
|
|
- ACCESS_ONCE(rdp->qlen)++;
|
|
|
+ ACCESS_ONCE(rdp->qlen) = rdp->qlen + 1;
|
|
|
if (lazy)
|
|
|
rdp->qlen_lazy++;
|
|
|
else
|
|
@@ -3257,7 +3257,7 @@ static void _rcu_barrier(struct rcu_state *rsp)
|
|
|
* ACCESS_ONCE() to prevent the compiler from speculating
|
|
|
* the increment to precede the early-exit check.
|
|
|
*/
|
|
|
- ACCESS_ONCE(rsp->n_barrier_done)++;
|
|
|
+ ACCESS_ONCE(rsp->n_barrier_done) = rsp->n_barrier_done + 1;
|
|
|
WARN_ON_ONCE((rsp->n_barrier_done & 0x1) != 1);
|
|
|
_rcu_barrier_trace(rsp, "Inc1", -1, rsp->n_barrier_done);
|
|
|
smp_mb(); /* Order ->n_barrier_done increment with below mechanism. */
|
|
@@ -3307,7 +3307,7 @@ static void _rcu_barrier(struct rcu_state *rsp)
|
|
|
|
|
|
/* Increment ->n_barrier_done to prevent duplicate work. */
|
|
|
smp_mb(); /* Keep increment after above mechanism. */
|
|
|
- ACCESS_ONCE(rsp->n_barrier_done)++;
|
|
|
+ ACCESS_ONCE(rsp->n_barrier_done) = rsp->n_barrier_done + 1;
|
|
|
WARN_ON_ONCE((rsp->n_barrier_done & 0x1) != 0);
|
|
|
_rcu_barrier_trace(rsp, "Inc2", -1, rsp->n_barrier_done);
|
|
|
smp_mb(); /* Keep increment before caller's subsequent code. */
|