Skip to content

Commit 1331e7a

Browse files
Paul E. McKenneypaulmck
Paul E. McKenney
authored andcommitted
rcu: Remove _rcu_barrier() dependency on __stop_machine()
Currently, _rcu_barrier() relies on preempt_disable() to prevent any CPU from going offline, which in turn depends on CPU hotplug's use of __stop_machine(). This patch therefore makes _rcu_barrier() use get_online_cpus() to block CPU-hotplug operations. This has the added benefit of removing the need for _rcu_barrier() to adopt callbacks: Because CPU-hotplug operations are excluded, there can be no callbacks to adopt. This commit simplifies the code accordingly. Signed-off-by: Paul E. McKenney <paul.mckenney@linaro.org> Signed-off-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com> Reviewed-by: Josh Triplett <josh@joshtriplett.org>
1 parent a10d206 commit 1331e7a

File tree

3 files changed

+13
-77
lines changed

3 files changed

+13
-77
lines changed

kernel/rcutree.c

+11-72
Original file line numberDiff line numberDiff line change
@@ -1392,17 +1392,6 @@ static void rcu_adopt_orphan_cbs(struct rcu_state *rsp)
13921392
int i;
13931393
struct rcu_data *rdp = __this_cpu_ptr(rsp->rda);
13941394

1395-
/*
1396-
* If there is an rcu_barrier() operation in progress, then
1397-
* only the task doing that operation is permitted to adopt
1398-
* callbacks. To do otherwise breaks rcu_barrier() and friends
1399-
* by causing them to fail to wait for the callbacks in the
1400-
* orphanage.
1401-
*/
1402-
if (rsp->rcu_barrier_in_progress &&
1403-
rsp->rcu_barrier_in_progress != current)
1404-
return;
1405-
14061395
/* Do the accounting first. */
14071396
rdp->qlen_lazy += rsp->qlen_lazy;
14081397
rdp->qlen += rsp->qlen;
@@ -1457,9 +1446,8 @@ static void rcu_cleanup_dying_cpu(struct rcu_state *rsp)
14571446
* The CPU has been completely removed, and some other CPU is reporting
14581447
* this fact from process context. Do the remainder of the cleanup,
14591448
* including orphaning the outgoing CPU's RCU callbacks, and also
1460-
* adopting them, if there is no _rcu_barrier() instance running.
1461-
* There can only be one CPU hotplug operation at a time, so no other
1462-
* CPU can be attempting to update rcu_cpu_kthread_task.
1449+
* adopting them. There can only be one CPU hotplug operation at a time,
1450+
* so no other CPU can be attempting to update rcu_cpu_kthread_task.
14631451
*/
14641452
static void rcu_cleanup_dead_cpu(int cpu, struct rcu_state *rsp)
14651453
{
@@ -1521,10 +1509,6 @@ static void rcu_cleanup_dead_cpu(int cpu, struct rcu_state *rsp)
15211509

15221510
#else /* #ifdef CONFIG_HOTPLUG_CPU */
15231511

1524-
static void rcu_adopt_orphan_cbs(struct rcu_state *rsp)
1525-
{
1526-
}
1527-
15281512
static void rcu_cleanup_dying_cpu(struct rcu_state *rsp)
15291513
{
15301514
}
@@ -2328,13 +2312,10 @@ static void rcu_barrier_func(void *type)
23282312
static void _rcu_barrier(struct rcu_state *rsp)
23292313
{
23302314
int cpu;
2331-
unsigned long flags;
23322315
struct rcu_data *rdp;
2333-
struct rcu_data rd;
23342316
unsigned long snap = ACCESS_ONCE(rsp->n_barrier_done);
23352317
unsigned long snap_done;
23362318

2337-
init_rcu_head_on_stack(&rd.barrier_head);
23382319
_rcu_barrier_trace(rsp, "Begin", -1, snap);
23392320

23402321
/* Take mutex to serialize concurrent rcu_barrier() requests. */
@@ -2374,70 +2355,30 @@ static void _rcu_barrier(struct rcu_state *rsp)
23742355
/*
23752356
* Initialize the count to one rather than to zero in order to
23762357
* avoid a too-soon return to zero in case of a short grace period
2377-
* (or preemption of this task). Also flag this task as doing
2378-
* an rcu_barrier(). This will prevent anyone else from adopting
2379-
* orphaned callbacks, which could cause otherwise failure if a
2380-
* CPU went offline and quickly came back online. To see this,
2381-
* consider the following sequence of events:
2382-
*
2383-
* 1. We cause CPU 0 to post an rcu_barrier_callback() callback.
2384-
* 2. CPU 1 goes offline, orphaning its callbacks.
2385-
* 3. CPU 0 adopts CPU 1's orphaned callbacks.
2386-
* 4. CPU 1 comes back online.
2387-
* 5. We cause CPU 1 to post an rcu_barrier_callback() callback.
2388-
* 6. Both rcu_barrier_callback() callbacks are invoked, awakening
2389-
* us -- but before CPU 1's orphaned callbacks are invoked!!!
2358+
* (or preemption of this task). Exclude CPU-hotplug operations
2359+
* to ensure that no offline CPU has callbacks queued.
23902360
*/
23912361
init_completion(&rsp->barrier_completion);
23922362
atomic_set(&rsp->barrier_cpu_count, 1);
2393-
raw_spin_lock_irqsave(&rsp->onofflock, flags);
2394-
rsp->rcu_barrier_in_progress = current;
2395-
raw_spin_unlock_irqrestore(&rsp->onofflock, flags);
2363+
get_online_cpus();
23962364

23972365
/*
2398-
* Force every CPU with callbacks to register a new callback
2399-
* that will tell us when all the preceding callbacks have
2400-
* been invoked. If an offline CPU has callbacks, wait for
2401-
* it to either come back online or to finish orphaning those
2402-
* callbacks.
2366+
* Force each CPU with callbacks to register a new callback.
2367+
* When that callback is invoked, we will know that all of the
2368+
* corresponding CPU's preceding callbacks have been invoked.
24032369
*/
2404-
for_each_possible_cpu(cpu) {
2405-
preempt_disable();
2370+
for_each_online_cpu(cpu) {
24062371
rdp = per_cpu_ptr(rsp->rda, cpu);
2407-
if (cpu_is_offline(cpu)) {
2408-
_rcu_barrier_trace(rsp, "Offline", cpu,
2409-
rsp->n_barrier_done);
2410-
preempt_enable();
2411-
while (cpu_is_offline(cpu) && ACCESS_ONCE(rdp->qlen))
2412-
schedule_timeout_interruptible(1);
2413-
} else if (ACCESS_ONCE(rdp->qlen)) {
2372+
if (ACCESS_ONCE(rdp->qlen)) {
24142373
_rcu_barrier_trace(rsp, "OnlineQ", cpu,
24152374
rsp->n_barrier_done);
24162375
smp_call_function_single(cpu, rcu_barrier_func, rsp, 1);
2417-
preempt_enable();
24182376
} else {
24192377
_rcu_barrier_trace(rsp, "OnlineNQ", cpu,
24202378
rsp->n_barrier_done);
2421-
preempt_enable();
24222379
}
24232380
}
2424-
2425-
/*
2426-
* Now that all online CPUs have rcu_barrier_callback() callbacks
2427-
* posted, we can adopt all of the orphaned callbacks and place
2428-
* an rcu_barrier_callback() callback after them. When that is done,
2429-
* we are guaranteed to have an rcu_barrier_callback() callback
2430-
* following every callback that could possibly have been
2431-
* registered before _rcu_barrier() was called.
2432-
*/
2433-
raw_spin_lock_irqsave(&rsp->onofflock, flags);
2434-
rcu_adopt_orphan_cbs(rsp);
2435-
rsp->rcu_barrier_in_progress = NULL;
2436-
raw_spin_unlock_irqrestore(&rsp->onofflock, flags);
2437-
atomic_inc(&rsp->barrier_cpu_count);
2438-
smp_mb__after_atomic_inc(); /* Ensure atomic_inc() before callback. */
2439-
rd.rsp = rsp;
2440-
rsp->call(&rd.barrier_head, rcu_barrier_callback);
2381+
put_online_cpus();
24412382

24422383
/*
24432384
* Now that we have an rcu_barrier_callback() callback on each
@@ -2458,8 +2399,6 @@ static void _rcu_barrier(struct rcu_state *rsp)
24582399

24592400
/* Other rcu_barrier() invocations can now safely proceed. */
24602401
mutex_unlock(&rsp->barrier_mutex);
2461-
2462-
destroy_rcu_head_on_stack(&rd.barrier_head);
24632402
}
24642403

24652404
/**

kernel/rcutree.h

-3
Original file line numberDiff line numberDiff line change
@@ -398,9 +398,6 @@ struct rcu_state {
398398
struct rcu_head **orphan_donetail; /* Tail of above. */
399399
long qlen_lazy; /* Number of lazy callbacks. */
400400
long qlen; /* Total number of callbacks. */
401-
struct task_struct *rcu_barrier_in_progress;
402-
/* Task doing rcu_barrier(), */
403-
/* or NULL if no barrier. */
404401
struct mutex barrier_mutex; /* Guards barrier fields. */
405402
atomic_t barrier_cpu_count; /* # CPUs waiting on. */
406403
struct completion barrier_completion; /* Wake at barrier end. */

kernel/rcutree_trace.c

+2-2
Original file line numberDiff line numberDiff line change
@@ -51,8 +51,8 @@ static int show_rcubarrier(struct seq_file *m, void *unused)
5151
struct rcu_state *rsp;
5252

5353
for_each_rcu_flavor(rsp)
54-
seq_printf(m, "%s: %c bcc: %d nbd: %lu\n",
55-
rsp->name, rsp->rcu_barrier_in_progress ? 'B' : '.',
54+
seq_printf(m, "%s: bcc: %d nbd: %lu\n",
55+
rsp->name,
5656
atomic_read(&rsp->barrier_cpu_count),
5757
rsp->n_barrier_done);
5858
return 0;

0 commit comments

Comments
 (0)