rcu: Avoid needlessly IPIing CPUs at GP end
If a CPU enters dyntick-idle mode with callbacks pending, it will need an IPI at the end of the grace period. However, if it exits dyntick-idle mode before the grace period ends, it will be needlessly IPIed at the end of the grace period. Therefore, this commit clears the per-CPU rcu_awake_at_gp_end flag when a CPU determines that it does not need it. This in turn requires disabling interrupts across much of rcu_prepare_for_idle() in order to avoid having nested interrupts clearing this state out from under us. Signed-off-by: Paul E. McKenney <paul.mckenney@linaro.org> Signed-off-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
This commit is contained in:
committed by
Paul E. McKenney
parent
3084f2f80c
commit
84ad00cb61
@@ -2027,6 +2027,9 @@ int rcu_needs_cpu(int cpu)
|
|||||||
static void rcu_prepare_for_idle(int cpu)
|
static void rcu_prepare_for_idle(int cpu)
|
||||||
{
|
{
|
||||||
int c = 0;
|
int c = 0;
|
||||||
|
unsigned long flags;
|
||||||
|
|
||||||
|
local_irq_save(flags);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* If there are no callbacks on this CPU or if RCU has no further
|
* If there are no callbacks on this CPU or if RCU has no further
|
||||||
@@ -2036,14 +2039,17 @@ static void rcu_prepare_for_idle(int cpu)
|
|||||||
if (!rcu_cpu_has_callbacks(cpu)) {
|
if (!rcu_cpu_has_callbacks(cpu)) {
|
||||||
per_cpu(rcu_dyntick_holdoff, cpu) = jiffies - 1;
|
per_cpu(rcu_dyntick_holdoff, cpu) = jiffies - 1;
|
||||||
per_cpu(rcu_dyntick_drain, cpu) = 0;
|
per_cpu(rcu_dyntick_drain, cpu) = 0;
|
||||||
|
per_cpu(rcu_awake_at_gp_end, cpu) = 0;
|
||||||
|
local_irq_restore(flags);
|
||||||
trace_rcu_prep_idle("No callbacks");
|
trace_rcu_prep_idle("No callbacks");
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
if (!rcu_pending(cpu)) {
|
if (!rcu_pending(cpu)) {
|
||||||
trace_rcu_prep_idle("Dyntick with callbacks");
|
|
||||||
per_cpu(rcu_dyntick_holdoff, cpu) = jiffies - 1;
|
per_cpu(rcu_dyntick_holdoff, cpu) = jiffies - 1;
|
||||||
per_cpu(rcu_dyntick_drain, cpu) = 0;
|
per_cpu(rcu_dyntick_drain, cpu) = 0;
|
||||||
per_cpu(rcu_awake_at_gp_end, cpu) = 1;
|
per_cpu(rcu_awake_at_gp_end, cpu) = 1;
|
||||||
|
local_irq_restore(flags);
|
||||||
|
trace_rcu_prep_idle("Dyntick with callbacks");
|
||||||
return; /* Nothing to do immediately. */
|
return; /* Nothing to do immediately. */
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -2052,6 +2058,7 @@ static void rcu_prepare_for_idle(int cpu)
|
|||||||
* refrained from disabling the scheduling-clock tick.
|
* refrained from disabling the scheduling-clock tick.
|
||||||
*/
|
*/
|
||||||
if (per_cpu(rcu_dyntick_holdoff, cpu) == jiffies) {
|
if (per_cpu(rcu_dyntick_holdoff, cpu) == jiffies) {
|
||||||
|
local_irq_restore(flags);
|
||||||
trace_rcu_prep_idle("In holdoff");
|
trace_rcu_prep_idle("In holdoff");
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
@@ -2060,9 +2067,11 @@ static void rcu_prepare_for_idle(int cpu)
|
|||||||
if (per_cpu(rcu_dyntick_drain, cpu) <= 0) {
|
if (per_cpu(rcu_dyntick_drain, cpu) <= 0) {
|
||||||
/* First time through, initialize the counter. */
|
/* First time through, initialize the counter. */
|
||||||
per_cpu(rcu_dyntick_drain, cpu) = RCU_NEEDS_CPU_FLUSHES;
|
per_cpu(rcu_dyntick_drain, cpu) = RCU_NEEDS_CPU_FLUSHES;
|
||||||
|
per_cpu(rcu_awake_at_gp_end, cpu) = 0;
|
||||||
} else if (--per_cpu(rcu_dyntick_drain, cpu) <= 0) {
|
} else if (--per_cpu(rcu_dyntick_drain, cpu) <= 0) {
|
||||||
/* We have hit the limit, so time to give up. */
|
/* We have hit the limit, so time to give up. */
|
||||||
per_cpu(rcu_dyntick_holdoff, cpu) = jiffies;
|
per_cpu(rcu_dyntick_holdoff, cpu) = jiffies;
|
||||||
|
local_irq_restore(flags);
|
||||||
trace_rcu_prep_idle("Begin holdoff");
|
trace_rcu_prep_idle("Begin holdoff");
|
||||||
invoke_rcu_core(); /* Force the CPU out of dyntick-idle. */
|
invoke_rcu_core(); /* Force the CPU out of dyntick-idle. */
|
||||||
return;
|
return;
|
||||||
@@ -2095,10 +2104,13 @@ static void rcu_prepare_for_idle(int cpu)
|
|||||||
* So try forcing the callbacks through the grace period.
|
* So try forcing the callbacks through the grace period.
|
||||||
*/
|
*/
|
||||||
if (c) {
|
if (c) {
|
||||||
|
local_irq_restore(flags);
|
||||||
trace_rcu_prep_idle("More callbacks");
|
trace_rcu_prep_idle("More callbacks");
|
||||||
invoke_rcu_core();
|
invoke_rcu_core();
|
||||||
} else
|
} else {
|
||||||
|
local_irq_restore(flags);
|
||||||
trace_rcu_prep_idle("Callbacks drained");
|
trace_rcu_prep_idle("Callbacks drained");
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
Reference in New Issue
Block a user