rcu: Split hierarchical RCU initialization into boot-time and CPU-online pieces
This patch divides the rcutree initialization into boot-time and hotplug-time components, so that the tree data structures are guaranteed to be fully linked at boot time regardless of what might happen in CPU hotplug operations. This makes RCU more resilient against CPU hotplug misbehavior (and vice versa), but more importantly, does a better job of compartmentalizing the code. Reported-by: Ingo Molnar <mingo@elte.hu> Signed-off-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com> Cc: laijs@cn.fujitsu.com Cc: dipankar@in.ibm.com Cc: josht@linux.vnet.ibm.com Cc: akpm@linux-foundation.org Cc: mathieu.desnoyers@polymtl.ca Cc: dvhltc@us.ibm.com Cc: niv@us.ibm.com Cc: peterz@infradead.org Cc: rostedt@goodmis.org Cc: hugh.dickins@tiscali.co.uk Cc: benh@kernel.crashing.org LKML-Reference: <1250355231152-git-send-email-> Signed-off-by: Ingo Molnar <mingo@elte.hu>
This commit is contained in:
committed by
Ingo Molnar
parent
fa08661af8
commit
27569620c7
@@ -1325,22 +1325,40 @@ int rcu_needs_cpu(int cpu)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Initialize a CPU's per-CPU RCU data. We take this "scorched earth"
|
* Do boot-time initialization of a CPU's per-CPU RCU data.
|
||||||
* approach so that we don't have to worry about how long the CPU has
|
*/
|
||||||
* been gone, or whether it ever was online previously. We do trust the
|
static void __init
|
||||||
* ->mynode field, as it is constant for a given struct rcu_data and
|
rcu_boot_init_percpu_data(int cpu, struct rcu_state *rsp)
|
||||||
* initialized during early boot.
|
{
|
||||||
*
|
unsigned long flags;
|
||||||
* Note that only one online or offline event can be happening at a given
|
int i;
|
||||||
* time. Note also that we can accept some slop in the rsp->completed
|
struct rcu_data *rdp = rsp->rda[cpu];
|
||||||
* access due to the fact that this CPU cannot possibly have any RCU
|
struct rcu_node *rnp = rcu_get_root(rsp);
|
||||||
* callbacks in flight yet.
|
|
||||||
|
/* Set up local state, ensuring consistent view of global state. */
|
||||||
|
spin_lock_irqsave(&rnp->lock, flags);
|
||||||
|
rdp->grpmask = 1UL << (cpu - rdp->mynode->grplo);
|
||||||
|
rdp->nxtlist = NULL;
|
||||||
|
for (i = 0; i < RCU_NEXT_SIZE; i++)
|
||||||
|
rdp->nxttail[i] = &rdp->nxtlist;
|
||||||
|
rdp->qlen = 0;
|
||||||
|
#ifdef CONFIG_NO_HZ
|
||||||
|
rdp->dynticks = &per_cpu(rcu_dynticks, cpu);
|
||||||
|
#endif /* #ifdef CONFIG_NO_HZ */
|
||||||
|
rdp->cpu = cpu;
|
||||||
|
spin_unlock_irqrestore(&rnp->lock, flags);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Initialize a CPU's per-CPU RCU data. Note that only one online or
|
||||||
|
* offline event can be happening at a given time. Note also that we
|
||||||
|
* can accept some slop in the rsp->completed access due to the fact
|
||||||
|
* that this CPU cannot possibly have any RCU callbacks in flight yet.
|
||||||
*/
|
*/
|
||||||
static void __cpuinit
|
static void __cpuinit
|
||||||
rcu_init_percpu_data(int cpu, struct rcu_state *rsp)
|
rcu_init_percpu_data(int cpu, struct rcu_state *rsp)
|
||||||
{
|
{
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
int i;
|
|
||||||
long lastcomp;
|
long lastcomp;
|
||||||
unsigned long mask;
|
unsigned long mask;
|
||||||
struct rcu_data *rdp = rsp->rda[cpu];
|
struct rcu_data *rdp = rsp->rda[cpu];
|
||||||
@@ -1355,16 +1373,7 @@ rcu_init_percpu_data(int cpu, struct rcu_state *rsp)
|
|||||||
rdp->qs_pending = 1; /* so set up to respond to current GP. */
|
rdp->qs_pending = 1; /* so set up to respond to current GP. */
|
||||||
rdp->beenonline = 1; /* We have now been online. */
|
rdp->beenonline = 1; /* We have now been online. */
|
||||||
rdp->passed_quiesc_completed = lastcomp - 1;
|
rdp->passed_quiesc_completed = lastcomp - 1;
|
||||||
rdp->grpmask = 1UL << (cpu - rdp->mynode->grplo);
|
|
||||||
rdp->nxtlist = NULL;
|
|
||||||
for (i = 0; i < RCU_NEXT_SIZE; i++)
|
|
||||||
rdp->nxttail[i] = &rdp->nxtlist;
|
|
||||||
rdp->qlen = 0;
|
|
||||||
rdp->blimit = blimit;
|
rdp->blimit = blimit;
|
||||||
#ifdef CONFIG_NO_HZ
|
|
||||||
rdp->dynticks = &per_cpu(rcu_dynticks, cpu);
|
|
||||||
#endif /* #ifdef CONFIG_NO_HZ */
|
|
||||||
rdp->cpu = cpu;
|
|
||||||
spin_unlock(&rnp->lock); /* irqs remain disabled. */
|
spin_unlock(&rnp->lock); /* irqs remain disabled. */
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@@ -1539,8 +1548,12 @@ void __init __rcu_init(void)
|
|||||||
#endif /* #ifdef CONFIG_RCU_CPU_STALL_DETECTOR */
|
#endif /* #ifdef CONFIG_RCU_CPU_STALL_DETECTOR */
|
||||||
rcu_init_one(&rcu_state);
|
rcu_init_one(&rcu_state);
|
||||||
RCU_DATA_PTR_INIT(&rcu_state, rcu_data);
|
RCU_DATA_PTR_INIT(&rcu_state, rcu_data);
|
||||||
|
for_each_possible_cpu(i)
|
||||||
|
rcu_boot_init_percpu_data(i, &rcu_state);
|
||||||
rcu_init_one(&rcu_bh_state);
|
rcu_init_one(&rcu_bh_state);
|
||||||
RCU_DATA_PTR_INIT(&rcu_bh_state, rcu_bh_data);
|
RCU_DATA_PTR_INIT(&rcu_bh_state, rcu_bh_data);
|
||||||
|
for_each_possible_cpu(i)
|
||||||
|
rcu_boot_init_percpu_data(i, &rcu_bh_state);
|
||||||
|
|
||||||
for_each_online_cpu(i)
|
for_each_online_cpu(i)
|
||||||
rcu_cpu_notify(&rcu_nb, CPU_UP_PREPARE, (void *)(long)i);
|
rcu_cpu_notify(&rcu_nb, CPU_UP_PREPARE, (void *)(long)i);
|
||||||
|
Reference in New Issue
Block a user